projects
/
fio.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
Silence put_file() complaint
[fio.git]
/
io_u.c
diff --git
a/io_u.c
b/io_u.c
index faa1c5e36a8d84ac558d6867a7e796ec5a71d8fe..92cdd71701a9df3973eb4158b0a8fcb6bab4eed5 100644
(file)
--- a/
io_u.c
+++ b/
io_u.c
@@
-24,11
+24,10
@@
struct io_completion_data {
* The ->file_map[] contains a map of blocks we have or have not done io
* to yet. Used to make sure we cover the entire range in a fair fashion.
*/
* The ->file_map[] contains a map of blocks we have or have not done io
* to yet. Used to make sure we cover the entire range in a fair fashion.
*/
-static int random_map_free(struct thread_data *td, struct fio_file *f,
- const unsigned long long block)
+static int random_map_free(struct fio_file *f, const unsigned long long block)
{
{
- unsigned int idx = RAND_MAP_IDX(
td,
f, block);
- unsigned int bit = RAND_MAP_BIT(
td,
f, block);
+ unsigned int idx = RAND_MAP_IDX(f, block);
+ unsigned int bit = RAND_MAP_BIT(f, block);
dprint(FD_RANDOM, "free: b=%llu, idx=%u, bit=%u\n", block, idx, bit);
dprint(FD_RANDOM, "free: b=%llu, idx=%u, bit=%u\n", block, idx, bit);
@@
-57,11
+56,11
@@
static void mark_random_map(struct thread_data *td, struct io_u *io_u)
* If we have a mixed random workload, we may
* encounter blocks we already did IO to.
*/
* If we have a mixed random workload, we may
* encounter blocks we already did IO to.
*/
- if ((td->o.ddir_nr == 1) && !random_map_free(
td,
f, block))
+ if ((td->o.ddir_nr == 1) && !random_map_free(f, block))
break;
break;
- idx = RAND_MAP_IDX(
td,
f, block);
- bit = RAND_MAP_BIT(
td,
f, block);
+ idx = RAND_MAP_IDX(f, block);
+ bit = RAND_MAP_BIT(f, block);
fio_assert(td, idx < f->num_maps);
fio_assert(td, idx < f->num_maps);
@@
-84,7
+83,7
@@
static inline unsigned long long last_block(struct thread_data *td,
if (!max_blocks)
return 0;
if (!max_blocks)
return 0;
- return max_blocks
- 1
;
+ return max_blocks;
}
/*
}
/*
@@
-124,18
+123,19
@@
static int get_next_rand_offset(struct thread_data *td, struct fio_file *f,
do {
r = os_random_long(&td->random_state);
dprint(FD_RANDOM, "off rand %llu\n", r);
do {
r = os_random_long(&td->random_state);
dprint(FD_RANDOM, "off rand %llu\n", r);
- *b = (last_block(td, f, ddir) - 1) * (r / ((unsigned long long) RAND_MAX + 1.0));
+ *b = (last_block(td, f, ddir) - 1)
+ * (r / ((unsigned long long) RAND_MAX + 1.0));
/*
* if we are not maintaining a random map, we are done.
*/
/*
* if we are not maintaining a random map, we are done.
*/
- if (
td->o.norandommap
)
+ if (
!file_randommap(td, f)
)
return 0;
/*
* calculate map offset and check if it's free
*/
return 0;
/*
* calculate map offset and check if it's free
*/
- if (random_map_free(
td,
f, *b))
+ if (random_map_free(f, *b))
return 0;
dprint(FD_RANDOM, "get_next_rand_offset: offset %llu busy\n",
return 0;
dprint(FD_RANDOM, "get_next_rand_offset: offset %llu busy\n",
@@
-188,7
+188,14
@@
static int get_next_offset(struct thread_data *td, struct io_u *io_u)
b = (f->last_pos - f->file_offset) / td->o.min_bs[ddir];
}
b = (f->last_pos - f->file_offset) / td->o.min_bs[ddir];
}
- io_u->offset = (b * td->o.min_bs[ddir]) + f->file_offset;
+ io_u->offset = b * td->o.min_bs[ddir];
+ if (io_u->offset >= f->io_size) {
+ dprint(FD_IO, "get_next_offset: offset %llu >= io_size %llu\n",
+ io_u->offset, f->io_size);
+ return 1;
+ }
+
+ io_u->offset += f->file_offset;
if (io_u->offset >= f->real_file_size) {
dprint(FD_IO, "get_next_offset: offset %llu >= size %llu\n",
io_u->offset, f->real_file_size);
if (io_u->offset >= f->real_file_size) {
dprint(FD_IO, "get_next_offset: offset %llu >= size %llu\n",
io_u->offset, f->real_file_size);
@@
-208,9
+215,11
@@
static unsigned int get_next_buflen(struct thread_data *td, struct io_u *io_u)
buflen = td->o.min_bs[ddir];
else {
r = os_random_long(&td->bsrange_state);
buflen = td->o.min_bs[ddir];
else {
r = os_random_long(&td->bsrange_state);
- if (!td->o.bssplit_nr)
- buflen = (unsigned int) (1 + (double) (td->o.max_bs[ddir] - 1) * r / (RAND_MAX + 1.0));
- else {
+ if (!td->o.bssplit_nr) {
+ buflen = (unsigned int)
+ (1 + (double) (td->o.max_bs[ddir] - 1)
+ * r / (RAND_MAX + 1.0));
+ } else {
long perc = 0;
unsigned int i;
long perc = 0;
unsigned int i;
@@
-223,8
+232,10
@@
static unsigned int get_next_buflen(struct thread_data *td, struct io_u *io_u)
break;
}
}
break;
}
}
- if (!td->o.bs_unaligned)
- buflen = (buflen + td->o.min_bs[ddir] - 1) & ~(td->o.min_bs[ddir] - 1);
+ if (!td->o.bs_unaligned) {
+ buflen = (buflen + td->o.min_bs[ddir] - 1)
+ & ~(td->o.min_bs[ddir] - 1);
+ }
}
if (io_u->offset + buflen > io_u->file->real_file_size) {
}
if (io_u->offset + buflen > io_u->file->real_file_size) {
@@
-238,7
+249,7
@@
static unsigned int get_next_buflen(struct thread_data *td, struct io_u *io_u)
static void set_rwmix_bytes(struct thread_data *td)
{
static void set_rwmix_bytes(struct thread_data *td)
{
- unsigned long
long rbyt
es;
+ unsigned long
issu
es;
unsigned int diff;
/*
unsigned int diff;
/*
@@
-246,10
+257,11
@@
static void set_rwmix_bytes(struct thread_data *td)
* buffered writes may issue a lot quicker than they complete,
* whereas reads do not.
*/
* buffered writes may issue a lot quicker than they complete,
* whereas reads do not.
*/
-
rbytes = td->io_bytes[td->rwmix_ddir] - td->rwmix_byt
es;
+
issues = td->io_issues[td->rwmix_ddir] - td->rwmix_issu
es;
diff = td->o.rwmix[td->rwmix_ddir ^ 1];
diff = td->o.rwmix[td->rwmix_ddir ^ 1];
- td->rwmix_bytes = td->io_bytes[td->rwmix_ddir] + (rbytes * ((100 - diff)) / diff);
+ td->rwmix_issues = td->io_issues[td->rwmix_ddir]
+ + (issues * ((100 - diff)) / diff);
}
static inline enum fio_ddir get_rand_ddir(struct thread_data *td)
}
static inline enum fio_ddir get_rand_ddir(struct thread_data *td)
@@
-273,27
+285,12
@@
static inline enum fio_ddir get_rand_ddir(struct thread_data *td)
static enum fio_ddir get_rw_ddir(struct thread_data *td)
{
if (td_rw(td)) {
static enum fio_ddir get_rw_ddir(struct thread_data *td)
{
if (td_rw(td)) {
- struct timeval now;
- unsigned long elapsed;
- unsigned int cycle;
-
- fio_gettime(&now, NULL);
- elapsed = mtime_since_now(&td->rwmix_switch);
-
- /*
- * if this is the first cycle, make it shorter
- */
- cycle = td->o.rwmixcycle;
- if (!td->rwmix_bytes)
- cycle /= 10;
-
/*
* Check if it's time to seed a new data direction.
*/
/*
* Check if it's time to seed a new data direction.
*/
- if (elapsed >= cycle ||
- td->io_bytes[td->rwmix_ddir] >= td->rwmix_bytes) {
+ if (td->io_issues[td->rwmix_ddir] >= td->rwmix_issues) {
unsigned long long max_bytes;
unsigned long long max_bytes;
- enum fio_ddir ddir;
+ enum fio_ddir ddir;
/*
* Put a top limit on how many bytes we do for
/*
* Put a top limit on how many bytes we do for
@@
-302,11
+299,11
@@
static enum fio_ddir get_rw_ddir(struct thread_data *td)
*/
ddir = get_rand_ddir(td);
max_bytes = td->this_io_bytes[ddir];
*/
ddir = get_rand_ddir(td);
max_bytes = td->this_io_bytes[ddir];
- if (max_bytes >= (td->o.size * td->o.rwmix[ddir] / 100)) {
- if (!td->rw_end_set[ddir]) {
+ if (max_bytes >=
+ (td->o.size * td->o.rwmix[ddir] / 100)) {
+ if (!td->rw_end_set[ddir])
td->rw_end_set[ddir] = 1;
td->rw_end_set[ddir] = 1;
- memcpy(&td->rw_end[ddir], &now, sizeof(now));
- }
+
ddir ^= 1;
}
ddir ^= 1;
}
@@
-314,7
+311,6
@@
static enum fio_ddir get_rw_ddir(struct thread_data *td)
set_rwmix_bytes(td);
td->rwmix_ddir = ddir;
set_rwmix_bytes(td);
td->rwmix_ddir = ddir;
- memcpy(&td->rwmix_switch, &now, sizeof(now));
}
return td->rwmix_ddir;
} else if (td_read(td))
}
return td->rwmix_ddir;
} else if (td_read(td))
@@
-323,13
+319,21
@@
static enum fio_ddir get_rw_ddir(struct thread_data *td)
return DDIR_WRITE;
}
return DDIR_WRITE;
}
+static void put_file_log(struct thread_data *td, struct fio_file *f)
+{
+ int ret = put_file(td, f);
+
+ if (ret)
+ td_verror(td, ret, "file close");
+}
+
void put_io_u(struct thread_data *td, struct io_u *io_u)
{
assert((io_u->flags & IO_U_F_FREE) == 0);
io_u->flags |= IO_U_F_FREE;
if (io_u->file)
void put_io_u(struct thread_data *td, struct io_u *io_u)
{
assert((io_u->flags & IO_U_F_FREE) == 0);
io_u->flags |= IO_U_F_FREE;
if (io_u->file)
- put_file(td, io_u->file);
+ put_file
_log
(td, io_u->file);
io_u->file = NULL;
list_del(&io_u->list);
io_u->file = NULL;
list_del(&io_u->list);
@@
-344,7
+348,7
@@
void requeue_io_u(struct thread_data *td, struct io_u **io_u)
__io_u->flags |= IO_U_F_FREE;
if ((__io_u->flags & IO_U_F_FLIGHT) && (__io_u->ddir != DDIR_SYNC))
td->io_issues[__io_u->ddir]--;
__io_u->flags |= IO_U_F_FREE;
if ((__io_u->flags & IO_U_F_FLIGHT) && (__io_u->ddir != DDIR_SYNC))
td->io_issues[__io_u->ddir]--;
-
+
__io_u->flags &= ~IO_U_F_FLIGHT;
list_del(&__io_u->list);
__io_u->flags &= ~IO_U_F_FLIGHT;
list_del(&__io_u->list);
@@
-404,7
+408,7
@@
static int fill_io_u(struct thread_data *td, struct io_u *io_u)
/*
* mark entry before potentially trimming io_u
*/
/*
* mark entry before potentially trimming io_u
*/
- if (td_random(td) &&
!td->o.norandommap
)
+ if (td_random(td) &&
file_randommap(td, io_u->file)
)
mark_random_map(td, io_u);
/*
mark_random_map(td, io_u);
/*
@@
-417,13
+421,10
@@
out:
return 0;
}
return 0;
}
-void io_u_mark_depth(struct thread_data *td,
struct io_u *io_u
)
+void io_u_mark_depth(struct thread_data *td,
unsigned int nr
)
{
int index = 0;
{
int index = 0;
- if (io_u->ddir == DDIR_SYNC)
- return;
-
switch (td->cur_depth) {
default:
index = 6;
switch (td->cur_depth) {
default:
index = 6;
@@
-446,8
+447,7
@@
void io_u_mark_depth(struct thread_data *td, struct io_u *io_u)
break;
}
break;
}
- td->ts.io_u_map[index]++;
- td->ts.total_io_u[io_u->ddir]++;
+ td->ts.io_u_map[index] += nr;
}
static void io_u_mark_lat_usec(struct thread_data *td, unsigned long usec)
}
static void io_u_mark_lat_usec(struct thread_data *td, unsigned long usec)
@@
-556,8
+556,9
@@
static struct fio_file *get_next_file_rand(struct thread_data *td, int goodf,
do {
long r = os_random_long(&td->next_file_state);
do {
long r = os_random_long(&td->next_file_state);
- fno = (unsigned int) ((double) td->o.nr_files * (r / (RAND_MAX + 1.0)));
- f = &td->files[fno];
+ fno = (unsigned int) ((double) td->o.nr_files
+ * (r / (RAND_MAX + 1.0)));
+ f = td->files[fno];
if (f->flags & FIO_FILE_DONE)
continue;
if (f->flags & FIO_FILE_DONE)
continue;
@@
-578,7
+579,7
@@
static struct fio_file *get_next_file_rr(struct thread_data *td, int goodf,
struct fio_file *f;
do {
struct fio_file *f;
do {
- f =
&
td->files[td->next_file];
+ f = td->files[td->next_file];
td->next_file++;
if (td->next_file >= td->o.nr_files)
td->next_file++;
if (td->next_file >= td->o.nr_files)
@@
-606,7
+607,10
@@
static struct fio_file *get_next_file(struct thread_data *td)
assert(td->o.nr_files <= td->files_index);
if (!td->nr_open_files || td->nr_done_files >= td->o.nr_files) {
assert(td->o.nr_files <= td->files_index);
if (!td->nr_open_files || td->nr_done_files >= td->o.nr_files) {
- dprint(FD_FILE, "get_next_file: nr_open=%d, nr_done=%d, nr_files=%d\n", td->nr_open_files, td->nr_done_files, td->o.nr_files);
+ dprint(FD_FILE, "get_next_file: nr_open=%d, nr_done=%d,"
+ " nr_files=%d\n", td->nr_open_files,
+ td->nr_done_files,
+ td->o.nr_files);
return NULL;
}
return NULL;
}
@@
-657,6
+661,16
@@
set_file:
if (!fill_io_u(td, io_u))
break;
if (!fill_io_u(td, io_u))
break;
+ /*
+ * optimization to prevent close/open of the same file. This
+ * way we preserve queueing etc.
+ */
+ if (td->o.nr_files == 1 && td->o.time_based) {
+ put_file_log(td, f);
+ fio_file_reset(f);
+ goto set_file;
+ }
+
/*
* td_io_close() does a put_file() as well, so no need to
* do that here.
/*
* td_io_close() does a put_file() as well, so no need to
* do that here.
@@
-744,7
+758,7
@@
struct io_u *get_io_u(struct thread_data *td)
dprint(FD_IO, "io_u %p, setting file failed\n", io_u);
goto err_put;
}
dprint(FD_IO, "io_u %p, setting file failed\n", io_u);
goto err_put;
}
-
+
f = io_u->file;
assert(f->flags & FIO_FILE_OPEN);
f = io_u->file;
assert(f->flags & FIO_FILE_OPEN);
@@
-788,7
+802,8
@@
void io_u_log_error(struct thread_data *td, struct io_u *io_u)
log_err(": %s\n", strerror(io_u->error));
log_err(": %s\n", strerror(io_u->error));
- log_err(" %s offset=%llu, buflen=%lu\n", msg[io_u->ddir], io_u->offset, io_u->xfer_buflen);
+ log_err(" %s offset=%llu, buflen=%lu\n", msg[io_u->ddir],
+ io_u->offset, io_u->xfer_buflen);
if (!td->error)
td_verror(td, io_u->error, "io_u error");
if (!td->error)
td_verror(td, io_u->error, "io_u error");
@@
-820,8
+835,6
@@
static void io_completed(struct thread_data *td, struct io_u *io_u,
td->io_bytes[idx] += bytes;
td->this_io_bytes[idx] += bytes;
td->io_bytes[idx] += bytes;
td->this_io_bytes[idx] += bytes;
- io_u->file->last_completed_pos = io_u->endpos;
-
usec = utime_since(&io_u->issue_time, &icd->time);
add_clat_sample(td, idx, usec);
usec = utime_since(&io_u->issue_time, &icd->time);
add_clat_sample(td, idx, usec);
@@
-949,7
+962,10
@@
static void io_u_dump(struct io_u *io_u)
unsigned long t_issue = mtime_since_now(&io_u->issue_time);
log_err("io_u=%p, t_start=%lu, t_issue=%lu\n", io_u, t_start, t_issue);
unsigned long t_issue = mtime_since_now(&io_u->issue_time);
log_err("io_u=%p, t_start=%lu, t_issue=%lu\n", io_u, t_start, t_issue);
- log_err(" buf=%p/%p, len=%lu/%lu, offset=%llu\n", io_u->buf, io_u->xfer_buf, io_u->buflen, io_u->xfer_buflen, io_u->offset);
+ log_err(" buf=%p/%p, len=%lu/%lu, offset=%llu\n", io_u->buf,
+ io_u->xfer_buf, io_u->buflen,
+ io_u->xfer_buflen,
+ io_u->offset);
log_err(" ddir=%d, fname=%s\n", io_u->ddir, io_u->file->file_name);
}
#else
log_err(" ddir=%d, fname=%s\n", io_u->ddir, io_u->file->file_name);
}
#else