projects
/
fio.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
[PATCH] Add seperate read/write block size options
[fio.git]
/
io_u.c
diff --git
a/io_u.c
b/io_u.c
index 738f3e20101eee5e3535f1ff52fd344286c45745..7698e84581c3399eb771d9ecf1f5df472d688311 100644
(file)
--- a/
io_u.c
+++ b/
io_u.c
@@
-27,10
+27,13
@@
static int random_map_free(struct thread_data *td, struct fio_file *f,
static void mark_random_map(struct thread_data *td, struct fio_file *f,
struct io_u *io_u)
{
static void mark_random_map(struct thread_data *td, struct fio_file *f,
struct io_u *io_u)
{
- unsigned long long block = io_u->offset / (unsigned long long) td->min_bs;
- unsigned int blocks = 0;
+ unsigned int min_bs = td->min_bs[io_u->ddir];
+ unsigned long long block;
+ unsigned int blocks;
- while (blocks < (io_u->buflen / td->min_bs)) {
+ block = io_u->offset / (unsigned long long) min_bs;
+ blocks = 0;
+ while (blocks < (io_u->buflen / min_bs)) {
unsigned int idx, bit;
if (!random_map_free(td, f, block))
unsigned int idx, bit;
if (!random_map_free(td, f, block))
@@
-46,8
+49,8
@@
static void mark_random_map(struct thread_data *td, struct fio_file *f,
blocks++;
}
blocks++;
}
- if ((blocks *
td->
min_bs) < io_u->buflen)
- io_u->buflen = blocks *
td->
min_bs;
+ if ((blocks * min_bs) < io_u->buflen)
+ io_u->buflen = blocks * min_bs;
}
/*
}
/*
@@
-60,7
+63,7
@@
static int get_next_free_block(struct thread_data *td, struct fio_file *f,
*b = 0;
i = 0;
*b = 0;
i = 0;
- while ((*b) * td->min_bs < f->file_size) {
+ while ((*b) * td->
rw_
min_bs < f->file_size) {
if (f->file_map[i] != -1UL) {
*b += ffz(f->file_map[i]);
return 0;
if (f->file_map[i] != -1UL) {
*b += ffz(f->file_map[i]);
return 0;
@@
-79,19
+82,21
@@
static int get_next_free_block(struct thread_data *td, struct fio_file *f,
* the last io issued.
*/
static int get_next_offset(struct thread_data *td, struct fio_file *f,
* the last io issued.
*/
static int get_next_offset(struct thread_data *td, struct fio_file *f,
- unsigned long long *offset)
+ unsigned long long *offset
, int ddir
)
{
unsigned long long b, rb;
long r;
if (!td->sequential) {
{
unsigned long long b, rb;
long r;
if (!td->sequential) {
- unsigned long long max_blocks = td->io_size / td->min_bs;
+ unsigned long long max_blocks = td->io_size / td->min_bs
[ddir]
;
int loops = 50;
do {
r = os_random_long(&td->random_state);
b = ((max_blocks - 1) * r / (unsigned long long) (RAND_MAX+1.0));
int loops = 50;
do {
r = os_random_long(&td->random_state);
b = ((max_blocks - 1) * r / (unsigned long long) (RAND_MAX+1.0));
- rb = b + (f->file_offset / td->min_bs);
+ if (td->norandommap)
+ break;
+ rb = b + (f->file_offset / td->min_bs[ddir]);
loops--;
} while (!random_map_free(td, f, rb) && loops);
loops--;
} while (!random_map_free(td, f, rb) && loops);
@@
-100,29
+105,30
@@
static int get_next_offset(struct thread_data *td, struct fio_file *f,
return 1;
}
} else
return 1;
}
} else
- b = f->last_pos / td->min_bs;
+ b = f->last_pos / td->min_bs
[ddir]
;
- *offset = (b * td->min_bs) + f->file_offset;
+ *offset = (b * td->min_bs
[ddir]
) + f->file_offset;
if (*offset > f->file_size)
return 1;
return 0;
}
if (*offset > f->file_size)
return 1;
return 0;
}
-static unsigned int get_next_buflen(struct thread_data *td)
+static unsigned int get_next_buflen(struct thread_data *td
, int ddir
)
{
unsigned int buflen;
long r;
{
unsigned int buflen;
long r;
- if (td->min_bs
== td->max_bs
)
- buflen = td->min_bs;
+ if (td->min_bs
[ddir] == td->max_bs[ddir]
)
+ buflen = td->min_bs
[ddir]
;
else {
r = os_random_long(&td->bsrange_state);
else {
r = os_random_long(&td->bsrange_state);
- buflen = (1 + (double) (td->max_bs - 1) * r / (RAND_MAX + 1.0));
- buflen = (buflen + td->min_bs - 1) & ~(td->min_bs - 1);
+ buflen = (1 + (double) (td->max_bs[ddir] - 1) * r / (RAND_MAX + 1.0));
+ if (!td->bs_unaligned)
+ buflen = (buflen + td->min_bs[ddir] - 1) & ~(td->min_bs[ddir] - 1);
}
}
- if (buflen > td->io_size - td->this_io_bytes[
td->
ddir]) {
+ if (buflen > td->io_size - td->this_io_bytes[ddir]) {
/*
* if using direct/raw io, we may not be able to
* shrink the size. so just fail it.
/*
* if using direct/raw io, we may not be able to
* shrink the size. so just fail it.
@@
-130,7
+136,7
@@
static unsigned int get_next_buflen(struct thread_data *td)
if (td->io_ops->flags & FIO_RAWIO)
return 0;
if (td->io_ops->flags & FIO_RAWIO)
return 0;
- buflen = td->io_size - td->this_io_bytes[
td->
ddir];
+ buflen = td->io_size - td->this_io_bytes[ddir];
}
return buflen;
}
return buflen;
@@
-199,19
+205,18
@@
static int fill_io_u(struct thread_data *td, struct fio_file *f,
return 0;
}
return 0;
}
+ io_u->ddir = get_rw_ddir(td);
+
/*
* No log, let the seq/rand engine retrieve the next position.
*/
/*
* No log, let the seq/rand engine retrieve the next position.
*/
- if (!get_next_offset(td, f, &io_u->offset)) {
- io_u->buflen = get_next_buflen(td);
-
+ if (!get_next_offset(td, f, &io_u->offset, io_u->ddir)) {
+ io_u->buflen = get_next_buflen(td, io_u->ddir);
if (io_u->buflen) {
if (io_u->buflen) {
- io_u->ddir = get_rw_ddir(td);
-
/*
* If using a write iolog, store this entry.
*/
/*
* If using a write iolog, store this entry.
*/
- if (td->write_iolog)
+ if (td->write_iolog
_file
)
write_iolog_put(td, io_u);
io_u->file = f;
write_iolog_put(td, io_u);
io_u->file = f;
@@
-229,6
+234,7
@@
struct io_u *__get_io_u(struct thread_data *td)
if (!queue_full(td)) {
io_u = list_entry(td->io_u_freelist.next, struct io_u, list);
if (!queue_full(td)) {
io_u = list_entry(td->io_u_freelist.next, struct io_u, list);
+ io_u->buflen = 0;
io_u->error = 0;
io_u->resid = 0;
list_del(&io_u->list);
io_u->error = 0;
io_u->resid = 0;
list_del(&io_u->list);
@@
-276,7
+282,7
@@
struct io_u *get_io_u(struct thread_data *td, struct fio_file *f)
return NULL;
}
return NULL;
}
- if (!td->read_iolog && !td->sequential)
+ if (!td->read_iolog && !td->sequential
&& !td->norandommap
)
mark_random_map(td, f, io_u);
f->last_pos += io_u->buflen;
mark_random_map(td, f, io_u);
f->last_pos += io_u->buflen;