projects
/
fio.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
t/axmap: don't print 'pass' on failure
[fio.git]
/
backend.c
diff --git
a/backend.c
b/backend.c
index d82d49487c84991b8105508cd419cd557b58182b..a7e91843a0be087ff552d050a48aef8220cc2a8d 100644
(file)
--- a/
backend.c
+++ b/
backend.c
@@
-22,29
+22,17
@@
*
*/
#include <unistd.h>
*
*/
#include <unistd.h>
-#include <fcntl.h>
#include <string.h>
#include <string.h>
-#include <limits.h>
#include <signal.h>
#include <signal.h>
-#include <time.h>
-#include <locale.h>
#include <assert.h>
#include <assert.h>
-#include <time.h>
#include <inttypes.h>
#include <sys/stat.h>
#include <sys/wait.h>
#include <inttypes.h>
#include <sys/stat.h>
#include <sys/wait.h>
-#include <sys/ipc.h>
-#include <sys/mman.h>
#include <math.h>
#include "fio.h"
#include <math.h>
#include "fio.h"
-#ifndef FIO_NO_HAVE_SHM_H
-#include <sys/shm.h>
-#endif
-#include "hash.h"
#include "smalloc.h"
#include "verify.h"
#include "smalloc.h"
#include "verify.h"
-#include "trim.h"
#include "diskutil.h"
#include "cgroup.h"
#include "profile.h"
#include "diskutil.h"
#include "cgroup.h"
#include "profile.h"
@@
-62,7
+50,7
@@
static struct fio_sem *startup_sem;
static struct flist_head *cgroup_list;
static struct fio_sem *startup_sem;
static struct flist_head *cgroup_list;
-static
char
*cgroup_mnt;
+static
struct cgroup_mnt
*cgroup_mnt;
static int exit_value;
static volatile int fio_abort;
static unsigned int nr_process = 0;
static int exit_value;
static volatile int fio_abort;
static unsigned int nr_process = 0;
@@
-280,7
+268,7
@@
static void cleanup_pending_aio(struct thread_data *td)
static bool fio_io_sync(struct thread_data *td, struct fio_file *f)
{
struct io_u *io_u = __get_io_u(td);
static bool fio_io_sync(struct thread_data *td, struct fio_file *f)
{
struct io_u *io_u = __get_io_u(td);
-
int
ret;
+
enum fio_q_status
ret;
if (!io_u)
return true;
if (!io_u)
return true;
@@
-295,16
+283,13
@@
static bool fio_io_sync(struct thread_data *td, struct fio_file *f)
requeue:
ret = td_io_queue(td, io_u);
requeue:
ret = td_io_queue(td, io_u);
- if (ret < 0) {
- td_verror(td, io_u->error, "td_io_queue");
- put_io_u(td, io_u);
- return true;
- } else if (ret == FIO_Q_QUEUED) {
- if (td_io_commit(td))
- return true;
+ switch (ret) {
+ case FIO_Q_QUEUED:
+ td_io_commit(td);
if (io_u_queued_complete(td, 1) < 0)
return true;
if (io_u_queued_complete(td, 1) < 0)
return true;
- } else if (ret == FIO_Q_COMPLETED) {
+ break;
+ case FIO_Q_COMPLETED:
if (io_u->error) {
td_verror(td, io_u->error, "td_io_queue");
return true;
if (io_u->error) {
td_verror(td, io_u->error, "td_io_queue");
return true;
@@
-312,9
+297,9
@@
requeue:
if (io_u_sync_complete(td, io_u) < 0)
return true;
if (io_u_sync_complete(td, io_u) < 0)
return true;
- } else if (ret == FIO_Q_BUSY) {
- if (td_io_commit(td))
-
return true
;
+ break;
+ case FIO_Q_BUSY:
+
td_io_commit(td)
;
goto requeue;
}
goto requeue;
}
@@
-447,9
+432,7
@@
static int wait_for_completions(struct thread_data *td, struct timespec *time)
if ((full && !min_evts) || !td->o.iodepth_batch_complete_min)
min_evts = 1;
if ((full && !min_evts) || !td->o.iodepth_batch_complete_min)
min_evts = 1;
- if (time && (__should_check_rate(td, DDIR_READ) ||
- __should_check_rate(td, DDIR_WRITE) ||
- __should_check_rate(td, DDIR_TRIM)))
+ if (time && __should_check_rate(td))
fio_gettime(time, NULL);
do {
fio_gettime(time, NULL);
do {
@@
-465,8
+448,6
@@
int io_queue_event(struct thread_data *td, struct io_u *io_u, int *ret,
enum fio_ddir ddir, uint64_t *bytes_issued, int from_verify,
struct timespec *comp_time)
{
enum fio_ddir ddir, uint64_t *bytes_issued, int from_verify,
struct timespec *comp_time)
{
- int ret2;
-
switch (*ret) {
case FIO_Q_COMPLETED:
if (io_u->error) {
switch (*ret) {
case FIO_Q_COMPLETED:
if (io_u->error) {
@@
-480,7
+461,7
@@
int io_queue_event(struct thread_data *td, struct io_u *io_u, int *ret,
*bytes_issued += bytes;
if (!from_verify)
*bytes_issued += bytes;
if (!from_verify)
- trim_io_piece(
td,
io_u);
+ trim_io_piece(io_u);
/*
* zero read, fail
/*
* zero read, fail
@@
-506,9
+487,7
@@
int io_queue_event(struct thread_data *td, struct io_u *io_u, int *ret,
requeue_io_u(td, &io_u);
} else {
sync_done:
requeue_io_u(td, &io_u);
} else {
sync_done:
- if (comp_time && (__should_check_rate(td, DDIR_READ) ||
- __should_check_rate(td, DDIR_WRITE) ||
- __should_check_rate(td, DDIR_TRIM)))
+ if (comp_time && __should_check_rate(td))
fio_gettime(comp_time, NULL);
*ret = io_u_sync_complete(td, io_u);
fio_gettime(comp_time, NULL);
*ret = io_u_sync_complete(td, io_u);
@@
-542,9
+521,7
@@
sync_done:
if (!from_verify)
unlog_io_piece(td, io_u);
requeue_io_u(td, &io_u);
if (!from_verify)
unlog_io_piece(td, io_u);
requeue_io_u(td, &io_u);
- ret2 = td_io_commit(td);
- if (ret2 < 0)
- *ret = ret2;
+ td_io_commit(td);
break;
default:
assert(*ret < 0);
break;
default:
assert(*ret < 0);
@@
-617,7
+594,7
@@
static bool in_flight_overlap(struct io_u_queue *q, struct io_u *io_u)
return overlap;
}
return overlap;
}
-static
int
io_u_submit(struct thread_data *td, struct io_u *io_u)
+static
enum fio_q_status
io_u_submit(struct thread_data *td, struct io_u *io_u)
{
/*
* Check for overlap if the user asked us to, and we have
{
/*
* Check for overlap if the user asked us to, and we have
@@
-735,6
+712,7
@@
static void do_verify(struct thread_data *td, uint64_t verify_bytes)
break;
} else if (io_u->ddir == DDIR_WRITE) {
io_u->ddir = DDIR_READ;
break;
} else if (io_u->ddir == DDIR_WRITE) {
io_u->ddir = DDIR_READ;
+ populate_verify_io_u(td, io_u);
break;
} else {
put_io_u(td, io_u);
break;
} else {
put_io_u(td, io_u);
@@
-910,6
+888,8
@@
static void handle_thinktime(struct thread_data *td, enum fio_ddir ddir)
over = (usperop - total) / usperop * -bs;
td->rate_io_issue_bytes[ddir] += (missed - over);
over = (usperop - total) / usperop * -bs;
td->rate_io_issue_bytes[ddir] += (missed - over);
+ /* adjust for rate_process=poisson */
+ td->last_usec[ddir] += total;
}
}
}
}
@@
-1007,6
+987,9
@@
static void do_io(struct thread_data *td, uint64_t *bytes_done)
break;
}
break;
}
+ if (io_u->ddir == DDIR_WRITE && td->flags & TD_F_DO_VERIFY)
+ populate_verify_io_u(td, io_u);
+
ddir = io_u->ddir;
/*
ddir = io_u->ddir;
/*
@@
-1051,7
+1034,7
@@
static void do_io(struct thread_data *td, uint64_t *bytes_done)
if (td->o.io_submit_mode == IO_MODE_OFFLOAD) {
const unsigned long blen = io_u->xfer_buflen;
if (td->o.io_submit_mode == IO_MODE_OFFLOAD) {
const unsigned long blen = io_u->xfer_buflen;
- const enum fio_ddir ddir = acct_ddir(io_u);
+ const enum fio_ddir
__
ddir = acct_ddir(io_u);
if (td->error)
break;
if (td->error)
break;
@@
-1059,14
+1042,14
@@
static void do_io(struct thread_data *td, uint64_t *bytes_done)
workqueue_enqueue(&td->io_wq, &io_u->work);
ret = FIO_Q_QUEUED;
workqueue_enqueue(&td->io_wq, &io_u->work);
ret = FIO_Q_QUEUED;
- if (ddir_rw(ddir)) {
- td->io_issues[ddir]++;
- td->io_issue_bytes[ddir] += blen;
- td->rate_io_issue_bytes[ddir] += blen;
+ if (ddir_rw(
__
ddir)) {
+ td->io_issues[
__
ddir]++;
+ td->io_issue_bytes[
__
ddir] += blen;
+ td->rate_io_issue_bytes[
__
ddir] += blen;
}
if (should_check_rate(td))
}
if (should_check_rate(td))
- td->rate_next_io_time[
ddir] = usec_for_io(td,
ddir);
+ td->rate_next_io_time[
__ddir] = usec_for_io(td, __
ddir);
} else {
ret = io_u_submit(td, io_u);
} else {
ret = io_u_submit(td, io_u);
@@
-1340,7
+1323,7
@@
static int init_io_u(struct thread_data *td)
static int switch_ioscheduler(struct thread_data *td)
{
#ifdef FIO_HAVE_IOSCHED_SWITCH
static int switch_ioscheduler(struct thread_data *td)
{
#ifdef FIO_HAVE_IOSCHED_SWITCH
- char tmp[256], tmp2[128];
+ char tmp[256], tmp2[128]
, *p
;
FILE *f;
int ret;
FILE *f;
int ret;
@@
-1376,17
+1359,19
@@
static int switch_ioscheduler(struct thread_data *td)
/*
* Read back and check that the selected scheduler is now the default.
*/
/*
* Read back and check that the selected scheduler is now the default.
*/
- memset(tmp, 0, sizeof(tmp));
- ret = fread(tmp, sizeof(tmp), 1, f);
+ ret = fread(tmp, 1, sizeof(tmp) - 1, f);
if (ferror(f) || ret < 0) {
td_verror(td, errno, "fread");
fclose(f);
return 1;
}
if (ferror(f) || ret < 0) {
td_verror(td, errno, "fread");
fclose(f);
return 1;
}
+ tmp[ret] = '\0';
/*
/*
- * either a list of io schedulers or "none\n" is expected.
+ * either a list of io schedulers or "none\n" is expected. Strip the
+ * trailing newline.
*/
*/
- tmp[strlen(tmp) - 1] = '\0';
+ p = tmp;
+ strsep(&p, "\n");
/*
* Write to "none" entry doesn't fail, so check the result here.
/*
* Write to "none" entry doesn't fail, so check the result here.
@@
-1544,7
+1529,7
@@
static void *thread_main(void *data)
} else
td->pid = gettid();
} else
td->pid = gettid();
- fio_local_clock_init(
o->use_thread
);
+ fio_local_clock_init();
dprint(FD_PROCESS, "jobs pid=%d started\n", (int) td->pid);
dprint(FD_PROCESS, "jobs pid=%d started\n", (int) td->pid);
@@
-1555,7
+1540,6
@@
static void *thread_main(void *data)
INIT_FLIST_HEAD(&td->io_hist_list);
INIT_FLIST_HEAD(&td->verify_list);
INIT_FLIST_HEAD(&td->trim_list);
INIT_FLIST_HEAD(&td->io_hist_list);
INIT_FLIST_HEAD(&td->verify_list);
INIT_FLIST_HEAD(&td->trim_list);
- INIT_FLIST_HEAD(&td->next_rand_list);
td->io_hist_tree = RB_ROOT;
ret = mutex_cond_init_pshared(&td->io_u_lock, &td->free_cond);
td->io_hist_tree = RB_ROOT;
ret = mutex_cond_init_pshared(&td->io_u_lock, &td->free_cond);
@@
-1684,7
+1668,7
@@
static void *thread_main(void *data)
* May alter parameters that init_io_u() will use, so we need to
* do this first.
*/
* May alter parameters that init_io_u() will use, so we need to
* do this first.
*/
- if (init_iolog(td))
+ if (
!
init_iolog(td))
goto err;
if (init_io_u(td))
goto err;
if (init_io_u(td))
@@
-1902,7
+1886,7
@@
err:
close_and_free_files(td);
cleanup_io_u(td);
close_ioengine(td);
close_and_free_files(td);
cleanup_io_u(td);
close_ioengine(td);
- cgroup_shutdown(td,
&
cgroup_mnt);
+ cgroup_shutdown(td, cgroup_mnt);
verify_free_state(td);
if (td->zone_state_index) {
verify_free_state(td);
if (td->zone_state_index) {
@@
-2489,7
+2473,8
@@
int fio_backend(struct sk_out *sk_out)
helper_thread_create(startup_sem, sk_out);
cgroup_list = smalloc(sizeof(*cgroup_list));
helper_thread_create(startup_sem, sk_out);
cgroup_list = smalloc(sizeof(*cgroup_list));
- INIT_FLIST_HEAD(cgroup_list);
+ if (cgroup_list)
+ INIT_FLIST_HEAD(cgroup_list);
run_threads(sk_out);
run_threads(sk_out);
@@
-2519,9
+2504,10
@@
int fio_backend(struct sk_out *sk_out)
}
free_disk_util();
}
free_disk_util();
- cgroup_kill(cgroup_list);
- sfree(cgroup_list);
- sfree(cgroup_mnt);
+ if (cgroup_list) {
+ cgroup_kill(cgroup_list);
+ sfree(cgroup_list);
+ }
fio_sem_remove(startup_sem);
stat_exit();
fio_sem_remove(startup_sem);
stat_exit();