X-Git-Url: https://git.kernel.dk/?a=blobdiff_plain;f=fio.c;h=9c1bed379df586ba8f3efea187aaaf319ce2d387;hb=d3cc4ebf58ec05037c3b0dd7c83c075a56f1dc62;hp=9d9027d792e4ccbb8cd4f6469b89b88ed210b6e0;hpb=9a793c2f91a47df348237f1a9b778253ca87ec2e;p=fio.git diff --git a/fio.c b/fio.c index 9d9027d7..9c1bed37 100644 --- a/fio.c +++ b/fio.c @@ -45,6 +45,7 @@ #include "cgroup.h" #include "profile.h" #include "lib/rand.h" +#include "memalign.h" unsigned long page_mask; unsigned long page_size; @@ -69,6 +70,8 @@ static pthread_t disk_util_thread; static struct flist_head *cgroup_list; static char *cgroup_mnt; +unsigned long arch_flags = 0; + struct io_log *agg_io_log[2]; #define TERMINATE_ALL (-1) @@ -101,7 +104,9 @@ static void terminate_threads(int group_id) /* * if the thread is running, just let it exit */ - if (td->runstate < TD_RUNNING) + if (!td->pid) + continue; + else if (td->runstate < TD_RAMP) kill(td->pid, SIGTERM); else { struct ioengine_ops *ops = td->io_ops; @@ -113,18 +118,12 @@ static void terminate_threads(int group_id) } } -/* - * Happens on thread runs with ctrl-c, ignore our own SIGQUIT - */ -static void sig_quit(int sig) -{ -} - static void sig_int(int sig) { if (threads) { log_info("\nfio: terminating on signal %d\n", sig); fflush(stdout); + exit_value = 128; terminate_threads(TERMINATE_ALL); } } @@ -146,13 +145,9 @@ static void *disk_thread_main(void *data) static int create_disk_util_thread(void) { - pthread_attr_t attr; int ret; - pthread_attr_init(&attr); - pthread_attr_setstacksize(&attr, PTHREAD_STACK_MIN); - ret = pthread_create(&disk_util_thread, &attr, disk_thread_main, NULL); - pthread_attr_destroy(&attr); + ret = pthread_create(&disk_util_thread, NULL, disk_thread_main, NULL); if (ret) { log_err("Can't create disk util thread: %s\n", strerror(ret)); return 1; @@ -180,7 +175,7 @@ static void set_sig_handlers(void) sigaction(SIGINT, &act, NULL); memset(&act, 0, sizeof(act)); - act.sa_handler = sig_quit; + act.sa_handler = sig_int; act.sa_flags = SA_RESTART; sigaction(SIGTERM, &act, NULL); } @@ -189,7 +184,7 @@ static void set_sig_handlers(void) * Check if we are above the minimum rate given. */ static int __check_min_rate(struct thread_data *td, struct timeval *now, - enum td_ddir ddir) + enum fio_ddir ddir) { unsigned long long bytes = 0; unsigned long iops = 0; @@ -381,10 +376,15 @@ requeue: return 0; } +static inline void __update_tv_cache(struct thread_data *td) +{ + fio_gettime(&td->tv_cache, NULL); +} + static inline void update_tv_cache(struct thread_data *td) { if ((++td->tv_cache_nr & td->tv_cache_mask) == td->tv_cache_mask) - fio_gettime(&td->tv_cache, NULL); + __update_tv_cache(td); } static int break_on_this_error(struct thread_data *td, int *retptr) @@ -470,8 +470,11 @@ static void do_verify(struct thread_data *td) update_tv_cache(td); if (runtime_exceeded(td, &td->tv_cache)) { - td->terminate = 1; - break; + __update_tv_cache(td); + if (runtime_exceeded(td, &td->tv_cache)) { + td->terminate = 1; + break; + } } io_u = __get_io_u(td); @@ -501,7 +504,6 @@ static void do_verify(struct thread_data *td) clear_io_u(td, io_u); } else if (io_u->resid) { int bytes = io_u->xfer_buflen - io_u->resid; - struct fio_file *f = io_u->file; /* * zero read, fail @@ -519,6 +521,7 @@ static void do_verify(struct thread_data *td) if (ddir_rw(io_u->ddir)) td->ts.short_io_u[io_u->ddir]++; + f = io_u->file; if (io_u->offset == f->real_file_size) goto sync_done; @@ -549,13 +552,14 @@ sync_done: /* * if we can queue more, do so. but check if there are - * completed io_u's first. + * completed io_u's first. Note that we can get BUSY even + * without IO queued, if the system is resource starved. */ - full = queue_full(td) || ret == FIO_Q_BUSY; + full = queue_full(td) || (ret == FIO_Q_BUSY && td->cur_depth); if (full || !td->o.iodepth_batch_complete) { min_events = min(td->o.iodepth_batch_complete, td->cur_depth); - if (full && !min_events) + if (full && !min_events && td->o.iodepth_batch_complete != 0) min_events = 1; do { @@ -616,8 +620,11 @@ static void do_io(struct thread_data *td) update_tv_cache(td); if (runtime_exceeded(td, &td->tv_cache)) { - td->terminate = 1; - break; + __update_tv_cache(td); + if (runtime_exceeded(td, &td->tv_cache)) { + td->terminate = 1; + break; + } } io_u = get_io_u(td); @@ -706,13 +713,15 @@ sync_done: break; /* - * See if we need to complete some commands + * See if we need to complete some commands. Note that we + * can get BUSY even without IO queued, if the system is + * resource starved. */ - full = queue_full(td) || ret == FIO_Q_BUSY; + full = queue_full(td) || (ret == FIO_Q_BUSY && td->cur_depth); if (full || !td->o.iodepth_batch_complete) { min_evts = min(td->o.iodepth_batch_complete, td->cur_depth); - if (full && !min_evts) + if (full && !min_evts && td->o.iodepth_batch_complete != 0) min_evts = 1; if (__should_check_rate(td, 0) || @@ -769,8 +778,11 @@ sync_done: struct fio_file *f; i = td->cur_depth; - if (i) + if (i) { ret = io_u_queued_complete(td, i, NULL); + if (td->o.fill_device && td->error == ENOSPC) + td->error = 0; + } if (should_fsync(td) && td->o.end_fsync) { td_set_runstate(td, TD_FSYNCING); @@ -800,7 +812,7 @@ static void cleanup_io_u(struct thread_data *td) io_u = flist_entry(entry, struct io_u, list); flist_del(&io_u->list); - free(io_u); + fio_memfree(io_u, sizeof(*io_u)); } free_io_mem(td); @@ -847,8 +859,9 @@ static int init_io_u(struct thread_data *td) if (td->terminate) return 1; - if (posix_memalign(&ptr, cl_align, sizeof(*io_u))) { - log_err("fio: posix_memalign=%s\n", strerror(errno)); + ptr = fio_memalign(cl_align, sizeof(*io_u)); + if (!ptr) { + log_err("fio: unable to allocate aligned memory\n"); break; } @@ -972,11 +985,6 @@ static void reset_io_counters(struct thread_data *td) */ if (td->o.time_based || td->o.loops) td->nr_done_files = 0; - - /* - * Set the same seed to get repeatable runs - */ - td_fill_rand_seeds(td); } void reset_all_stats(struct thread_data *td) @@ -1010,6 +1018,11 @@ static void clear_io_state(struct thread_data *td) close_files(td); for_each_file(td, f, i) fio_file_clear_done(f); + + /* + * Set the same seed to get repeatable runs + */ + td_fill_rand_seeds(td); } static int exec_string(const char *string) @@ -1039,10 +1052,11 @@ static void *thread_main(void *data) pthread_condattr_t attr; int clear_state; - if (!td->o.use_thread) + if (!td->o.use_thread) { setsid(); - - td->pid = getpid(); + td->pid = getpid(); + } else + td->pid = gettid(); dprint(FD_PROCESS, "jobs pid=%d started\n", (int) td->pid); @@ -1086,6 +1100,22 @@ static void *thread_main(void *data) goto err; } + /* + * If we have a gettimeofday() thread, make sure we exclude that + * thread from this job + */ + if (td->o.gtod_cpu) + fio_cpu_clear(&td->o.cpumask, td->o.gtod_cpu); + + /* + * Set affinity first, in case it has an impact on the memory + * allocations. + */ + if (td->o.cpumask_set && fio_setaffinity(td->pid, td->o.cpumask) == -1) { + td_verror(td, errno, "cpu_set_affinity"); + goto err; + } + /* * May alter parameters that init_io_u() will use, so we need to * do this first. @@ -1099,23 +1129,6 @@ static void *thread_main(void *data) if (td->o.verify_async && verify_async_init(td)) goto err; - if (td->o.cpumask_set && fio_setaffinity(td->pid, td->o.cpumask) == -1) { - td_verror(td, errno, "cpu_set_affinity"); - goto err; - } - - /* - * If we have a gettimeofday() thread, make sure we exclude that - * thread from this job - */ - if (td->o.gtod_cpu) { - fio_cpu_clear(&td->o.cpumask, td->o.gtod_cpu); - if (fio_setaffinity(td->pid, td->o.cpumask) == -1) { - td_verror(td, errno, "cpu_set_affinity"); - goto err; - } - } - if (td->ioprio_set) { if (ioprio_set(IOPRIO_WHO_PROCESS, 0, td->ioprio) == -1) { td_verror(td, errno, "ioprio_set"); @@ -1289,6 +1302,7 @@ static int fork_main(int shmid, int offset) struct thread_data *td; void *data, *ret; +#ifndef __hpux data = shmat(shmid, NULL, 0); if (data == (void *) -1) { int __err = errno; @@ -1296,6 +1310,12 @@ static int fork_main(int shmid, int offset) perror("shmat"); return __err; } +#else + /* + * HP-UX inherits shm mappings? + */ + data = threads; +#endif td = data + offset * sizeof(struct thread_data); ret = thread_main(td); @@ -1501,14 +1521,14 @@ static void run_threads(void) todo--; } else { struct fio_file *f; - unsigned int i; + unsigned int j; /* * for sharing to work, each job must always open * its own files. so close them, if we opened them * for creation */ - for_each_file(td, f, i) { + for_each_file(td, f, j) { if (fio_file_open(f)) td_io_close_file(td, f); } @@ -1520,7 +1540,7 @@ static void run_threads(void) set_genesis_time(); while (todo) { - struct thread_data *map[MAX_JOBS]; + struct thread_data *map[REAL_MAX_JOBS]; struct timeval this_start; int this_jobs = 0, left; @@ -1672,10 +1692,12 @@ static void run_threads(void) fio_unpin_memory(); } -int main(int argc, char *argv[]) +int main(int argc, char *argv[], char *envp[]) { long ps; + arch_init(envp); + sinit(); init_rand(&__fio_rand_state);