X-Git-Url: https://git.kernel.dk/?p=fio.git;a=blobdiff_plain;f=fio.c;h=73964214d3468d93a4f0ed72f4ecbd330f57b9c6;hp=fa30019bfbeddd2bbbb756d069384468d343d1d2;hb=e07f72d341ded8cfa539abf377322ab1b7473960;hpb=687c6a230dfeae89454ed430ee49ed5813e146ad diff --git a/fio.c b/fio.c index fa30019b..73964214 100644 --- a/fio.c +++ b/fio.c @@ -102,7 +102,9 @@ static void terminate_threads(int group_id) /* * if the thread is running, just let it exit */ - if (td->runstate < TD_RAMP) + if (!td->pid) + continue; + else if (td->runstate < TD_RAMP) kill(td->pid, SIGTERM); else { struct ioengine_ops *ops = td->io_ops; @@ -548,13 +550,14 @@ sync_done: /* * if we can queue more, do so. but check if there are - * completed io_u's first. + * completed io_u's first. Note that we can get BUSY even + * without IO queued, if the system is resource starved. */ - full = queue_full(td) || ret == FIO_Q_BUSY; + full = queue_full(td) || (ret == FIO_Q_BUSY && td->cur_depth); if (full || !td->o.iodepth_batch_complete) { min_events = min(td->o.iodepth_batch_complete, td->cur_depth); - if (full && !min_events) + if (full && !min_events && td->o.iodepth_batch_complete != 0) min_events = 1; do { @@ -708,13 +711,15 @@ sync_done: break; /* - * See if we need to complete some commands + * See if we need to complete some commands. Note that we + * can get BUSY even without IO queued, if the system is + * resource starved. */ - full = queue_full(td) || ret == FIO_Q_BUSY; + full = queue_full(td) || (ret == FIO_Q_BUSY && td->cur_depth); if (full || !td->o.iodepth_batch_complete) { min_evts = min(td->o.iodepth_batch_complete, td->cur_depth); - if (full && !min_evts) + if (full && !min_evts && td->o.iodepth_batch_complete != 0) min_evts = 1; if (__should_check_rate(td, 0) || @@ -978,11 +983,6 @@ static void reset_io_counters(struct thread_data *td) */ if (td->o.time_based || td->o.loops) td->nr_done_files = 0; - - /* - * Set the same seed to get repeatable runs - */ - td_fill_rand_seeds(td); } void reset_all_stats(struct thread_data *td) @@ -1016,6 +1016,11 @@ static void clear_io_state(struct thread_data *td) close_files(td); for_each_file(td, f, i) fio_file_clear_done(f); + + /* + * Set the same seed to get repeatable runs + */ + td_fill_rand_seeds(td); } static int exec_string(const char *string) @@ -1045,10 +1050,11 @@ static void *thread_main(void *data) pthread_condattr_t attr; int clear_state; - if (!td->o.use_thread) + if (!td->o.use_thread) { setsid(); - - td->pid = getpid(); + td->pid = getpid(); + } else + td->pid = gettid(); dprint(FD_PROCESS, "jobs pid=%d started\n", (int) td->pid); @@ -1092,6 +1098,22 @@ static void *thread_main(void *data) goto err; } + /* + * If we have a gettimeofday() thread, make sure we exclude that + * thread from this job + */ + if (td->o.gtod_cpu) + fio_cpu_clear(&td->o.cpumask, td->o.gtod_cpu); + + /* + * Set affinity first, in case it has an impact on the memory + * allocations. + */ + if (td->o.cpumask_set && fio_setaffinity(td->pid, td->o.cpumask) == -1) { + td_verror(td, errno, "cpu_set_affinity"); + goto err; + } + /* * May alter parameters that init_io_u() will use, so we need to * do this first. @@ -1105,23 +1127,6 @@ static void *thread_main(void *data) if (td->o.verify_async && verify_async_init(td)) goto err; - if (td->o.cpumask_set && fio_setaffinity(td->pid, td->o.cpumask) == -1) { - td_verror(td, errno, "cpu_set_affinity"); - goto err; - } - - /* - * If we have a gettimeofday() thread, make sure we exclude that - * thread from this job - */ - if (td->o.gtod_cpu) { - fio_cpu_clear(&td->o.cpumask, td->o.gtod_cpu); - if (fio_setaffinity(td->pid, td->o.cpumask) == -1) { - td_verror(td, errno, "cpu_set_affinity"); - goto err; - } - } - if (td->ioprio_set) { if (ioprio_set(IOPRIO_WHO_PROCESS, 0, td->ioprio) == -1) { td_verror(td, errno, "ioprio_set"); @@ -1295,6 +1300,7 @@ static int fork_main(int shmid, int offset) struct thread_data *td; void *data, *ret; +#ifndef __hpux data = shmat(shmid, NULL, 0); if (data == (void *) -1) { int __err = errno; @@ -1302,6 +1308,12 @@ static int fork_main(int shmid, int offset) perror("shmat"); return __err; } +#else + /* + * HP-UX inherits shm mappings? + */ + data = threads; +#endif td = data + offset * sizeof(struct thread_data); ret = thread_main(td);