X-Git-Url: https://git.kernel.dk/?p=fio.git;a=blobdiff_plain;f=fio.c;h=d1faacae860deaf97c90e9a8e8142aabb899f205;hp=76c6ebd73b25e0eefb1b093ef5914530ee5b9389;hb=cc0df00ad5076d4adbc439899f24d9b0db26075d;hpb=b20b6be43e835cf84a87821d268751642ab64de5 diff --git a/fio.c b/fio.c index 76c6ebd7..d1faacae 100644 --- a/fio.c +++ b/fio.c @@ -46,6 +46,7 @@ #include "profile.h" #include "lib/rand.h" #include "memalign.h" +#include "server.h" unsigned long page_mask; unsigned long page_size; @@ -70,9 +71,10 @@ static pthread_t disk_util_thread; static struct flist_head *cgroup_list; static char *cgroup_mnt; +unsigned long arch_flags = 0; + struct io_log *agg_io_log[2]; -#define TERMINATE_ALL (-1) #define JOB_START_TIMEOUT (5 * 1000) void td_set_runstate(struct thread_data *td, int runstate) @@ -85,7 +87,7 @@ void td_set_runstate(struct thread_data *td, int runstate) td->runstate = runstate; } -static void terminate_threads(int group_id) +void fio_terminate_threads(int group_id) { struct thread_data *td; int i; @@ -102,7 +104,9 @@ static void terminate_threads(int group_id) /* * if the thread is running, just let it exit */ - if (td->runstate < TD_RUNNING) + if (!td->pid) + continue; + else if (td->runstate < TD_RAMP) kill(td->pid, SIGTERM); else { struct ioengine_ops *ops = td->io_ops; @@ -118,9 +122,10 @@ static void sig_int(int sig) { if (threads) { log_info("\nfio: terminating on signal %d\n", sig); + exit_backend = 1; fflush(stdout); exit_value = 128; - terminate_threads(TERMINATE_ALL); + fio_terminate_threads(TERMINATE_ALL); } } @@ -133,7 +138,11 @@ static void *disk_thread_main(void *data) if (!threads) break; update_io_ticks(); - print_thread_status(); + + if (is_backend) + fio_server_send_status(); + else + print_thread_status(); } return NULL; @@ -174,13 +183,20 @@ static void set_sig_handlers(void) act.sa_handler = sig_int; act.sa_flags = SA_RESTART; sigaction(SIGTERM, &act, NULL); + + if (is_backend) { + memset(&act, 0, sizeof(act)); + act.sa_handler = sig_int; + act.sa_flags = SA_RESTART; + sigaction(SIGPIPE, &act, NULL); + } } /* * Check if we are above the minimum rate given. */ static int __check_min_rate(struct thread_data *td, struct timeval *now, - enum td_ddir ddir) + enum fio_ddir ddir) { unsigned long long bytes = 0; unsigned long iops = 0; @@ -372,10 +388,15 @@ requeue: return 0; } +static inline void __update_tv_cache(struct thread_data *td) +{ + fio_gettime(&td->tv_cache, NULL); +} + static inline void update_tv_cache(struct thread_data *td) { if ((++td->tv_cache_nr & td->tv_cache_mask) == td->tv_cache_mask) - fio_gettime(&td->tv_cache, NULL); + __update_tv_cache(td); } static int break_on_this_error(struct thread_data *td, int *retptr) @@ -461,8 +482,11 @@ static void do_verify(struct thread_data *td) update_tv_cache(td); if (runtime_exceeded(td, &td->tv_cache)) { - td->terminate = 1; - break; + __update_tv_cache(td); + if (runtime_exceeded(td, &td->tv_cache)) { + td->terminate = 1; + break; + } } io_u = __get_io_u(td); @@ -540,13 +564,14 @@ sync_done: /* * if we can queue more, do so. but check if there are - * completed io_u's first. + * completed io_u's first. Note that we can get BUSY even + * without IO queued, if the system is resource starved. */ - full = queue_full(td) || ret == FIO_Q_BUSY; + full = queue_full(td) || (ret == FIO_Q_BUSY && td->cur_depth); if (full || !td->o.iodepth_batch_complete) { min_events = min(td->o.iodepth_batch_complete, td->cur_depth); - if (full && !min_events) + if (full && !min_events && td->o.iodepth_batch_complete != 0) min_events = 1; do { @@ -607,8 +632,11 @@ static void do_io(struct thread_data *td) update_tv_cache(td); if (runtime_exceeded(td, &td->tv_cache)) { - td->terminate = 1; - break; + __update_tv_cache(td); + if (runtime_exceeded(td, &td->tv_cache)) { + td->terminate = 1; + break; + } } io_u = get_io_u(td); @@ -697,13 +725,15 @@ sync_done: break; /* - * See if we need to complete some commands + * See if we need to complete some commands. Note that we + * can get BUSY even without IO queued, if the system is + * resource starved. */ - full = queue_full(td) || ret == FIO_Q_BUSY; + full = queue_full(td) || (ret == FIO_Q_BUSY && td->cur_depth); if (full || !td->o.iodepth_batch_complete) { min_evts = min(td->o.iodepth_batch_complete, td->cur_depth); - if (full && !min_evts) + if (full && !min_evts && td->o.iodepth_batch_complete != 0) min_evts = 1; if (__should_check_rate(td, 0) || @@ -726,7 +756,7 @@ sync_done: if (!in_ramp_time(td) && should_check_rate(td, bytes_done)) { if (check_min_rate(td, &comp_time, bytes_done)) { if (exitall_on_terminate) - terminate_threads(td->groupid); + fio_terminate_threads(td->groupid); td_verror(td, EIO, "check_min_rate"); break; } @@ -760,8 +790,11 @@ sync_done: struct fio_file *f; i = td->cur_depth; - if (i) + if (i) { ret = io_u_queued_complete(td, i, NULL); + if (td->o.fill_device && td->error == ENOSPC) + td->error = 0; + } if (should_fsync(td) && td->o.end_fsync) { td_set_runstate(td, TD_FSYNCING); @@ -853,9 +886,9 @@ static int init_io_u(struct thread_data *td) io_u->buf = p + max_bs * i; dprint(FD_MEM, "io_u %p, mem %p\n", io_u, io_u->buf); - if (td_write(td) && !td->o.refill_buffers) + if (td_write(td)) io_u_fill_buffer(td, io_u, max_bs); - else if (td_write(td) && td->o.verify_pattern_bytes) { + if (td_write(td) && td->o.verify_pattern_bytes) { /* * Fill the buffer with the pattern if we are * going to be doing writes. @@ -951,7 +984,7 @@ static int keep_running(struct thread_data *td) static void reset_io_counters(struct thread_data *td) { - td->ts.stat_io_bytes[0] = td->ts.stat_io_bytes[1] = 0; + td->stat_io_bytes[0] = td->stat_io_bytes[1] = 0; td->this_io_bytes[0] = td->this_io_bytes[1] = 0; td->zone_bytes = 0; td->rate_bytes[0] = td->rate_bytes[1] = 0; @@ -964,11 +997,6 @@ static void reset_io_counters(struct thread_data *td) */ if (td->o.time_based || td->o.loops) td->nr_done_files = 0; - - /* - * Set the same seed to get repeatable runs - */ - td_fill_rand_seeds(td); } void reset_all_stats(struct thread_data *td) @@ -1002,6 +1030,11 @@ static void clear_io_state(struct thread_data *td) close_files(td); for_each_file(td, f, i) fio_file_clear_done(f); + + /* + * Set the same seed to get repeatable runs + */ + td_fill_rand_seeds(td); } static int exec_string(const char *string) @@ -1031,10 +1064,11 @@ static void *thread_main(void *data) pthread_condattr_t attr; int clear_state; - if (!td->o.use_thread) + if (!td->o.use_thread) { setsid(); - - td->pid = getpid(); + td->pid = getpid(); + } else + td->pid = gettid(); dprint(FD_PROCESS, "jobs pid=%d started\n", (int) td->pid); @@ -1078,6 +1112,22 @@ static void *thread_main(void *data) goto err; } + /* + * If we have a gettimeofday() thread, make sure we exclude that + * thread from this job + */ + if (td->o.gtod_cpu) + fio_cpu_clear(&td->o.cpumask, td->o.gtod_cpu); + + /* + * Set affinity first, in case it has an impact on the memory + * allocations. + */ + if (td->o.cpumask_set && fio_setaffinity(td->pid, td->o.cpumask) == -1) { + td_verror(td, errno, "cpu_set_affinity"); + goto err; + } + /* * May alter parameters that init_io_u() will use, so we need to * do this first. @@ -1091,23 +1141,6 @@ static void *thread_main(void *data) if (td->o.verify_async && verify_async_init(td)) goto err; - if (td->o.cpumask_set && fio_setaffinity(td->pid, td->o.cpumask) == -1) { - td_verror(td, errno, "cpu_set_affinity"); - goto err; - } - - /* - * If we have a gettimeofday() thread, make sure we exclude that - * thread from this job - */ - if (td->o.gtod_cpu) { - fio_cpu_clear(&td->o.cpumask, td->o.gtod_cpu); - if (fio_setaffinity(td->pid, td->o.cpumask) == -1) { - td_verror(td, errno, "cpu_set_affinity"); - goto err; - } - } - if (td->ioprio_set) { if (ioprio_set(IOPRIO_WHO_PROCESS, 0, td->ioprio) == -1) { td_verror(td, errno, "ioprio_set"); @@ -1146,19 +1179,17 @@ static void *thread_main(void *data) } fio_gettime(&td->epoch, NULL); - getrusage(RUSAGE_SELF, &td->ts.ru_start); + getrusage(RUSAGE_SELF, &td->ru_start); clear_state = 0; while (keep_running(td)) { fio_gettime(&td->start, NULL); - memcpy(&td->ts.stat_sample_time[0], &td->start, - sizeof(td->start)); - memcpy(&td->ts.stat_sample_time[1], &td->start, - sizeof(td->start)); + memcpy(&td->stat_sample_time[0], &td->start, sizeof(td->start)); + memcpy(&td->stat_sample_time[1], &td->start, sizeof(td->start)); memcpy(&td->tv_cache, &td->start, sizeof(td->start)); if (td->o.ratemin[0] || td->o.ratemin[1]) - memcpy(&td->lastrate, &td->ts.stat_sample_time, + memcpy(&td->lastrate, &td->stat_sample_time, sizeof(td->lastrate)); if (clear_state) @@ -1207,40 +1238,40 @@ static void *thread_main(void *data) td->ts.io_bytes[1] = td->io_bytes[1]; fio_mutex_down(writeout_mutex); - if (td->ts.bw_log) { + if (td->bw_log) { if (td->o.bw_log_file) { - finish_log_named(td, td->ts.bw_log, + finish_log_named(td, td->bw_log, td->o.bw_log_file, "bw"); } else - finish_log(td, td->ts.bw_log, "bw"); + finish_log(td, td->bw_log, "bw"); } - if (td->ts.lat_log) { + if (td->lat_log) { if (td->o.lat_log_file) { - finish_log_named(td, td->ts.lat_log, + finish_log_named(td, td->lat_log, td->o.lat_log_file, "lat"); } else - finish_log(td, td->ts.lat_log, "lat"); + finish_log(td, td->lat_log, "lat"); } - if (td->ts.slat_log) { + if (td->slat_log) { if (td->o.lat_log_file) { - finish_log_named(td, td->ts.slat_log, + finish_log_named(td, td->slat_log, td->o.lat_log_file, "slat"); } else - finish_log(td, td->ts.slat_log, "slat"); + finish_log(td, td->slat_log, "slat"); } - if (td->ts.clat_log) { + if (td->clat_log) { if (td->o.lat_log_file) { - finish_log_named(td, td->ts.clat_log, + finish_log_named(td, td->clat_log, td->o.lat_log_file, "clat"); } else - finish_log(td, td->ts.clat_log, "clat"); + finish_log(td, td->clat_log, "clat"); } fio_mutex_up(writeout_mutex); if (td->o.exec_postrun) exec_string(td->o.exec_postrun); if (exitall_on_terminate) - terminate_threads(td->groupid); + fio_terminate_threads(td->groupid); err: if (td->error) @@ -1281,6 +1312,7 @@ static int fork_main(int shmid, int offset) struct thread_data *td; void *data, *ret; +#ifndef __hpux data = shmat(shmid, NULL, 0); if (data == (void *) -1) { int __err = errno; @@ -1288,6 +1320,12 @@ static int fork_main(int shmid, int offset) perror("shmat"); return __err; } +#else + /* + * HP-UX inherits shm mappings? + */ + data = threads; +#endif td = data + offset * sizeof(struct thread_data); ret = thread_main(td); @@ -1387,7 +1425,7 @@ reaped: } if (*nr_running == cputhreads && !pending && realthreads) - terminate_threads(TERMINATE_ALL); + fio_terminate_threads(TERMINATE_ALL); } static void *gtod_thread_main(void *data) @@ -1449,6 +1487,8 @@ static void run_threads(void) if (fio_gtod_offload && fio_start_gtod_thread()) return; + set_sig_handlers(); + if (!terse_output) { log_info("Starting "); if (nr_thread) @@ -1464,8 +1504,6 @@ static void run_threads(void) fflush(stdout); } - set_sig_handlers(); - todo = thread_number; nr_running = 0; nr_started = 0; @@ -1474,10 +1512,8 @@ static void run_threads(void) for_each_td(td, i) { print_status_init(td->thread_number - 1); - if (!td->o.create_serialize) { - init_disk_util(td); + if (!td->o.create_serialize) continue; - } /* * do file setup here so it happens sequentially, @@ -1505,14 +1541,12 @@ static void run_threads(void) td_io_close_file(td, f); } } - - init_disk_util(td); } set_genesis_time(); while (todo) { - struct thread_data *map[MAX_JOBS]; + struct thread_data *map[REAL_MAX_JOBS]; struct timeval this_start; int this_jobs = 0, left; @@ -1545,6 +1579,8 @@ static void run_threads(void) break; } + init_disk_util(td); + /* * Set state to created. Thread will transition * to TD_INITIALIZED when it's done setting up. @@ -1583,7 +1619,7 @@ static void run_threads(void) dprint(FD_MUTEX, "wait on startup_mutex\n"); if (fio_mutex_down_timeout(startup_mutex, 10)) { log_err("fio: job startup hung? exiting.\n"); - terminate_threads(TERMINATE_ALL); + fio_terminate_threads(TERMINATE_ALL); fio_abort = 1; nr_started--; break; @@ -1657,6 +1693,10 @@ static void run_threads(void) while (nr_running) { reap_threads(&nr_running, &t_rate, &m_rate); + + if (is_backend) + fio_server_idle_loop(); + usleep(10000); } @@ -1664,34 +1704,10 @@ static void run_threads(void) fio_unpin_memory(); } -int main(int argc, char *argv[]) +int exec_run(void) { - long ps; - - sinit(); - init_rand(&__fio_rand_state); - - /* - * We need locale for number printing, if it isn't set then just - * go with the US format. - */ - if (!getenv("LC_NUMERIC")) - setlocale(LC_NUMERIC, "en_US"); - - ps = sysconf(_SC_PAGESIZE); - if (ps < 0) { - log_err("Failed to get page size\n"); - return 1; - } - - page_size = ps; - page_mask = ps - 1; - - fio_keywords_init(); - - if (parse_options(argc, argv)) - return 1; - + if (nr_clients) + return fio_handle_clients(); if (exec_profile && load_profile(exec_profile)) return 1; @@ -1735,3 +1751,44 @@ int main(int argc, char *argv[]) fio_mutex_remove(writeout_mutex); return exit_value; } + +void reset_fio_state(void) +{ + groupid = 0; + thread_number = 0; + nr_process = 0; + nr_thread = 0; + done_secs = 0; +} + +int main(int argc, char *argv[], char *envp[]) +{ + long ps; + + arch_init(envp); + + sinit(); + + /* + * We need locale for number printing, if it isn't set then just + * go with the US format. + */ + if (!getenv("LC_NUMERIC")) + setlocale(LC_NUMERIC, "en_US"); + + ps = sysconf(_SC_PAGESIZE); + if (ps < 0) { + log_err("Failed to get page size\n"); + return 1; + } + + page_size = ps; + page_mask = ps - 1; + + fio_keywords_init(); + + if (parse_options(argc, argv)) + return 1; + + return exec_run(); +}