static void sig_handler(int sig)
{
+ if (!threads)
+ return;
+
switch (sig) {
case SIGALRM:
update_io_ticks();
*/
min_events = 0;
if (queue_full(td) || ret == FIO_Q_BUSY) {
- min_events = 1;
-
- if (td->cur_depth > td->o.iodepth_low)
+ if (td->cur_depth >= td->o.iodepth_low)
min_events = td->cur_depth - td->o.iodepth_low;
+ if (!min_events)
+ min_events = 1;
}
/*
/*
* See if we need to complete some commands
*/
- if (ret == FIO_Q_QUEUED || ret == FIO_Q_BUSY) {
+ if (queue_full(td) || ret == FIO_Q_BUSY) {
min_evts = 0;
- if (queue_full(td) || ret == FIO_Q_BUSY) {
+ if (td->cur_depth >= td->o.iodepth_low)
+ min_evts = td->cur_depth - td->o.iodepth_low;
+ if (!min_evts)
min_evts = 1;
-
- if (td->cur_depth > td->o.iodepth_low)
- min_evts = td->cur_depth
- - td->o.iodepth_low;
- }
-
fio_gettime(&comp_time, NULL);
bytes_done = io_u_queued_complete(td, min_evts);
if (bytes_done < 0)
free_io_mem(td);
}
-/*
- * "randomly" fill the buffer contents
- */
-static void fill_io_buf(struct thread_data *td, struct io_u *io_u, int max_bs)
-{
- long *ptr = io_u->buf;
-
- if (!td->o.zero_buffers) {
- while ((void *) ptr - io_u->buf < max_bs) {
- *ptr = rand() * GOLDEN_RATIO_PRIME;
- ptr++;
- }
- } else
- memset(ptr, 0, max_bs);
-}
-
static int init_io_u(struct thread_data *td)
{
struct io_u *io_u;
if (!(td->io_ops->flags & FIO_NOIO)) {
io_u->buf = p + max_bs * i;
- if (td_write(td))
- fill_io_buf(td, io_u, max_bs);
+ if (td_write(td) && !td->o.refill_buffers)
+ io_u_fill_buffer(td, io_u, max_bs);
}
io_u->index = i;
if (pthread_detach(td->thread) < 0)
perror("pthread_detach");
} else {
+ pid_t pid;
dprint(FD_PROCESS, "will fork\n");
- if (!fork()) {
+ pid = fork();
+ if (!pid) {
int ret = fork_main(shm_id, i);
- exit(ret);
- }
+ _exit(ret);
+ } else if (i == fio_debug_jobno)
+ *fio_debug_jobp = pid;
}
fio_mutex_down(startup_mutex);
}