One step closer to group reporting
[fio.git] / engines / syslet-rw.c
index 74a6b31145167c00be9c6d9cadb974a3d69c43d5..8ab217f40b407e0c628016d8794bd90ed5f1e6aa 100644 (file)
@@ -21,41 +21,29 @@ struct syslet_data {
        struct syslet_uatom **ring;
 
        struct syslet_uatom *head, *tail;
-       struct syslet_uatom **event_map;
-       unsigned int event_map_idx;
 };
 
 static void fio_syslet_complete_atom(struct thread_data *td,
                                     struct syslet_uatom *atom)
 {
        struct syslet_data *sd = td->io_ops->data;
+       struct syslet_uatom *last;
        struct io_u *io_u;
-       int i, end;
-
-       if (!sd->event_map_idx)
-               return;
 
        /*
-        * Find the start of the string of atoms for this sequence
+        * complete from the beginning of the sequence up to (and
+        * including) this atom
         */
-       for (end = sd->event_map_idx - 1; end >= 0; end--)
-               if (atom == sd->event_map[end])
-                       break;
-
-       if (end < 0 || atom != sd->event_map[end]) {
-               printf("didn't find atom\n");
-               return;
-       }
-
-       //printf("end=%d, total %d\n", end, sd->event_map_idx);
+       last = atom;
+       io_u = atom->private;
+       atom = io_u->req.head;
 
        /*
         * now complete in right order
         */
-       for (i = 0; i <= end; i++) {
+       do {
                long ret;
 
-               atom = sd->event_map[i];
                io_u = atom->private;
                ret = *atom->ret_ptr;
                if (ret > 0)
@@ -65,19 +53,14 @@ static void fio_syslet_complete_atom(struct thread_data *td,
 
                assert(sd->nr_events < td->iodepth);
                sd->events[sd->nr_events++] = io_u;
-       }
 
-       /*
-        * Move later completions to the front, if we didn't complete all
-        */
-       if (end == (int) sd->event_map_idx - 1)
-               sd->event_map_idx = 0;
-       else {
-               int nr = sd->event_map_idx - end - 1;
+               if (atom == last)
+                       break;
 
-               memmove(sd->event_map, &sd->event_map[end + 1], nr * sizeof(struct syslet_uatom *));
-               sd->event_map_idx = nr;
-       }
+               atom = atom->next;
+       } while (1);
+
+       assert(!last->next);
 }
 
 /*
@@ -194,14 +177,31 @@ static int fio_syslet_prep(struct thread_data fio_unused *td, struct io_u *io_u)
 static void cachemiss_thread_start(void)
 {
        while (1)
-               async_thread();
+               async_thread(NULL, NULL);
 }
 
 #define THREAD_STACK_SIZE (16384)
 
 static unsigned long thread_stack_alloc()
 {
-       return (unsigned long)malloc(THREAD_STACK_SIZE) + THREAD_STACK_SIZE;
+       return (unsigned long) malloc(THREAD_STACK_SIZE) + THREAD_STACK_SIZE;
+}
+
+static void fio_syslet_queued(struct thread_data *td, struct syslet_data *sd)
+{
+       struct syslet_uatom *atom;
+       struct timeval now;
+
+       fio_gettime(&now, NULL);
+
+       atom = sd->head;
+       while (atom) {
+               struct io_u *io_u = atom->private;
+
+               memcpy(&io_u->issue_time, &now, sizeof(now));
+               io_u_queued(td, io_u);
+               atom = atom->next;
+       }
 }
 
 static int fio_syslet_commit(struct thread_data *td)
@@ -212,9 +212,13 @@ static int fio_syslet_commit(struct thread_data *td)
        if (!sd->head)
                return 0;
 
+       assert(!sd->tail->next);
+
        if (!sd->ahu.new_thread_stack)
                sd->ahu.new_thread_stack = thread_stack_alloc();
 
+       fio_syslet_queued(td, sd);
+
        /*
         * On sync completion, the atom is returned. So on NULL return
         * it's queued asynchronously.
@@ -239,7 +243,7 @@ static int fio_syslet_queue(struct thread_data *td, struct io_u *io_u)
        } else
                sd->head = sd->tail = &io_u->req.atom;
 
-       sd->event_map[sd->event_map_idx++] = sd->tail;
+       io_u->req.head = sd->head;
        return FIO_Q_QUEUED;
 }
 
@@ -257,8 +261,8 @@ static int async_head_init(struct syslet_data *sd, unsigned int depth)
        sd->ahu.completion_ring = sd->ring;
        sd->ahu.ring_size_bytes = ring_size;
        sd->ahu.head_stack = thread_stack_alloc();
-       sd->ahu.head_eip = (unsigned long)cachemiss_thread_start;
-       sd->ahu.new_thread_eip = (unsigned long)cachemiss_thread_start;
+       sd->ahu.head_eip = (unsigned long) cachemiss_thread_start;
+       sd->ahu.new_thread_eip = (unsigned long) cachemiss_thread_start;
 
        return 0;
 }
@@ -275,7 +279,6 @@ static void fio_syslet_cleanup(struct thread_data *td)
        if (sd) {
                async_head_exit(sd);
                free(sd->events);
-               free(sd->event_map);
                free(sd);
                td->io_ops->data = NULL;
        }
@@ -290,8 +293,6 @@ static int fio_syslet_init(struct thread_data *td)
        memset(sd, 0, sizeof(*sd));
        sd->events = malloc(sizeof(struct io_u *) * td->iodepth);
        memset(sd->events, 0, sizeof(struct io_u *) * td->iodepth);
-       sd->event_map = malloc(sizeof(struct syslet_uatom *) * td->iodepth);
-       memset(sd->event_map, 0, sizeof(struct syslet_uatom *) * td->iodepth);
 
        /*
         * This will handily fail for kernels where syslet isn't available