* including) this atom
*/
last = atom;
- io_u = atom->private;
+ io_u = (struct io_u *)atom->private;
atom = io_u->req.head;
/*
do {
long ret;
- io_u = atom->private;
- ret = *atom->ret_ptr;
+ io_u = (struct io_u *)atom->private;
+ ret = *(long *)atom->ret_ptr;
if (ret >= 0)
io_u->resid = io_u->xfer_buflen - ret;
else if (ret < 0)
if (atom == last)
break;
- atom = atom->next;
+ atom = (struct syslet_uatom *)atom->next;
} while (1);
assert(!last->next);
{
atom->flags = flags;
atom->nr = nr;
- atom->ret_ptr = ret_ptr;
- atom->next = NULL;
- atom->arg_ptr[0] = arg0;
- atom->arg_ptr[1] = arg1;
- atom->arg_ptr[2] = arg2;
- atom->arg_ptr[3] = arg3;
- atom->arg_ptr[4] = atom->arg_ptr[5] = NULL;
- atom->private = priv;
+ atom->ret_ptr = (uint64_t)ret_ptr;
+ atom->next = 0;
+ atom->arg_ptr[0] = (uint64_t)arg0;
+ atom->arg_ptr[1] = (uint64_t)arg1;
+ atom->arg_ptr[2] = (uint64_t)arg2;
+ atom->arg_ptr[3] = (uint64_t)arg3;
+ atom->arg_ptr[4] = 0;
+ atom->arg_ptr[5] = 0;
+ atom->private = (uint64_t)priv;
}
/*
atom = sd->head;
while (atom) {
- struct io_u *io_u = atom->private;
+ struct io_u *io_u = (struct io_u *)atom->private;
memcpy(&io_u->issue_time, &now, sizeof(now));
io_u_queued(td, io_u);
- atom = atom->next;
+ atom = (struct syslet_uatom *)atom->next;
}
}
fio_ro_check(td, io_u);
if (sd->tail) {
- sd->tail->next = &io_u->req.atom;
+ sd->tail->next = (uint64_t)&io_u->req.atom;
sd->tail = &io_u->req.atom;
} else
- sd->head = sd->tail = &io_u->req.atom;
+ sd->head = sd->tail = (struct syslet_uatom *)&io_u->req.atom;
io_u->req.head = sd->head;
return FIO_Q_QUEUED;
memset(sd->ring, 0, ring_size);
sd->ahu.user_ring_idx = 0;
- sd->ahu.completion_ring = sd->ring;
+ sd->ahu.completion_ring_ptr = (uint64_t)sd->ring;
sd->ahu.ring_size_bytes = ring_size;
sd->ahu.head_stack = thread_stack_alloc();
- sd->ahu.head_eip = (unsigned long) cachemiss_thread_start;
- sd->ahu.new_thread_eip = (unsigned long) cachemiss_thread_start;
+ sd->ahu.head_ip = (uint64_t)cachemiss_thread_start;
+ sd->ahu.new_thread_ip = (uint64_t)cachemiss_thread_start;
sd->ahu.new_thread_stack = thread_stack_alloc();
return 0;
* jump a full syslet_uatom number of bytes.) ]
*/
struct syslet_uatom {
- unsigned long flags;
- unsigned long nr;
- long __user *ret_ptr;
- struct syslet_uatom __user *next;
- unsigned long __user *arg_ptr[6];
+ uint32_t flags;
+ uint32_t nr;
+ uint64_t ret_ptr;
+ uint64_t next;
+ uint64_t arg_ptr[6];
/*
* User-space can put anything in here, kernel will not
* touch it:
*/
- void __user *private;
+ uint64_t private;
};
/*
/*
* Current completion ring index - managed by the kernel:
*/
- unsigned long kernel_ring_idx;
+ uint64_t kernel_ring_idx;
/*
* User-side ring index:
*/
- unsigned long user_ring_idx;
+ uint64_t user_ring_idx;
/*
* Ring of pointers to completed async syslets (i.e. syslets that
* Note: the final atom that generated the exit condition is
* queued here. Normally this would be the last atom of a syslet.
*/
- struct syslet_uatom __user **completion_ring;
+ uint64_t completion_ring_ptr;
/*
* Ring size in bytes:
*/
- unsigned long ring_size_bytes;
+ uint64_t ring_size_bytes;
/*
* The head task can become a cachemiss thread later on
* too, if it blocks - so it needs its separate thread
* stack and start address too:
*/
- unsigned long head_stack;
- unsigned long head_eip;
+ uint64_t head_stack;
+ uint64_t head_ip;
/*
* Newly started async kernel threads will take their
* code has to check for new_thread_stack going to NULL
* and has to refill it with a new stack if that happens.
*/
- unsigned long new_thread_stack;
- unsigned long new_thread_eip;
+ uint64_t new_thread_stack;
+ uint64_t new_thread_ip;
};
#endif