}
void rate_throttle(struct thread_data *td, unsigned long time_spent,
- unsigned int bytes, int ddir)
+ unsigned int bytes)
{
unsigned long usec_cycle;
+ unsigned int bs;
if (!td->rate)
return;
- usec_cycle = td->rate_usec_cycle * (bytes / td->min_bs[ddir]);
+ if (td_rw(td))
+ bs = td->rw_min_bs;
+ else if (td_read(td))
+ bs = td->min_bs[DDIR_READ];
+ else
+ bs = td->min_bs[DDIR_WRITE];
+
+ usec_cycle = td->rate_usec_cycle * (bytes / bs);
if (time_spent < usec_cycle) {
unsigned long s = usec_cycle - time_spent;
{
int i;
- fio_gettime(&genesis, NULL);
-
/*
* Check the granularity of the nanosleep function
*/
}
}
+void set_genesis_time(void)
+{
+ fio_gettime(&genesis, NULL);
+}
+
void fill_start_time(struct timeval *t)
{
memcpy(t, &genesis, sizeof(genesis));