t/zbd: update test case 42
[fio.git] / flow.c
CommitLineData
9e684a49 1#include "fio.h"
971caeb1 2#include "fio_sem.h"
9e684a49
DE
3#include "smalloc.h"
4#include "flist.h"
5
6struct fio_flow {
7 unsigned int refs;
9e684a49 8 unsigned int id;
2b7968d3 9 struct flist_head list;
5200b0f8 10 unsigned long flow_counter;
d4e74fda 11 unsigned int total_weight;
9e684a49
DE
12};
13
14static struct flist_head *flow_list;
971caeb1 15static struct fio_sem *flow_lock;
9e684a49
DE
16
17int flow_threshold_exceeded(struct thread_data *td)
18{
19 struct fio_flow *flow = td->flow;
d4e74fda 20 double flow_counter_ratio, flow_weight_ratio;
9e684a49
DE
21
22 if (!flow)
23 return 0;
24
d4e74fda
DB
25 flow_counter_ratio = (double)td->flow_counter /
26 atomic_load_relaxed(&flow->flow_counter);
27 flow_weight_ratio = (double)td->o.flow /
28 atomic_load_relaxed(&flow->total_weight);
29
30 /*
31 * each thread/process executing a fio job will stall based on the
32 * expected user ratio for a given flow_id group. the idea is to keep
33 * 2 counters, flow and job-specific counter to test if the
34 * ratio between them is proportional to other jobs in the same flow_id
35 */
36 if (flow_counter_ratio > flow_weight_ratio) {
cb44aa1f
JA
37 if (td->o.flow_sleep) {
38 io_u_quiesce(td);
9e684a49 39 usleep(td->o.flow_sleep);
731365d8
AR
40 } else if (td->o.zone_mode == ZONE_MODE_ZBD) {
41 io_u_quiesce(td);
cb44aa1f
JA
42 }
43
9e684a49
DE
44 return 1;
45 }
46
d4e74fda
DB
47 /*
48 * increment flow(shared counter, therefore atomically)
49 * and job-specific counter
50 */
51 atomic_add(&flow->flow_counter, 1);
52 ++td->flow_counter;
53
9e684a49
DE
54 return 0;
55}
56
57static struct fio_flow *flow_get(unsigned int id)
58{
8e8b225d 59 struct fio_flow *flow = NULL;
9e684a49
DE
60 struct flist_head *n;
61
fba5c5ff
JA
62 if (!flow_lock)
63 return NULL;
64
971caeb1 65 fio_sem_down(flow_lock);
9e684a49
DE
66
67 flist_for_each(n, flow_list) {
68 flow = flist_entry(n, struct fio_flow, list);
69 if (flow->id == id)
70 break;
71
72 flow = NULL;
73 }
74
75 if (!flow) {
76 flow = smalloc(sizeof(*flow));
fba5c5ff 77 if (!flow) {
971caeb1 78 fio_sem_up(flow_lock);
fba5c5ff
JA
79 return NULL;
80 }
9e684a49
DE
81 flow->refs = 0;
82 INIT_FLIST_HEAD(&flow->list);
83 flow->id = id;
d4e74fda
DB
84 flow->flow_counter = 1;
85 flow->total_weight = 0;
9e684a49
DE
86
87 flist_add_tail(&flow->list, flow_list);
88 }
89
90 flow->refs++;
971caeb1 91 fio_sem_up(flow_lock);
9e684a49
DE
92 return flow;
93}
94
5200b0f8 95static void flow_put(struct fio_flow *flow, unsigned long flow_counter,
d4e74fda 96 unsigned int weight)
9e684a49 97{
fba5c5ff
JA
98 if (!flow_lock)
99 return;
100
971caeb1 101 fio_sem_down(flow_lock);
9e684a49 102
d4e74fda
DB
103 atomic_sub(&flow->flow_counter, flow_counter);
104 atomic_sub(&flow->total_weight, weight);
105
9e684a49 106 if (!--flow->refs) {
d4e74fda 107 assert(flow->flow_counter == 1);
9e684a49
DE
108 flist_del(&flow->list);
109 sfree(flow);
110 }
111
971caeb1 112 fio_sem_up(flow_lock);
9e684a49
DE
113}
114
115void flow_init_job(struct thread_data *td)
116{
d4e74fda 117 if (td->o.flow) {
9e684a49 118 td->flow = flow_get(td->o.flow_id);
d4e74fda
DB
119 td->flow_counter = 0;
120 atomic_add(&td->flow->total_weight, td->o.flow);
121 }
9e684a49
DE
122}
123
124void flow_exit_job(struct thread_data *td)
125{
126 if (td->flow) {
d4e74fda 127 flow_put(td->flow, td->flow_counter, td->o.flow);
9e684a49
DE
128 td->flow = NULL;
129 }
130}
131
132void flow_init(void)
133{
9e684a49 134 flow_list = smalloc(sizeof(*flow_list));
fba5c5ff
JA
135 if (!flow_list) {
136 log_err("fio: smalloc pool exhausted\n");
137 return;
138 }
139
971caeb1 140 flow_lock = fio_sem_init(FIO_SEM_UNLOCKED);
fba5c5ff
JA
141 if (!flow_lock) {
142 log_err("fio: failed to allocate flow lock\n");
143 sfree(flow_list);
144 return;
145 }
146
9e684a49
DE
147 INIT_FLIST_HEAD(flow_list);
148}
149
150void flow_exit(void)
151{
fba5c5ff 152 if (flow_lock)
971caeb1 153 fio_sem_remove(flow_lock);
fba5c5ff
JA
154 if (flow_list)
155 sfree(flow_list);
9e684a49 156}