Commit | Line | Data |
---|---|---|
2874c5fd | 1 | // SPDX-License-Identifier: GPL-2.0-or-later |
96b2d69b | 2 | /* Processing of received RxRPC packets |
17926a79 | 3 | * |
96b2d69b | 4 | * Copyright (C) 2020 Red Hat, Inc. All Rights Reserved. |
17926a79 | 5 | * Written by David Howells (dhowells@redhat.com) |
17926a79 DH |
6 | */ |
7 | ||
9b6d5398 JP |
8 | #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt |
9 | ||
17926a79 DH |
10 | #include "ar-internal.h" |
11 | ||
57af281e DH |
12 | static void rxrpc_proto_abort(struct rxrpc_call *call, rxrpc_seq_t seq, |
13 | enum rxrpc_abort_reason why) | |
248f219c | 14 | { |
57af281e | 15 | rxrpc_abort_call(call, seq, RX_PROTOCOL_ERROR, -EBADMSG, why); |
248f219c DH |
16 | } |
17 | ||
57494343 DH |
18 | /* |
19 | * Do TCP-style congestion management [RFC 5681]. | |
20 | */ | |
21 | static void rxrpc_congestion_management(struct rxrpc_call *call, | |
22 | struct sk_buff *skb, | |
ed1e8679 DH |
23 | struct rxrpc_ack_summary *summary, |
24 | rxrpc_serial_t acked_serial) | |
57494343 DH |
25 | { |
26 | enum rxrpc_congest_change change = rxrpc_cong_no_change; | |
57494343 DH |
27 | unsigned int cumulative_acks = call->cong_cumul_acks; |
28 | unsigned int cwnd = call->cong_cwnd; | |
29 | bool resend = false; | |
30 | ||
31 | summary->flight_size = | |
a4ea4c47 | 32 | (call->tx_top - call->acks_hard_ack) - summary->nr_acks; |
57494343 DH |
33 | |
34 | if (test_and_clear_bit(RXRPC_CALL_RETRANS_TIMEOUT, &call->flags)) { | |
35 | summary->retrans_timeo = true; | |
36 | call->cong_ssthresh = max_t(unsigned int, | |
37 | summary->flight_size / 2, 2); | |
38 | cwnd = 1; | |
8782def2 | 39 | if (cwnd >= call->cong_ssthresh && |
57494343 DH |
40 | call->cong_mode == RXRPC_CALL_SLOW_START) { |
41 | call->cong_mode = RXRPC_CALL_CONGEST_AVOIDANCE; | |
42 | call->cong_tstamp = skb->tstamp; | |
43 | cumulative_acks = 0; | |
44 | } | |
45 | } | |
46 | ||
47 | cumulative_acks += summary->nr_new_acks; | |
48 | cumulative_acks += summary->nr_rot_new_acks; | |
49 | if (cumulative_acks > 255) | |
50 | cumulative_acks = 255; | |
51 | ||
52 | summary->mode = call->cong_mode; | |
53 | summary->cwnd = call->cong_cwnd; | |
54 | summary->ssthresh = call->cong_ssthresh; | |
55 | summary->cumulative_acks = cumulative_acks; | |
56 | summary->dup_acks = call->cong_dup_acks; | |
57 | ||
58 | switch (call->cong_mode) { | |
59 | case RXRPC_CALL_SLOW_START: | |
d57a3a15 | 60 | if (summary->saw_nacks) |
57494343 DH |
61 | goto packet_loss_detected; |
62 | if (summary->cumulative_acks > 0) | |
63 | cwnd += 1; | |
8782def2 | 64 | if (cwnd >= call->cong_ssthresh) { |
57494343 DH |
65 | call->cong_mode = RXRPC_CALL_CONGEST_AVOIDANCE; |
66 | call->cong_tstamp = skb->tstamp; | |
67 | } | |
68 | goto out; | |
69 | ||
70 | case RXRPC_CALL_CONGEST_AVOIDANCE: | |
d57a3a15 | 71 | if (summary->saw_nacks) |
57494343 DH |
72 | goto packet_loss_detected; |
73 | ||
74 | /* We analyse the number of packets that get ACK'd per RTT | |
75 | * period and increase the window if we managed to fill it. | |
76 | */ | |
c410bf01 | 77 | if (call->peer->rtt_count == 0) |
57494343 DH |
78 | goto out; |
79 | if (ktime_before(skb->tstamp, | |
c410bf01 DH |
80 | ktime_add_us(call->cong_tstamp, |
81 | call->peer->srtt_us >> 3))) | |
57494343 DH |
82 | goto out_no_clear_ca; |
83 | change = rxrpc_cong_rtt_window_end; | |
84 | call->cong_tstamp = skb->tstamp; | |
85 | if (cumulative_acks >= cwnd) | |
86 | cwnd++; | |
87 | goto out; | |
88 | ||
89 | case RXRPC_CALL_PACKET_LOSS: | |
d57a3a15 | 90 | if (!summary->saw_nacks) |
57494343 DH |
91 | goto resume_normality; |
92 | ||
93 | if (summary->new_low_nack) { | |
94 | change = rxrpc_cong_new_low_nack; | |
95 | call->cong_dup_acks = 1; | |
96 | if (call->cong_extra > 1) | |
97 | call->cong_extra = 1; | |
98 | goto send_extra_data; | |
99 | } | |
100 | ||
101 | call->cong_dup_acks++; | |
102 | if (call->cong_dup_acks < 3) | |
103 | goto send_extra_data; | |
104 | ||
105 | change = rxrpc_cong_begin_retransmission; | |
106 | call->cong_mode = RXRPC_CALL_FAST_RETRANSMIT; | |
107 | call->cong_ssthresh = max_t(unsigned int, | |
108 | summary->flight_size / 2, 2); | |
109 | cwnd = call->cong_ssthresh + 3; | |
110 | call->cong_extra = 0; | |
111 | call->cong_dup_acks = 0; | |
112 | resend = true; | |
113 | goto out; | |
114 | ||
115 | case RXRPC_CALL_FAST_RETRANSMIT: | |
116 | if (!summary->new_low_nack) { | |
117 | if (summary->nr_new_acks == 0) | |
118 | cwnd += 1; | |
119 | call->cong_dup_acks++; | |
120 | if (call->cong_dup_acks == 2) { | |
121 | change = rxrpc_cong_retransmit_again; | |
122 | call->cong_dup_acks = 0; | |
123 | resend = true; | |
124 | } | |
125 | } else { | |
126 | change = rxrpc_cong_progress; | |
127 | cwnd = call->cong_ssthresh; | |
d57a3a15 | 128 | if (!summary->saw_nacks) |
57494343 DH |
129 | goto resume_normality; |
130 | } | |
131 | goto out; | |
132 | ||
133 | default: | |
134 | BUG(); | |
135 | goto out; | |
136 | } | |
137 | ||
138 | resume_normality: | |
139 | change = rxrpc_cong_cleared_nacks; | |
140 | call->cong_dup_acks = 0; | |
141 | call->cong_extra = 0; | |
142 | call->cong_tstamp = skb->tstamp; | |
8782def2 | 143 | if (cwnd < call->cong_ssthresh) |
57494343 DH |
144 | call->cong_mode = RXRPC_CALL_SLOW_START; |
145 | else | |
146 | call->cong_mode = RXRPC_CALL_CONGEST_AVOIDANCE; | |
147 | out: | |
148 | cumulative_acks = 0; | |
149 | out_no_clear_ca: | |
a4ea4c47 DH |
150 | if (cwnd >= RXRPC_TX_MAX_WINDOW) |
151 | cwnd = RXRPC_TX_MAX_WINDOW; | |
57494343 DH |
152 | call->cong_cwnd = cwnd; |
153 | call->cong_cumul_acks = cumulative_acks; | |
ed1e8679 | 154 | trace_rxrpc_congest(call, summary, acked_serial, change); |
5e6ef4f1 DH |
155 | if (resend) |
156 | rxrpc_resend(call, skb); | |
57494343 DH |
157 | return; |
158 | ||
159 | packet_loss_detected: | |
160 | change = rxrpc_cong_saw_nack; | |
161 | call->cong_mode = RXRPC_CALL_PACKET_LOSS; | |
162 | call->cong_dup_acks = 0; | |
163 | goto send_extra_data; | |
164 | ||
165 | send_extra_data: | |
166 | /* Send some previously unsent DATA if we have some to advance the ACK | |
167 | * state. | |
168 | */ | |
a4ea4c47 DH |
169 | if (test_bit(RXRPC_CALL_TX_LAST, &call->flags) || |
170 | summary->nr_acks != call->tx_top - call->acks_hard_ack) { | |
57494343 DH |
171 | call->cong_extra++; |
172 | wake_up(&call->waitq); | |
173 | } | |
174 | goto out_no_clear_ca; | |
175 | } | |
176 | ||
5086d9a9 DH |
177 | /* |
178 | * Degrade the congestion window if we haven't transmitted a packet for >1RTT. | |
179 | */ | |
180 | void rxrpc_congestion_degrade(struct rxrpc_call *call) | |
181 | { | |
182 | ktime_t rtt, now; | |
183 | ||
184 | if (call->cong_mode != RXRPC_CALL_SLOW_START && | |
185 | call->cong_mode != RXRPC_CALL_CONGEST_AVOIDANCE) | |
186 | return; | |
187 | if (call->state == RXRPC_CALL_CLIENT_AWAIT_REPLY) | |
188 | return; | |
189 | ||
190 | rtt = ns_to_ktime(call->peer->srtt_us * (1000 / 8)); | |
191 | now = ktime_get_real(); | |
192 | if (!ktime_before(ktime_add(call->tx_last_sent, rtt), now)) | |
193 | return; | |
194 | ||
195 | trace_rxrpc_reset_cwnd(call, now); | |
196 | rxrpc_inc_stat(call->rxnet, stat_tx_data_cwnd_reset); | |
197 | call->tx_last_sent = now; | |
198 | call->cong_mode = RXRPC_CALL_SLOW_START; | |
199 | call->cong_ssthresh = max_t(unsigned int, call->cong_ssthresh, | |
200 | call->cong_cwnd * 3 / 4); | |
201 | call->cong_cwnd = max_t(unsigned int, call->cong_cwnd / 2, RXRPC_MIN_CWND); | |
202 | } | |
203 | ||
17926a79 | 204 | /* |
248f219c | 205 | * Apply a hard ACK by advancing the Tx window. |
17926a79 | 206 | */ |
c479d5f2 | 207 | static bool rxrpc_rotate_tx_window(struct rxrpc_call *call, rxrpc_seq_t to, |
31a1b989 | 208 | struct rxrpc_ack_summary *summary) |
17926a79 | 209 | { |
a4ea4c47 | 210 | struct rxrpc_txbuf *txb; |
c479d5f2 | 211 | bool rot_last = false; |
17926a79 | 212 | |
a4ea4c47 DH |
213 | list_for_each_entry_rcu(txb, &call->tx_buffer, call_link, false) { |
214 | if (before_eq(txb->seq, call->acks_hard_ack)) | |
215 | continue; | |
d57a3a15 | 216 | summary->nr_rot_new_acks++; |
a4ea4c47 | 217 | if (test_bit(RXRPC_TXBUF_LAST, &txb->flags)) { |
70790dbe | 218 | set_bit(RXRPC_CALL_TX_LAST, &call->flags); |
c479d5f2 DH |
219 | rot_last = true; |
220 | } | |
a4ea4c47 DH |
221 | if (txb->seq == to) |
222 | break; | |
248f219c | 223 | } |
17926a79 | 224 | |
a4ea4c47 DH |
225 | if (rot_last) |
226 | set_bit(RXRPC_CALL_TX_ALL_ACKED, &call->flags); | |
17926a79 | 227 | |
a4ea4c47 | 228 | _enter("%x,%x,%x,%d", to, call->acks_hard_ack, call->tx_top, rot_last); |
bc4abfcf | 229 | |
a4ea4c47 DH |
230 | if (call->acks_lowest_nak == call->acks_hard_ack) { |
231 | call->acks_lowest_nak = to; | |
1fc4fa2a | 232 | } else if (after(to, call->acks_lowest_nak)) { |
a4ea4c47 DH |
233 | summary->new_low_nack = true; |
234 | call->acks_lowest_nak = to; | |
17926a79 | 235 | } |
c479d5f2 | 236 | |
a4ea4c47 DH |
237 | smp_store_release(&call->acks_hard_ack, to); |
238 | ||
239 | trace_rxrpc_txqueue(call, (rot_last ? | |
240 | rxrpc_txqueue_rotate_last : | |
241 | rxrpc_txqueue_rotate)); | |
242 | wake_up(&call->waitq); | |
c479d5f2 | 243 | return rot_last; |
248f219c | 244 | } |
17926a79 | 245 | |
248f219c DH |
246 | /* |
247 | * End the transmission phase of a call. | |
248 | * | |
249 | * This occurs when we get an ACKALL packet, the first DATA packet of a reply, | |
250 | * or a final ACK packet. | |
251 | */ | |
57af281e DH |
252 | static void rxrpc_end_tx_phase(struct rxrpc_call *call, bool reply_begun, |
253 | enum rxrpc_abort_reason abort_why) | |
248f219c | 254 | { |
dfe99522 | 255 | unsigned int state; |
17926a79 | 256 | |
70790dbe | 257 | ASSERT(test_bit(RXRPC_CALL_TX_LAST, &call->flags)); |
17926a79 | 258 | |
248f219c | 259 | write_lock(&call->state_lock); |
651350d1 | 260 | |
dfe99522 DH |
261 | state = call->state; |
262 | switch (state) { | |
70790dbe | 263 | case RXRPC_CALL_CLIENT_SEND_REQUEST: |
248f219c | 264 | case RXRPC_CALL_CLIENT_AWAIT_REPLY: |
70790dbe | 265 | if (reply_begun) |
dfe99522 | 266 | call->state = state = RXRPC_CALL_CLIENT_RECV_REPLY; |
70790dbe | 267 | else |
dfe99522 | 268 | call->state = state = RXRPC_CALL_CLIENT_AWAIT_REPLY; |
248f219c | 269 | break; |
70790dbe | 270 | |
248f219c DH |
271 | case RXRPC_CALL_SERVER_AWAIT_ACK: |
272 | __rxrpc_call_completed(call); | |
dfe99522 | 273 | state = call->state; |
248f219c | 274 | break; |
70790dbe DH |
275 | |
276 | default: | |
277 | goto bad_state; | |
17926a79 | 278 | } |
17926a79 | 279 | |
248f219c | 280 | write_unlock(&call->state_lock); |
dfe99522 | 281 | if (state == RXRPC_CALL_CLIENT_AWAIT_REPLY) |
a4ea4c47 | 282 | trace_rxrpc_txqueue(call, rxrpc_txqueue_await_reply); |
dfe99522 | 283 | else |
a4ea4c47 | 284 | trace_rxrpc_txqueue(call, rxrpc_txqueue_end); |
248f219c | 285 | _leave(" = ok"); |
57af281e | 286 | return; |
70790dbe DH |
287 | |
288 | bad_state: | |
289 | write_unlock(&call->state_lock); | |
290 | kdebug("end_tx %s", rxrpc_call_states[call->state]); | |
57af281e | 291 | rxrpc_proto_abort(call, call->tx_top, abort_why); |
70790dbe DH |
292 | } |
293 | ||
294 | /* | |
295 | * Begin the reply reception phase of a call. | |
296 | */ | |
297 | static bool rxrpc_receiving_reply(struct rxrpc_call *call) | |
298 | { | |
31a1b989 | 299 | struct rxrpc_ack_summary summary = { 0 }; |
a158bdd3 | 300 | unsigned long now, timo; |
70790dbe DH |
301 | rxrpc_seq_t top = READ_ONCE(call->tx_top); |
302 | ||
dd7c1ee5 | 303 | if (call->ackr_reason) { |
a158bdd3 DH |
304 | now = jiffies; |
305 | timo = now + MAX_JIFFY_OFFSET; | |
306 | WRITE_ONCE(call->resend_at, timo); | |
530403d9 | 307 | WRITE_ONCE(call->delay_ack_at, timo); |
a158bdd3 | 308 | trace_rxrpc_timer(call, rxrpc_timer_init_for_reply, now); |
dd7c1ee5 DH |
309 | } |
310 | ||
70790dbe | 311 | if (!test_bit(RXRPC_CALL_TX_LAST, &call->flags)) { |
c479d5f2 | 312 | if (!rxrpc_rotate_tx_window(call, top, &summary)) { |
57af281e | 313 | rxrpc_proto_abort(call, top, rxrpc_eproto_early_reply); |
c479d5f2 DH |
314 | return false; |
315 | } | |
70790dbe | 316 | } |
57af281e DH |
317 | |
318 | rxrpc_end_tx_phase(call, true, rxrpc_eproto_unexpected_reply); | |
319 | return true; | |
248f219c DH |
320 | } |
321 | ||
93368b6b DH |
322 | /* |
323 | * End the packet reception phase. | |
324 | */ | |
325 | static void rxrpc_end_rx_phase(struct rxrpc_call *call, rxrpc_serial_t serial) | |
326 | { | |
327 | rxrpc_seq_t whigh = READ_ONCE(call->rx_highest_seq); | |
328 | ||
329 | _enter("%d,%s", call->debug_id, rxrpc_call_states[call->state]); | |
330 | ||
331 | trace_rxrpc_receive(call, rxrpc_receive_end, 0, whigh); | |
332 | ||
333 | if (rxrpc_call_state(call) == RXRPC_CALL_CLIENT_RECV_REPLY) | |
334 | rxrpc_propose_delay_ACK(call, serial, rxrpc_propose_ack_terminal_ack); | |
335 | ||
336 | write_lock(&call->state_lock); | |
337 | ||
338 | switch (call->state) { | |
339 | case RXRPC_CALL_CLIENT_RECV_REPLY: | |
340 | __rxrpc_call_completed(call); | |
341 | write_unlock(&call->state_lock); | |
342 | break; | |
343 | ||
344 | case RXRPC_CALL_SERVER_RECV_REQUEST: | |
345 | call->state = RXRPC_CALL_SERVER_ACK_REQUEST; | |
346 | call->expect_req_by = jiffies + MAX_JIFFY_OFFSET; | |
347 | write_unlock(&call->state_lock); | |
348 | rxrpc_propose_delay_ACK(call, serial, | |
349 | rxrpc_propose_ack_processing_op); | |
350 | break; | |
351 | default: | |
352 | write_unlock(&call->state_lock); | |
353 | break; | |
354 | } | |
355 | } | |
356 | ||
5d7edbc9 DH |
357 | static void rxrpc_input_update_ack_window(struct rxrpc_call *call, |
358 | rxrpc_seq_t window, rxrpc_seq_t wtop) | |
359 | { | |
360 | atomic64_set_release(&call->ackr_window, ((u64)wtop) << 32 | window); | |
361 | } | |
362 | ||
248f219c | 363 | /* |
5d7edbc9 DH |
364 | * Push a DATA packet onto the Rx queue. |
365 | */ | |
366 | static void rxrpc_input_queue_data(struct rxrpc_call *call, struct sk_buff *skb, | |
367 | rxrpc_seq_t window, rxrpc_seq_t wtop, | |
368 | enum rxrpc_receive_trace why) | |
369 | { | |
370 | struct rxrpc_skb_priv *sp = rxrpc_skb(skb); | |
371 | bool last = sp->hdr.flags & RXRPC_LAST_PACKET; | |
372 | ||
373 | __skb_queue_tail(&call->recvmsg_queue, skb); | |
374 | rxrpc_input_update_ack_window(call, window, wtop); | |
5d7edbc9 | 375 | trace_rxrpc_receive(call, last ? why + 1 : why, sp->hdr.serial, sp->hdr.seq); |
93368b6b DH |
376 | if (last) |
377 | rxrpc_end_rx_phase(call, sp->hdr.serial); | |
5d7edbc9 DH |
378 | } |
379 | ||
380 | /* | |
381 | * Process a DATA packet. | |
248f219c | 382 | */ |
2d1faf7a DH |
383 | static void rxrpc_input_data_one(struct rxrpc_call *call, struct sk_buff *skb, |
384 | bool *_notify) | |
248f219c DH |
385 | { |
386 | struct rxrpc_skb_priv *sp = rxrpc_skb(skb); | |
5d7edbc9 | 387 | struct sk_buff *oos; |
d4d02d8b | 388 | rxrpc_serial_t serial = sp->hdr.serial; |
5d7edbc9 DH |
389 | u64 win = atomic64_read(&call->ackr_window); |
390 | rxrpc_seq_t window = lower_32_bits(win); | |
391 | rxrpc_seq_t wtop = upper_32_bits(win); | |
392 | rxrpc_seq_t wlimit = window + call->rx_winsize - 1; | |
393 | rxrpc_seq_t seq = sp->hdr.seq; | |
d4d02d8b | 394 | bool last = sp->hdr.flags & RXRPC_LAST_PACKET; |
5d7edbc9 | 395 | int ack_reason = -1; |
248f219c | 396 | |
d4d02d8b DH |
397 | rxrpc_inc_stat(call->rxnet, stat_rx_data); |
398 | if (sp->hdr.flags & RXRPC_REQUEST_ACK) | |
399 | rxrpc_inc_stat(call->rxnet, stat_rx_data_reqack); | |
400 | if (sp->hdr.flags & RXRPC_JUMBO_PACKET) | |
401 | rxrpc_inc_stat(call->rxnet, stat_rx_data_jumbo); | |
c3c9e3df | 402 | |
d4d02d8b | 403 | if (last) { |
5d7edbc9 | 404 | if (test_and_set_bit(RXRPC_CALL_RX_LAST, &call->flags) && |
57af281e DH |
405 | seq + 1 != wtop) |
406 | return rxrpc_proto_abort(call, seq, rxrpc_eproto_different_last); | |
d4d02d8b DH |
407 | } else { |
408 | if (test_bit(RXRPC_CALL_RX_LAST, &call->flags) && | |
5d7edbc9 DH |
409 | after_eq(seq, wtop)) { |
410 | pr_warn("Packet beyond last: c=%x q=%x window=%x-%x wlimit=%x\n", | |
411 | call->debug_id, seq, window, wtop, wlimit); | |
57af281e | 412 | return rxrpc_proto_abort(call, seq, rxrpc_eproto_data_after_last); |
d4d02d8b | 413 | } |
c3c9e3df | 414 | } |
248f219c | 415 | |
5d7edbc9 DH |
416 | if (after(seq, call->rx_highest_seq)) |
417 | call->rx_highest_seq = seq; | |
418 | ||
d4d02d8b | 419 | trace_rxrpc_rx_data(call->debug_id, seq, serial, sp->hdr.flags); |
17926a79 | 420 | |
5d7edbc9 DH |
421 | if (before(seq, window)) { |
422 | ack_reason = RXRPC_ACK_DUPLICATE; | |
423 | goto send_ack; | |
d4d02d8b | 424 | } |
5d7edbc9 DH |
425 | if (after(seq, wlimit)) { |
426 | ack_reason = RXRPC_ACK_EXCEEDS_WINDOW; | |
427 | goto send_ack; | |
d4d02d8b DH |
428 | } |
429 | ||
5d7edbc9 DH |
430 | /* Queue the packet. */ |
431 | if (seq == window) { | |
432 | rxrpc_seq_t reset_from; | |
433 | bool reset_sack = false; | |
d4d02d8b | 434 | |
5d7edbc9 DH |
435 | if (sp->hdr.flags & RXRPC_REQUEST_ACK) |
436 | ack_reason = RXRPC_ACK_REQUESTED; | |
437 | /* Send an immediate ACK if we fill in a hole */ | |
438 | else if (!skb_queue_empty(&call->rx_oos_queue)) | |
439 | ack_reason = RXRPC_ACK_DELAY; | |
5e6ef4f1 DH |
440 | else |
441 | atomic_inc_return(&call->ackr_nr_unacked); | |
d4d02d8b | 442 | |
5d7edbc9 DH |
443 | window++; |
444 | if (after(window, wtop)) | |
445 | wtop = window; | |
d4d02d8b | 446 | |
2d1faf7a DH |
447 | rxrpc_get_skb(skb, rxrpc_skb_get_to_recvmsg); |
448 | ||
5d7edbc9 DH |
449 | spin_lock(&call->recvmsg_queue.lock); |
450 | rxrpc_input_queue_data(call, skb, window, wtop, rxrpc_receive_queue); | |
2d1faf7a | 451 | *_notify = true; |
5d7edbc9 DH |
452 | |
453 | while ((oos = skb_peek(&call->rx_oos_queue))) { | |
454 | struct rxrpc_skb_priv *osp = rxrpc_skb(oos); | |
455 | ||
456 | if (after(osp->hdr.seq, window)) | |
457 | break; | |
458 | ||
459 | __skb_unlink(oos, &call->rx_oos_queue); | |
460 | last = osp->hdr.flags & RXRPC_LAST_PACKET; | |
461 | seq = osp->hdr.seq; | |
462 | if (!reset_sack) { | |
463 | reset_from = seq; | |
464 | reset_sack = true; | |
465 | } | |
466 | ||
467 | window++; | |
468 | rxrpc_input_queue_data(call, oos, window, wtop, | |
469 | rxrpc_receive_queue_oos); | |
d4d02d8b | 470 | } |
d4d02d8b | 471 | |
5d7edbc9 | 472 | spin_unlock(&call->recvmsg_queue.lock); |
d4d02d8b | 473 | |
5d7edbc9 DH |
474 | if (reset_sack) { |
475 | do { | |
476 | call->ackr_sack_table[reset_from % RXRPC_SACK_SIZE] = 0; | |
477 | } while (reset_from++, before(reset_from, window)); | |
478 | } | |
d4d02d8b | 479 | } else { |
5d7edbc9 | 480 | bool keep = false; |
d4d02d8b | 481 | |
5d7edbc9 DH |
482 | ack_reason = RXRPC_ACK_OUT_OF_SEQUENCE; |
483 | ||
484 | if (!call->ackr_sack_table[seq % RXRPC_SACK_SIZE]) { | |
485 | call->ackr_sack_table[seq % RXRPC_SACK_SIZE] = 1; | |
486 | keep = 1; | |
487 | } | |
488 | ||
489 | if (after(seq + 1, wtop)) { | |
490 | wtop = seq + 1; | |
491 | rxrpc_input_update_ack_window(call, window, wtop); | |
492 | } | |
493 | ||
494 | if (!keep) { | |
495 | ack_reason = RXRPC_ACK_DUPLICATE; | |
496 | goto send_ack; | |
497 | } | |
498 | ||
499 | skb_queue_walk(&call->rx_oos_queue, oos) { | |
500 | struct rxrpc_skb_priv *osp = rxrpc_skb(oos); | |
501 | ||
502 | if (after(osp->hdr.seq, seq)) { | |
2d1faf7a | 503 | rxrpc_get_skb(skb, rxrpc_skb_get_to_recvmsg_oos); |
5d7edbc9 DH |
504 | __skb_queue_before(&call->rx_oos_queue, oos, skb); |
505 | goto oos_queued; | |
506 | } | |
d4d02d8b | 507 | } |
5d7edbc9 | 508 | |
2d1faf7a | 509 | rxrpc_get_skb(skb, rxrpc_skb_get_to_recvmsg_oos); |
5d7edbc9 DH |
510 | __skb_queue_tail(&call->rx_oos_queue, skb); |
511 | oos_queued: | |
512 | trace_rxrpc_receive(call, last ? rxrpc_receive_oos_last : rxrpc_receive_oos, | |
513 | sp->hdr.serial, sp->hdr.seq); | |
d4d02d8b DH |
514 | } |
515 | ||
5d7edbc9 | 516 | send_ack: |
5d7edbc9 DH |
517 | if (ack_reason >= 0) |
518 | rxrpc_send_ACK(call, ack_reason, serial, | |
d4d02d8b DH |
519 | rxrpc_propose_ack_input_data); |
520 | else | |
521 | rxrpc_propose_delay_ACK(call, serial, | |
522 | rxrpc_propose_ack_input_data); | |
17926a79 DH |
523 | } |
524 | ||
525 | /* | |
d4d02d8b | 526 | * Split a jumbo packet and file the bits separately. |
17926a79 | 527 | */ |
d4d02d8b | 528 | static bool rxrpc_input_split_jumbo(struct rxrpc_call *call, struct sk_buff *skb) |
17926a79 | 529 | { |
d4d02d8b DH |
530 | struct rxrpc_jumbo_header jhdr; |
531 | struct rxrpc_skb_priv *sp = rxrpc_skb(skb), *jsp; | |
532 | struct sk_buff *jskb; | |
533 | unsigned int offset = sizeof(struct rxrpc_wire_header); | |
534 | unsigned int len = skb->len - offset; | |
2d1faf7a | 535 | bool notify = false; |
17926a79 | 536 | |
d4d02d8b DH |
537 | while (sp->hdr.flags & RXRPC_JUMBO_PACKET) { |
538 | if (len < RXRPC_JUMBO_SUBPKTLEN) | |
539 | goto protocol_error; | |
540 | if (sp->hdr.flags & RXRPC_LAST_PACKET) | |
541 | goto protocol_error; | |
542 | if (skb_copy_bits(skb, offset + RXRPC_JUMBO_DATALEN, | |
543 | &jhdr, sizeof(jhdr)) < 0) | |
544 | goto protocol_error; | |
545 | ||
5e6ef4f1 | 546 | jskb = skb_clone(skb, GFP_NOFS); |
d4d02d8b DH |
547 | if (!jskb) { |
548 | kdebug("couldn't clone"); | |
549 | return false; | |
550 | } | |
9a36a6bc | 551 | rxrpc_new_skb(jskb, rxrpc_skb_new_jumbo_subpacket); |
d4d02d8b DH |
552 | jsp = rxrpc_skb(jskb); |
553 | jsp->offset = offset; | |
554 | jsp->len = RXRPC_JUMBO_DATALEN; | |
2d1faf7a DH |
555 | rxrpc_input_data_one(call, jskb, ¬ify); |
556 | rxrpc_free_skb(jskb, rxrpc_skb_put_jumbo_subpacket); | |
d4d02d8b DH |
557 | |
558 | sp->hdr.flags = jhdr.flags; | |
559 | sp->hdr._rsvd = ntohs(jhdr._rsvd); | |
560 | sp->hdr.seq++; | |
561 | sp->hdr.serial++; | |
562 | offset += RXRPC_JUMBO_SUBPKTLEN; | |
563 | len -= RXRPC_JUMBO_SUBPKTLEN; | |
248f219c | 564 | } |
d4d02d8b DH |
565 | |
566 | sp->offset = offset; | |
567 | sp->len = len; | |
2d1faf7a DH |
568 | rxrpc_input_data_one(call, skb, ¬ify); |
569 | if (notify) { | |
570 | trace_rxrpc_notify_socket(call->debug_id, sp->hdr.serial); | |
571 | rxrpc_notify_socket(call); | |
572 | } | |
d4d02d8b DH |
573 | return true; |
574 | ||
575 | protocol_error: | |
576 | return false; | |
248f219c | 577 | } |
17926a79 | 578 | |
248f219c | 579 | /* |
4858e403 DH |
580 | * Process a DATA packet, adding the packet to the Rx ring. The caller's |
581 | * packet ref must be passed on or discarded. | |
248f219c | 582 | */ |
e8c3af6b | 583 | static void rxrpc_input_data(struct rxrpc_call *call, struct sk_buff *skb) |
248f219c DH |
584 | { |
585 | struct rxrpc_skb_priv *sp = rxrpc_skb(skb); | |
146d8fef | 586 | enum rxrpc_call_state state; |
d4d02d8b DH |
587 | rxrpc_serial_t serial = sp->hdr.serial; |
588 | rxrpc_seq_t seq0 = sp->hdr.seq; | |
17926a79 | 589 | |
5d7edbc9 DH |
590 | _enter("{%llx,%x},{%u,%x}", |
591 | atomic64_read(&call->ackr_window), call->rx_highest_seq, | |
592 | skb->len, seq0); | |
17926a79 | 593 | |
146d8fef | 594 | state = READ_ONCE(call->state); |
2d1faf7a | 595 | if (state >= RXRPC_CALL_COMPLETE) |
248f219c | 596 | return; |
17926a79 | 597 | |
a95d25dd | 598 | if (state == RXRPC_CALL_SERVER_RECV_REQUEST) { |
a158bdd3 DH |
599 | unsigned long timo = READ_ONCE(call->next_req_timo); |
600 | unsigned long now, expect_req_by; | |
601 | ||
602 | if (timo) { | |
603 | now = jiffies; | |
604 | expect_req_by = now + timo; | |
605 | WRITE_ONCE(call->expect_req_by, expect_req_by); | |
606 | rxrpc_reduce_call_timer(call, expect_req_by, now, | |
607 | rxrpc_timer_set_for_idle); | |
608 | } | |
609 | } | |
610 | ||
248f219c DH |
611 | /* Received data implicitly ACKs all of the request packets we sent |
612 | * when we're acting as a client. | |
613 | */ | |
146d8fef DH |
614 | if ((state == RXRPC_CALL_CLIENT_SEND_REQUEST || |
615 | state == RXRPC_CALL_CLIENT_AWAIT_REPLY) && | |
70790dbe | 616 | !rxrpc_receiving_reply(call)) |
5e6ef4f1 | 617 | goto out_notify; |
72f0c6fb | 618 | |
d4d02d8b | 619 | if (!rxrpc_input_split_jumbo(call, skb)) { |
57af281e | 620 | rxrpc_proto_abort(call, sp->hdr.seq, rxrpc_badmsg_bad_jumbo); |
5e6ef4f1 | 621 | goto out_notify; |
248f219c | 622 | } |
d4d02d8b | 623 | skb = NULL; |
17926a79 | 624 | |
5e6ef4f1 | 625 | out_notify: |
f71dbf2f DH |
626 | trace_rxrpc_notify_socket(call->debug_id, serial); |
627 | rxrpc_notify_socket(call); | |
248f219c | 628 | _leave(" [queued]"); |
17926a79 DH |
629 | } |
630 | ||
50235c4b | 631 | /* |
4700c4d8 | 632 | * See if there's a cached RTT probe to complete. |
50235c4b | 633 | */ |
4700c4d8 DH |
634 | static void rxrpc_complete_rtt_probe(struct rxrpc_call *call, |
635 | ktime_t resp_time, | |
636 | rxrpc_serial_t acked_serial, | |
637 | rxrpc_serial_t ack_serial, | |
638 | enum rxrpc_rtt_rx_trace type) | |
50235c4b | 639 | { |
4700c4d8 DH |
640 | rxrpc_serial_t orig_serial; |
641 | unsigned long avail; | |
50235c4b | 642 | ktime_t sent_at; |
4700c4d8 DH |
643 | bool matched = false; |
644 | int i; | |
50235c4b | 645 | |
4700c4d8 DH |
646 | avail = READ_ONCE(call->rtt_avail); |
647 | smp_rmb(); /* Read avail bits before accessing data. */ | |
50235c4b | 648 | |
4700c4d8 DH |
649 | for (i = 0; i < ARRAY_SIZE(call->rtt_serial); i++) { |
650 | if (!test_bit(i + RXRPC_CALL_RTT_PEND_SHIFT, &avail)) | |
50235c4b | 651 | continue; |
b604dd98 | 652 | |
4700c4d8 DH |
653 | sent_at = call->rtt_sent_at[i]; |
654 | orig_serial = call->rtt_serial[i]; | |
655 | ||
656 | if (orig_serial == acked_serial) { | |
657 | clear_bit(i + RXRPC_CALL_RTT_PEND_SHIFT, &call->rtt_avail); | |
658 | smp_mb(); /* Read data before setting avail bit */ | |
659 | set_bit(i, &call->rtt_avail); | |
660 | if (type != rxrpc_rtt_rx_cancel) | |
661 | rxrpc_peer_add_rtt(call, type, i, acked_serial, ack_serial, | |
662 | sent_at, resp_time); | |
663 | else | |
664 | trace_rxrpc_rtt_rx(call, rxrpc_rtt_rx_cancel, i, | |
665 | orig_serial, acked_serial, 0, 0); | |
666 | matched = true; | |
667 | } | |
668 | ||
669 | /* If a later serial is being acked, then mark this slot as | |
670 | * being available. | |
671 | */ | |
672 | if (after(acked_serial, orig_serial)) { | |
673 | trace_rxrpc_rtt_rx(call, rxrpc_rtt_rx_obsolete, i, | |
674 | orig_serial, acked_serial, 0, 0); | |
675 | clear_bit(i + RXRPC_CALL_RTT_PEND_SHIFT, &call->rtt_avail); | |
676 | smp_wmb(); | |
677 | set_bit(i, &call->rtt_avail); | |
678 | } | |
679 | } | |
50235c4b | 680 | |
4700c4d8 DH |
681 | if (!matched) |
682 | trace_rxrpc_rtt_rx(call, rxrpc_rtt_rx_lost, 9, 0, acked_serial, 0, 0); | |
50235c4b DH |
683 | } |
684 | ||
17926a79 | 685 | /* |
248f219c | 686 | * Process the extra information that may be appended to an ACK packet |
17926a79 | 687 | */ |
248f219c DH |
688 | static void rxrpc_input_ackinfo(struct rxrpc_call *call, struct sk_buff *skb, |
689 | struct rxrpc_ackinfo *ackinfo) | |
17926a79 | 690 | { |
248f219c DH |
691 | struct rxrpc_skb_priv *sp = rxrpc_skb(skb); |
692 | struct rxrpc_peer *peer; | |
693 | unsigned int mtu; | |
702f2ac8 | 694 | bool wake = false; |
01fd0742 | 695 | u32 rwind = ntohl(ackinfo->rwind); |
248f219c | 696 | |
a4ea4c47 DH |
697 | if (rwind > RXRPC_TX_MAX_WINDOW) |
698 | rwind = RXRPC_TX_MAX_WINDOW; | |
702f2ac8 | 699 | if (call->tx_winsize != rwind) { |
702f2ac8 DH |
700 | if (rwind > call->tx_winsize) |
701 | wake = true; | |
a2ad7c21 | 702 | trace_rxrpc_rx_rwind_change(call, sp->hdr.serial, rwind, wake); |
702f2ac8 DH |
703 | call->tx_winsize = rwind; |
704 | } | |
705 | ||
08511150 DH |
706 | if (call->cong_ssthresh > rwind) |
707 | call->cong_ssthresh = rwind; | |
248f219c DH |
708 | |
709 | mtu = min(ntohl(ackinfo->rxMTU), ntohl(ackinfo->maxMTU)); | |
710 | ||
711 | peer = call->peer; | |
712 | if (mtu < peer->maxdata) { | |
3dd9c8b5 | 713 | spin_lock(&peer->lock); |
248f219c DH |
714 | peer->maxdata = mtu; |
715 | peer->mtu = mtu + peer->hdrsize; | |
3dd9c8b5 | 716 | spin_unlock(&peer->lock); |
248f219c | 717 | } |
702f2ac8 DH |
718 | |
719 | if (wake) | |
720 | wake_up(&call->waitq); | |
248f219c | 721 | } |
17926a79 | 722 | |
248f219c DH |
723 | /* |
724 | * Process individual soft ACKs. | |
725 | * | |
726 | * Each ACK in the array corresponds to one packet and can be either an ACK or | |
727 | * a NAK. If we get find an explicitly NAK'd packet we resend immediately; | |
728 | * packets that lie beyond the end of the ACK list are scheduled for resend by | |
729 | * the timer on the basis that the peer might just not have processed them at | |
730 | * the time the ACK was sent. | |
731 | */ | |
732 | static void rxrpc_input_soft_acks(struct rxrpc_call *call, u8 *acks, | |
31a1b989 DH |
733 | rxrpc_seq_t seq, int nr_acks, |
734 | struct rxrpc_ack_summary *summary) | |
248f219c | 735 | { |
d57a3a15 | 736 | unsigned int i; |
a4ea4c47 | 737 | |
d57a3a15 DH |
738 | for (i = 0; i < nr_acks; i++) { |
739 | if (acks[i] == RXRPC_ACK_TYPE_ACK) { | |
31a1b989 | 740 | summary->nr_acks++; |
31a1b989 | 741 | summary->nr_new_acks++; |
d57a3a15 DH |
742 | } else { |
743 | if (!summary->saw_nacks && | |
744 | call->acks_lowest_nak != seq + i) { | |
745 | call->acks_lowest_nak = seq + i; | |
31a1b989 DH |
746 | summary->new_low_nack = true; |
747 | } | |
d57a3a15 | 748 | summary->saw_nacks = true; |
17926a79 | 749 | } |
17926a79 DH |
750 | } |
751 | } | |
752 | ||
441fdee1 DH |
753 | /* |
754 | * Return true if the ACK is valid - ie. it doesn't appear to have regressed | |
755 | * with respect to the ack state conveyed by preceding ACKs. | |
756 | */ | |
757 | static bool rxrpc_is_ack_valid(struct rxrpc_call *call, | |
758 | rxrpc_seq_t first_pkt, rxrpc_seq_t prev_pkt) | |
759 | { | |
8940ba3c | 760 | rxrpc_seq_t base = READ_ONCE(call->acks_first_seq); |
441fdee1 DH |
761 | |
762 | if (after(first_pkt, base)) | |
763 | return true; /* The window advanced */ | |
764 | ||
765 | if (before(first_pkt, base)) | |
766 | return false; /* firstPacket regressed */ | |
767 | ||
8940ba3c | 768 | if (after_eq(prev_pkt, call->acks_prev_seq)) |
441fdee1 DH |
769 | return true; /* previousPacket hasn't regressed. */ |
770 | ||
771 | /* Some rx implementations put a serial number in previousPacket. */ | |
772 | if (after_eq(prev_pkt, base + call->tx_winsize)) | |
773 | return false; | |
774 | return true; | |
775 | } | |
776 | ||
17926a79 | 777 | /* |
248f219c DH |
778 | * Process an ACK packet. |
779 | * | |
780 | * ack.firstPacket is the sequence number of the first soft-ACK'd/NAK'd packet | |
781 | * in the ACK array. Anything before that is hard-ACK'd and may be discarded. | |
782 | * | |
783 | * A hard-ACK means that a packet has been processed and may be discarded; a | |
784 | * soft-ACK means that the packet may be discarded and retransmission | |
785 | * requested. A phase is complete when all packets are hard-ACK'd. | |
17926a79 | 786 | */ |
e8c3af6b | 787 | static void rxrpc_input_ack(struct rxrpc_call *call, struct sk_buff *skb) |
17926a79 | 788 | { |
31a1b989 | 789 | struct rxrpc_ack_summary summary = { 0 }; |
d57a3a15 | 790 | struct rxrpc_ackpacket ack; |
17926a79 | 791 | struct rxrpc_skb_priv *sp = rxrpc_skb(skb); |
d57a3a15 | 792 | struct rxrpc_ackinfo info; |
68528d93 | 793 | rxrpc_serial_t ack_serial, acked_serial; |
1a2391c3 | 794 | rxrpc_seq_t first_soft_ack, hard_ack, prev_pkt; |
775e5b71 | 795 | int nr_acks, offset, ioffset; |
248f219c DH |
796 | |
797 | _enter(""); | |
798 | ||
775e5b71 | 799 | offset = sizeof(struct rxrpc_wire_header); |
5e6ef4f1 | 800 | if (skb_copy_bits(skb, offset, &ack, sizeof(ack)) < 0) |
57af281e | 801 | return rxrpc_proto_abort(call, 0, rxrpc_badmsg_short_ack); |
d57a3a15 | 802 | offset += sizeof(ack); |
248f219c | 803 | |
68528d93 | 804 | ack_serial = sp->hdr.serial; |
d57a3a15 DH |
805 | acked_serial = ntohl(ack.serial); |
806 | first_soft_ack = ntohl(ack.firstPacket); | |
807 | prev_pkt = ntohl(ack.previousPacket); | |
248f219c | 808 | hard_ack = first_soft_ack - 1; |
d57a3a15 DH |
809 | nr_acks = ack.nAcks; |
810 | summary.ack_reason = (ack.reason < RXRPC_ACK__INVALID ? | |
811 | ack.reason : RXRPC_ACK__INVALID); | |
248f219c | 812 | |
68528d93 | 813 | trace_rxrpc_rx_ack(call, ack_serial, acked_serial, |
1a2391c3 | 814 | first_soft_ack, prev_pkt, |
b1d9f7fd | 815 | summary.ack_reason, nr_acks); |
d57a3a15 | 816 | rxrpc_inc_stat(call->rxnet, stat_rx_acks[ack.reason]); |
ec71eb9a | 817 | |
d57a3a15 | 818 | switch (ack.reason) { |
4700c4d8 | 819 | case RXRPC_ACK_PING_RESPONSE: |
4700c4d8 DH |
820 | rxrpc_complete_rtt_probe(call, skb->tstamp, acked_serial, ack_serial, |
821 | rxrpc_rtt_rx_ping_response); | |
822 | break; | |
823 | case RXRPC_ACK_REQUESTED: | |
824 | rxrpc_complete_rtt_probe(call, skb->tstamp, acked_serial, ack_serial, | |
825 | rxrpc_rtt_rx_requested_ack); | |
826 | break; | |
827 | default: | |
828 | if (acked_serial != 0) | |
829 | rxrpc_complete_rtt_probe(call, skb->tstamp, acked_serial, ack_serial, | |
830 | rxrpc_rtt_rx_cancel); | |
831 | break; | |
832 | } | |
8e83134d | 833 | |
d57a3a15 | 834 | if (ack.reason == RXRPC_ACK_PING) { |
72f0c6fb DH |
835 | rxrpc_send_ACK(call, RXRPC_ACK_PING_RESPONSE, ack_serial, |
836 | rxrpc_propose_ack_respond_to_ping); | |
248f219c | 837 | } else if (sp->hdr.flags & RXRPC_REQUEST_ACK) { |
72f0c6fb DH |
838 | rxrpc_send_ACK(call, RXRPC_ACK_REQUESTED, ack_serial, |
839 | rxrpc_propose_ack_respond_to_ack); | |
17926a79 DH |
840 | } |
841 | ||
adc9613f DH |
842 | /* If we get an EXCEEDS_WINDOW ACK from the server, it probably |
843 | * indicates that the client address changed due to NAT. The server | |
844 | * lost the call because it switched to a different peer. | |
845 | */ | |
d57a3a15 | 846 | if (unlikely(ack.reason == RXRPC_ACK_EXCEEDS_WINDOW) && |
adc9613f DH |
847 | first_soft_ack == 1 && |
848 | prev_pkt == 0 && | |
849 | rxrpc_is_client_call(call)) { | |
850 | rxrpc_set_call_completion(call, RXRPC_CALL_REMOTELY_ABORTED, | |
851 | 0, -ENETRESET); | |
5e6ef4f1 | 852 | return; |
adc9613f DH |
853 | } |
854 | ||
855 | /* If we get an OUT_OF_SEQUENCE ACK from the server, that can also | |
856 | * indicate a change of address. However, we can retransmit the call | |
857 | * if we still have it buffered to the beginning. | |
858 | */ | |
d57a3a15 | 859 | if (unlikely(ack.reason == RXRPC_ACK_OUT_OF_SEQUENCE) && |
adc9613f DH |
860 | first_soft_ack == 1 && |
861 | prev_pkt == 0 && | |
a4ea4c47 | 862 | call->acks_hard_ack == 0 && |
adc9613f DH |
863 | rxrpc_is_client_call(call)) { |
864 | rxrpc_set_call_completion(call, RXRPC_CALL_REMOTELY_ABORTED, | |
865 | 0, -ENETRESET); | |
5e6ef4f1 | 866 | return; |
adc9613f DH |
867 | } |
868 | ||
1a2391c3 | 869 | /* Discard any out-of-order or duplicate ACKs (outside lock). */ |
441fdee1 | 870 | if (!rxrpc_is_ack_valid(call, first_soft_ack, prev_pkt)) { |
68528d93 | 871 | trace_rxrpc_rx_discard_ack(call->debug_id, ack_serial, |
8940ba3c DH |
872 | first_soft_ack, call->acks_first_seq, |
873 | prev_pkt, call->acks_prev_seq); | |
5e6ef4f1 | 874 | return; |
d1f12947 | 875 | } |
c1e15b49 | 876 | |
d57a3a15 | 877 | info.rxMTU = 0; |
775e5b71 | 878 | ioffset = offset + nr_acks + 3; |
d57a3a15 | 879 | if (skb->len >= ioffset + sizeof(info) && |
5e6ef4f1 | 880 | skb_copy_bits(skb, ioffset, &info, sizeof(info)) < 0) |
57af281e | 881 | return rxrpc_proto_abort(call, 0, rxrpc_badmsg_short_ack_info); |
d57a3a15 DH |
882 | |
883 | if (nr_acks > 0) | |
884 | skb_condense(skb); | |
c1e15b49 | 885 | |
298bc15b | 886 | call->acks_latest_ts = skb->tstamp; |
8940ba3c DH |
887 | call->acks_first_seq = first_soft_ack; |
888 | call->acks_prev_seq = prev_pkt; | |
1a2391c3 | 889 | |
d57a3a15 DH |
890 | switch (ack.reason) { |
891 | case RXRPC_ACK_PING: | |
892 | break; | |
d57a3a15 DH |
893 | default: |
894 | if (after(acked_serial, call->acks_highest_serial)) | |
895 | call->acks_highest_serial = acked_serial; | |
896 | break; | |
897 | } | |
589a0c1e | 898 | |
298bc15b | 899 | /* Parse rwind and mtu sizes if provided. */ |
d57a3a15 DH |
900 | if (info.rxMTU) |
901 | rxrpc_input_ackinfo(call, skb, &info); | |
17926a79 | 902 | |
5e6ef4f1 | 903 | if (first_soft_ack == 0) |
57af281e | 904 | return rxrpc_proto_abort(call, 0, rxrpc_eproto_ackr_zero); |
17926a79 | 905 | |
248f219c | 906 | /* Ignore ACKs unless we are or have just been transmitting. */ |
146d8fef | 907 | switch (READ_ONCE(call->state)) { |
248f219c DH |
908 | case RXRPC_CALL_CLIENT_SEND_REQUEST: |
909 | case RXRPC_CALL_CLIENT_AWAIT_REPLY: | |
910 | case RXRPC_CALL_SERVER_SEND_REPLY: | |
911 | case RXRPC_CALL_SERVER_AWAIT_ACK: | |
912 | break; | |
17926a79 | 913 | default: |
5e6ef4f1 | 914 | return; |
248f219c | 915 | } |
17926a79 | 916 | |
a4ea4c47 | 917 | if (before(hard_ack, call->acks_hard_ack) || |
5e6ef4f1 | 918 | after(hard_ack, call->tx_top)) |
57af281e | 919 | return rxrpc_proto_abort(call, 0, rxrpc_eproto_ackr_outside_window); |
5e6ef4f1 | 920 | if (nr_acks > call->tx_top - hard_ack) |
57af281e | 921 | return rxrpc_proto_abort(call, 0, rxrpc_eproto_ackr_sack_overflow); |
17926a79 | 922 | |
a4ea4c47 | 923 | if (after(hard_ack, call->acks_hard_ack)) { |
c479d5f2 | 924 | if (rxrpc_rotate_tx_window(call, hard_ack, &summary)) { |
57af281e | 925 | rxrpc_end_tx_phase(call, false, rxrpc_eproto_unexpected_ack); |
5e6ef4f1 | 926 | return; |
c479d5f2 DH |
927 | } |
928 | } | |
17926a79 | 929 | |
70790dbe | 930 | if (nr_acks > 0) { |
5e6ef4f1 | 931 | if (offset > (int)skb->len - nr_acks) |
57af281e | 932 | return rxrpc_proto_abort(call, 0, rxrpc_eproto_ackr_short_sack); |
d57a3a15 DH |
933 | rxrpc_input_soft_acks(call, skb->data + offset, first_soft_ack, |
934 | nr_acks, &summary); | |
70790dbe DH |
935 | } |
936 | ||
a4ea4c47 | 937 | if (test_bit(RXRPC_CALL_TX_LAST, &call->flags) && |
a9f312d9 DH |
938 | summary.nr_acks == call->tx_top - hard_ack && |
939 | rxrpc_is_client_call(call)) | |
72f0c6fb DH |
940 | rxrpc_propose_ping(call, ack_serial, |
941 | rxrpc_propose_ack_ping_for_lost_reply); | |
57494343 | 942 | |
c1e15b49 | 943 | rxrpc_congestion_management(call, skb, &summary, acked_serial); |
17926a79 DH |
944 | } |
945 | ||
946 | /* | |
248f219c | 947 | * Process an ACKALL packet. |
17926a79 | 948 | */ |
248f219c | 949 | static void rxrpc_input_ackall(struct rxrpc_call *call, struct sk_buff *skb) |
17926a79 | 950 | { |
31a1b989 | 951 | struct rxrpc_ack_summary summary = { 0 }; |
17926a79 | 952 | |
c479d5f2 | 953 | if (rxrpc_rotate_tx_window(call, call->tx_top, &summary)) |
57af281e | 954 | rxrpc_end_tx_phase(call, false, rxrpc_eproto_unexpected_ackall); |
248f219c | 955 | } |
17926a79 | 956 | |
248f219c | 957 | /* |
005ede28 | 958 | * Process an ABORT packet directed at a call. |
248f219c DH |
959 | */ |
960 | static void rxrpc_input_abort(struct rxrpc_call *call, struct sk_buff *skb) | |
961 | { | |
962 | struct rxrpc_skb_priv *sp = rxrpc_skb(skb); | |
17926a79 | 963 | |
f14febd8 | 964 | trace_rxrpc_rx_abort(call, sp->hdr.serial, skb->priority); |
005ede28 | 965 | |
5ac0d622 | 966 | rxrpc_set_call_completion(call, RXRPC_CALL_REMOTELY_ABORTED, |
f14febd8 | 967 | skb->priority, -ECONNABORTED); |
17926a79 DH |
968 | } |
969 | ||
970 | /* | |
248f219c | 971 | * Process an incoming call packet. |
17926a79 | 972 | */ |
5e6ef4f1 | 973 | void rxrpc_input_call_packet(struct rxrpc_call *call, struct sk_buff *skb) |
17926a79 | 974 | { |
248f219c | 975 | struct rxrpc_skb_priv *sp = rxrpc_skb(skb); |
a158bdd3 | 976 | unsigned long timo; |
17926a79 | 977 | |
7727640c | 978 | _enter("%p,%p", call, skb); |
17926a79 | 979 | |
5e6ef4f1 DH |
980 | if (sp->hdr.serviceId != call->dest_srx.srx_service) |
981 | call->dest_srx.srx_service = sp->hdr.serviceId; | |
982 | if ((int)sp->hdr.serial - (int)call->rx_serial > 0) | |
983 | call->rx_serial = sp->hdr.serial; | |
984 | if (!test_bit(RXRPC_CALL_RX_HEARD, &call->flags)) | |
985 | set_bit(RXRPC_CALL_RX_HEARD, &call->flags); | |
986 | ||
a158bdd3 DH |
987 | timo = READ_ONCE(call->next_rx_timo); |
988 | if (timo) { | |
989 | unsigned long now = jiffies, expect_rx_by; | |
990 | ||
c54e43d7 | 991 | expect_rx_by = now + timo; |
a158bdd3 DH |
992 | WRITE_ONCE(call->expect_rx_by, expect_rx_by); |
993 | rxrpc_reduce_call_timer(call, expect_rx_by, now, | |
994 | rxrpc_timer_set_for_normal); | |
995 | } | |
3d7682af | 996 | |
248f219c DH |
997 | switch (sp->hdr.type) { |
998 | case RXRPC_PACKET_TYPE_DATA: | |
57af281e | 999 | return rxrpc_input_data(call, skb); |
f5c17aae | 1000 | |
248f219c | 1001 | case RXRPC_PACKET_TYPE_ACK: |
57af281e | 1002 | return rxrpc_input_ack(call, skb); |
17926a79 | 1003 | |
248f219c | 1004 | case RXRPC_PACKET_TYPE_BUSY: |
248f219c DH |
1005 | /* Just ignore BUSY packets from the server; the retry and |
1006 | * lifespan timers will take care of business. BUSY packets | |
1007 | * from the client don't make sense. | |
1008 | */ | |
57af281e | 1009 | return; |
17926a79 | 1010 | |
248f219c | 1011 | case RXRPC_PACKET_TYPE_ABORT: |
57af281e | 1012 | return rxrpc_input_abort(call, skb); |
17926a79 | 1013 | |
248f219c | 1014 | case RXRPC_PACKET_TYPE_ACKALL: |
57af281e | 1015 | return rxrpc_input_ackall(call, skb); |
f5c17aae | 1016 | |
248f219c | 1017 | default: |
248f219c | 1018 | break; |
17926a79 | 1019 | } |
17926a79 DH |
1020 | } |
1021 | ||
b3156274 | 1022 | /* |
c1e15b49 DH |
1023 | * Handle a new service call on a channel implicitly completing the preceding |
1024 | * call on that channel. This does not apply to client conns. | |
b3156274 DH |
1025 | * |
1026 | * TODO: If callNumber > call_id + 1, renegotiate security. | |
1027 | */ | |
5e6ef4f1 | 1028 | void rxrpc_implicit_end_call(struct rxrpc_call *call, struct sk_buff *skb) |
b3156274 | 1029 | { |
146d8fef | 1030 | switch (READ_ONCE(call->state)) { |
b3156274 DH |
1031 | case RXRPC_CALL_SERVER_AWAIT_ACK: |
1032 | rxrpc_call_completed(call); | |
df561f66 | 1033 | fallthrough; |
b3156274 DH |
1034 | case RXRPC_CALL_COMPLETE: |
1035 | break; | |
1036 | default: | |
57af281e DH |
1037 | rxrpc_abort_call(call, 0, RX_CALL_DEAD, -ESHUTDOWN, |
1038 | rxrpc_eproto_improper_term); | |
c1e15b49 | 1039 | trace_rxrpc_improper_term(call); |
b3156274 DH |
1040 | break; |
1041 | } | |
1042 | ||
5e6ef4f1 | 1043 | rxrpc_input_call_event(call, skb); |
b3156274 | 1044 | } |