]> bbs.cooldavid.org Git - net-next-2.6.git/blame - net/dccp/input.c
dccp: Integration of dynamic feature activation - part 3 (client side)
[net-next-2.6.git] / net / dccp / input.c
CommitLineData
7c657876
ACM
1/*
2 * net/dccp/input.c
8109b02b 3 *
7c657876
ACM
4 * An implementation of the DCCP protocol
5 * Arnaldo Carvalho de Melo <acme@conectiva.com.br>
6 *
7 * This program is free software; you can redistribute it and/or
8 * modify it under the terms of the GNU General Public License
9 * as published by the Free Software Foundation; either version
10 * 2 of the License, or (at your option) any later version.
11 */
12
7c657876
ACM
13#include <linux/dccp.h>
14#include <linux/skbuff.h>
15
16#include <net/sock.h>
17
ae31c339 18#include "ackvec.h"
7c657876
ACM
19#include "ccid.h"
20#include "dccp.h"
21
bd5435e7
IM
22/* rate-limit for syncs in reply to sequence-invalid packets; RFC 4340, 7.5.4 */
23int sysctl_dccp_sync_ratelimit __read_mostly = HZ / 8;
24
69567d0b 25static void dccp_enqueue_skb(struct sock *sk, struct sk_buff *skb)
7c657876 26{
7c657876
ACM
27 __skb_pull(skb, dccp_hdr(skb)->dccph_doff * 4);
28 __skb_queue_tail(&sk->sk_receive_queue, skb);
29 skb_set_owner_r(skb, sk);
30 sk->sk_data_ready(sk, 0);
31}
32
69567d0b
GR
33static void dccp_fin(struct sock *sk, struct sk_buff *skb)
34{
35 /*
36 * On receiving Close/CloseReq, both RD/WR shutdown are performed.
37 * RFC 4340, 8.3 says that we MAY send further Data/DataAcks after
38 * receiving the closing segment, but there is no guarantee that such
39 * data will be processed at all.
40 */
41 sk->sk_shutdown = SHUTDOWN_MASK;
42 sock_set_flag(sk, SOCK_DONE);
43 dccp_enqueue_skb(sk, skb);
44}
45
0c869620 46static int dccp_rcv_close(struct sock *sk, struct sk_buff *skb)
7c657876 47{
0c869620
GR
48 int queued = 0;
49
50 switch (sk->sk_state) {
51 /*
52 * We ignore Close when received in one of the following states:
53 * - CLOSED (may be a late or duplicate packet)
54 * - PASSIVE_CLOSEREQ (the peer has sent a CloseReq earlier)
55 * - RESPOND (already handled by dccp_check_req)
56 */
57 case DCCP_CLOSING:
58 /*
59 * Simultaneous-close: receiving a Close after sending one. This
60 * can happen if both client and server perform active-close and
61 * will result in an endless ping-pong of crossing and retrans-
62 * mitted Close packets, which only terminates when one of the
63 * nodes times out (min. 64 seconds). Quicker convergence can be
64 * achieved when one of the nodes acts as tie-breaker.
65 * This is ok as both ends are done with data transfer and each
66 * end is just waiting for the other to acknowledge termination.
67 */
68 if (dccp_sk(sk)->dccps_role != DCCP_ROLE_CLIENT)
69 break;
70 /* fall through */
71 case DCCP_REQUESTING:
72 case DCCP_ACTIVE_CLOSEREQ:
73 dccp_send_reset(sk, DCCP_RESET_CODE_CLOSED);
74 dccp_done(sk);
75 break;
76 case DCCP_OPEN:
77 case DCCP_PARTOPEN:
78 /* Give waiting application a chance to read pending data */
79 queued = 1;
80 dccp_fin(sk, skb);
81 dccp_set_state(sk, DCCP_PASSIVE_CLOSE);
82 /* fall through */
83 case DCCP_PASSIVE_CLOSE:
84 /*
85 * Retransmitted Close: we have already enqueued the first one.
86 */
87 sk_wake_async(sk, SOCK_WAKE_WAITD, POLL_HUP);
88 }
89 return queued;
7c657876
ACM
90}
91
0c869620 92static int dccp_rcv_closereq(struct sock *sk, struct sk_buff *skb)
7c657876 93{
0c869620
GR
94 int queued = 0;
95
7c657876
ACM
96 /*
97 * Step 7: Check for unexpected packet types
98 * If (S.is_server and P.type == CloseReq)
99 * Send Sync packet acknowledging P.seqno
100 * Drop packet and return
101 */
102 if (dccp_sk(sk)->dccps_role != DCCP_ROLE_CLIENT) {
e92ae93a 103 dccp_send_sync(sk, DCCP_SKB_CB(skb)->dccpd_seq, DCCP_PKT_SYNC);
0c869620 104 return queued;
7c657876
ACM
105 }
106
0c869620
GR
107 /* Step 13: process relevant Client states < CLOSEREQ */
108 switch (sk->sk_state) {
109 case DCCP_REQUESTING:
110 dccp_send_close(sk, 0);
811265b8 111 dccp_set_state(sk, DCCP_CLOSING);
0c869620
GR
112 break;
113 case DCCP_OPEN:
114 case DCCP_PARTOPEN:
115 /* Give waiting application a chance to read pending data */
116 queued = 1;
117 dccp_fin(sk, skb);
118 dccp_set_state(sk, DCCP_PASSIVE_CLOSEREQ);
119 /* fall through */
120 case DCCP_PASSIVE_CLOSEREQ:
121 sk_wake_async(sk, SOCK_WAKE_WAITD, POLL_HUP);
122 }
123 return queued;
7c657876
ACM
124}
125
d8ef2c29
GR
126static u8 dccp_reset_code_convert(const u8 code)
127{
128 const u8 error_code[] = {
129 [DCCP_RESET_CODE_CLOSED] = 0, /* normal termination */
130 [DCCP_RESET_CODE_UNSPECIFIED] = 0, /* nothing known */
131 [DCCP_RESET_CODE_ABORTED] = ECONNRESET,
132
133 [DCCP_RESET_CODE_NO_CONNECTION] = ECONNREFUSED,
134 [DCCP_RESET_CODE_CONNECTION_REFUSED] = ECONNREFUSED,
135 [DCCP_RESET_CODE_TOO_BUSY] = EUSERS,
136 [DCCP_RESET_CODE_AGGRESSION_PENALTY] = EDQUOT,
137
138 [DCCP_RESET_CODE_PACKET_ERROR] = ENOMSG,
139 [DCCP_RESET_CODE_BAD_INIT_COOKIE] = EBADR,
140 [DCCP_RESET_CODE_BAD_SERVICE_CODE] = EBADRQC,
141 [DCCP_RESET_CODE_OPTION_ERROR] = EILSEQ,
142 [DCCP_RESET_CODE_MANDATORY_ERROR] = EOPNOTSUPP,
143 };
144
145 return code >= DCCP_MAX_RESET_CODES ? 0 : error_code[code];
146}
147
148static void dccp_rcv_reset(struct sock *sk, struct sk_buff *skb)
149{
150 u8 err = dccp_reset_code_convert(dccp_hdr_reset(skb)->dccph_reset_code);
151
152 sk->sk_err = err;
153
154 /* Queue the equivalent of TCP fin so that dccp_recvmsg exits the loop */
155 dccp_fin(sk, skb);
156
157 if (err && !sock_flag(sk, SOCK_DEAD))
8d8ad9d7 158 sk_wake_async(sk, SOCK_WAKE_IO, POLL_ERR);
d8ef2c29
GR
159 dccp_time_wait(sk, DCCP_TIME_WAIT, 0);
160}
161
410e27a4 162static void dccp_event_ack_recv(struct sock *sk, struct sk_buff *skb)
7c657876 163{
410e27a4 164 struct dccp_sock *dp = dccp_sk(sk);
7c657876 165
410e27a4
GR
166 if (dccp_msk(sk)->dccpms_send_ack_vector)
167 dccp_ackvec_check_rcv_ackno(dp->dccps_hc_rx_ackvec, sk,
168 DCCP_SKB_CB(skb)->dccpd_ack_seq);
7c657876
ACM
169}
170
8e8c71f1
GR
171static void dccp_deliver_input_to_ccids(struct sock *sk, struct sk_buff *skb)
172{
173 const struct dccp_sock *dp = dccp_sk(sk);
174
175 /* Don't deliver to RX CCID when node has shut down read end. */
176 if (!(sk->sk_shutdown & RCV_SHUTDOWN))
177 ccid_hc_rx_packet_recv(dp->dccps_hc_rx_ccid, sk, skb);
178 /*
179 * Until the TX queue has been drained, we can not honour SHUT_WR, since
180 * we need received feedback as input to adjust congestion control.
181 */
182 if (sk->sk_write_queue.qlen > 0 || !(sk->sk_shutdown & SEND_SHUTDOWN))
183 ccid_hc_tx_packet_recv(dp->dccps_hc_tx_ccid, sk, skb);
184}
185
7c657876
ACM
186static int dccp_check_seqno(struct sock *sk, struct sk_buff *skb)
187{
188 const struct dccp_hdr *dh = dccp_hdr(skb);
189 struct dccp_sock *dp = dccp_sk(sk);
cbe1f5f8
GR
190 u64 lswl, lawl, seqno = DCCP_SKB_CB(skb)->dccpd_seq,
191 ackno = DCCP_SKB_CB(skb)->dccpd_ack_seq;
7c657876
ACM
192
193 /*
194 * Step 5: Prepare sequence numbers for Sync
195 * If P.type == Sync or P.type == SyncAck,
196 * If S.AWL <= P.ackno <= S.AWH and P.seqno >= S.SWL,
197 * / * P is valid, so update sequence number variables
198 * accordingly. After this update, P will pass the tests
199 * in Step 6. A SyncAck is generated if necessary in
200 * Step 15 * /
201 * Update S.GSR, S.SWL, S.SWH
202 * Otherwise,
203 * Drop packet and return
204 */
8109b02b 205 if (dh->dccph_type == DCCP_PKT_SYNC ||
7c657876 206 dh->dccph_type == DCCP_PKT_SYNCACK) {
cbe1f5f8
GR
207 if (between48(ackno, dp->dccps_awl, dp->dccps_awh) &&
208 dccp_delta_seqno(dp->dccps_swl, seqno) >= 0)
209 dccp_update_gsr(sk, seqno);
7c657876
ACM
210 else
211 return -1;
e92ae93a 212 }
c9eaf173 213
7c657876
ACM
214 /*
215 * Step 6: Check sequence numbers
216 * Let LSWL = S.SWL and LAWL = S.AWL
217 * If P.type == CloseReq or P.type == Close or P.type == Reset,
218 * LSWL := S.GSR + 1, LAWL := S.GAR
219 * If LSWL <= P.seqno <= S.SWH
220 * and (P.ackno does not exist or LAWL <= P.ackno <= S.AWH),
221 * Update S.GSR, S.SWL, S.SWH
222 * If P.type != Sync,
223 * Update S.GAR
7c657876 224 */
e92ae93a
ACM
225 lswl = dp->dccps_swl;
226 lawl = dp->dccps_awl;
227
228 if (dh->dccph_type == DCCP_PKT_CLOSEREQ ||
c59eab46
ACM
229 dh->dccph_type == DCCP_PKT_CLOSE ||
230 dh->dccph_type == DCCP_PKT_RESET) {
cbe1f5f8 231 lswl = ADD48(dp->dccps_gsr, 1);
7c657876
ACM
232 lawl = dp->dccps_gar;
233 }
234
cbe1f5f8
GR
235 if (between48(seqno, lswl, dp->dccps_swh) &&
236 (ackno == DCCP_PKT_WITHOUT_ACK_SEQ ||
237 between48(ackno, lawl, dp->dccps_awh))) {
238 dccp_update_gsr(sk, seqno);
7c657876
ACM
239
240 if (dh->dccph_type != DCCP_PKT_SYNC &&
cbe1f5f8
GR
241 (ackno != DCCP_PKT_WITHOUT_ACK_SEQ))
242 dp->dccps_gar = ackno;
7c657876 243 } else {
a94f0f97
GR
244 unsigned long now = jiffies;
245 /*
246 * Step 6: Check sequence numbers
247 * Otherwise,
248 * If P.type == Reset,
249 * Send Sync packet acknowledging S.GSR
250 * Otherwise,
251 * Send Sync packet acknowledging P.seqno
252 * Drop packet and return
253 *
254 * These Syncs are rate-limited as per RFC 4340, 7.5.4:
255 * at most 1 / (dccp_sync_rate_limit * HZ) Syncs per second.
256 */
257 if (time_before(now, (dp->dccps_rate_last +
258 sysctl_dccp_sync_ratelimit)))
259 return 0;
260
59348b19
GR
261 DCCP_WARN("DCCP: Step 6 failed for %s packet, "
262 "(LSWL(%llu) <= P.seqno(%llu) <= S.SWH(%llu)) and "
263 "(P.ackno %s or LAWL(%llu) <= P.ackno(%llu) <= S.AWH(%llu), "
264 "sending SYNC...\n", dccp_packet_name(dh->dccph_type),
cbe1f5f8 265 (unsigned long long) lswl, (unsigned long long) seqno,
59348b19 266 (unsigned long long) dp->dccps_swh,
cbe1f5f8
GR
267 (ackno == DCCP_PKT_WITHOUT_ACK_SEQ) ? "doesn't exist"
268 : "exists",
269 (unsigned long long) lawl, (unsigned long long) ackno,
59348b19 270 (unsigned long long) dp->dccps_awh);
a94f0f97
GR
271
272 dp->dccps_rate_last = now;
273
e155d769
GR
274 if (dh->dccph_type == DCCP_PKT_RESET)
275 seqno = dp->dccps_gsr;
cbe1f5f8 276 dccp_send_sync(sk, seqno, DCCP_PKT_SYNC);
7c657876
ACM
277 return -1;
278 }
279
280 return 0;
281}
282
c25a18ba
ACM
283static int __dccp_rcv_established(struct sock *sk, struct sk_buff *skb,
284 const struct dccp_hdr *dh, const unsigned len)
7c657876
ACM
285{
286 struct dccp_sock *dp = dccp_sk(sk);
287
7c657876
ACM
288 switch (dccp_hdr(skb)->dccph_type) {
289 case DCCP_PKT_DATAACK:
290 case DCCP_PKT_DATA:
291 /*
8e8c71f1
GR
292 * FIXME: schedule DATA_DROPPED (RFC 4340, 11.7.2) if and when
293 * - sk_shutdown == RCV_SHUTDOWN, use Code 1, "Not Listening"
294 * - sk_receive_queue is full, use Code 2, "Receive Buffer"
7c657876 295 */
69567d0b 296 dccp_enqueue_skb(sk, skb);
7c657876
ACM
297 return 0;
298 case DCCP_PKT_ACK:
299 goto discard;
300 case DCCP_PKT_RESET:
301 /*
302 * Step 9: Process Reset
303 * If P.type == Reset,
304 * Tear down connection
305 * S.state := TIMEWAIT
306 * Set TIMEWAIT timer
307 * Drop packet and return
d8ef2c29
GR
308 */
309 dccp_rcv_reset(sk, skb);
7c657876
ACM
310 return 0;
311 case DCCP_PKT_CLOSEREQ:
0c869620
GR
312 if (dccp_rcv_closereq(sk, skb))
313 return 0;
7c657876
ACM
314 goto discard;
315 case DCCP_PKT_CLOSE:
0c869620
GR
316 if (dccp_rcv_close(sk, skb))
317 return 0;
318 goto discard;
7c657876 319 case DCCP_PKT_REQUEST:
8109b02b
ACM
320 /* Step 7
321 * or (S.is_server and P.type == Response)
7c657876
ACM
322 * or (S.is_client and P.type == Request)
323 * or (S.state >= OPEN and P.type == Request
324 * and P.seqno >= S.OSR)
325 * or (S.state >= OPEN and P.type == Response
326 * and P.seqno >= S.OSR)
327 * or (S.state == RESPOND and P.type == Data),
328 * Send Sync packet acknowledging P.seqno
329 * Drop packet and return
330 */
331 if (dp->dccps_role != DCCP_ROLE_LISTEN)
332 goto send_sync;
333 goto check_seq;
334 case DCCP_PKT_RESPONSE:
335 if (dp->dccps_role != DCCP_ROLE_CLIENT)
336 goto send_sync;
337check_seq:
8d13bf9a
GR
338 if (dccp_delta_seqno(dp->dccps_osr,
339 DCCP_SKB_CB(skb)->dccpd_seq) >= 0) {
7c657876 340send_sync:
e92ae93a
ACM
341 dccp_send_sync(sk, DCCP_SKB_CB(skb)->dccpd_seq,
342 DCCP_PKT_SYNC);
7c657876
ACM
343 }
344 break;
e92ae93a
ACM
345 case DCCP_PKT_SYNC:
346 dccp_send_sync(sk, DCCP_SKB_CB(skb)->dccpd_seq,
347 DCCP_PKT_SYNCACK);
348 /*
0e64e94e 349 * From RFC 4340, sec. 5.7
e92ae93a
ACM
350 *
351 * As with DCCP-Ack packets, DCCP-Sync and DCCP-SyncAck packets
352 * MAY have non-zero-length application data areas, whose
0e64e94e 353 * contents receivers MUST ignore.
e92ae93a
ACM
354 */
355 goto discard;
7c657876
ACM
356 }
357
358 DCCP_INC_STATS_BH(DCCP_MIB_INERRS);
359discard:
360 __kfree_skb(skb);
361 return 0;
362}
363
709dd3aa
AB
364int dccp_rcv_established(struct sock *sk, struct sk_buff *skb,
365 const struct dccp_hdr *dh, const unsigned len)
366{
410e27a4
GR
367 struct dccp_sock *dp = dccp_sk(sk);
368
709dd3aa
AB
369 if (dccp_check_seqno(sk, skb))
370 goto discard;
371
8b819412 372 if (dccp_parse_options(sk, NULL, skb))
ba1a6c7b 373 return 1;
709dd3aa 374
410e27a4
GR
375 if (DCCP_SKB_CB(skb)->dccpd_ack_seq != DCCP_PKT_WITHOUT_ACK_SEQ)
376 dccp_event_ack_recv(sk, skb);
377
378 if (dccp_msk(sk)->dccpms_send_ack_vector &&
379 dccp_ackvec_add(dp->dccps_hc_rx_ackvec, sk,
380 DCCP_SKB_CB(skb)->dccpd_seq,
381 DCCP_ACKVEC_STATE_RECEIVED))
382 goto discard;
8e8c71f1 383 dccp_deliver_input_to_ccids(sk, skb);
709dd3aa
AB
384
385 return __dccp_rcv_established(sk, skb, dh, len);
386discard:
387 __kfree_skb(skb);
388 return 0;
389}
390
f21e68ca
ACM
391EXPORT_SYMBOL_GPL(dccp_rcv_established);
392
7c657876
ACM
393static int dccp_rcv_request_sent_state_process(struct sock *sk,
394 struct sk_buff *skb,
395 const struct dccp_hdr *dh,
396 const unsigned len)
397{
8109b02b 398 /*
7c657876
ACM
399 * Step 4: Prepare sequence numbers in REQUEST
400 * If S.state == REQUEST,
401 * If (P.type == Response or P.type == Reset)
402 * and S.AWL <= P.ackno <= S.AWH,
403 * / * Set sequence number variables corresponding to the
404 * other endpoint, so P will pass the tests in Step 6 * /
405 * Set S.GSR, S.ISR, S.SWL, S.SWH
406 * / * Response processing continues in Step 10; Reset
407 * processing continues in Step 9 * /
408 */
409 if (dh->dccph_type == DCCP_PKT_RESPONSE) {
410 const struct inet_connection_sock *icsk = inet_csk(sk);
411 struct dccp_sock *dp = dccp_sk(sk);
3393da82 412 long tstamp = dccp_timestamp();
7c657876 413
7690af3f
ACM
414 if (!between48(DCCP_SKB_CB(skb)->dccpd_ack_seq,
415 dp->dccps_awl, dp->dccps_awh)) {
416 dccp_pr_debug("invalid ackno: S.AWL=%llu, "
417 "P.ackno=%llu, S.AWH=%llu \n",
418 (unsigned long long)dp->dccps_awl,
419 (unsigned long long)DCCP_SKB_CB(skb)->dccpd_ack_seq,
420 (unsigned long long)dp->dccps_awh);
7c657876
ACM
421 goto out_invalid_packet;
422 }
9e377202 423
991d927c
GR
424 /*
425 * If option processing (Step 8) failed, return 1 here so that
426 * dccp_v4_do_rcv() sends a Reset. The Reset code depends on
427 * the option type and is set in dccp_parse_options().
428 */
8b819412 429 if (dccp_parse_options(sk, NULL, skb))
991d927c 430 return 1;
afe00251 431
3393da82
GR
432 /* Obtain usec RTT sample from SYN exchange (used by CCID 3) */
433 if (likely(dp->dccps_options_received.dccpor_timestamp_echo))
434 dp->dccps_syn_rtt = dccp_sample_rtt(sk, 10 * (tstamp -
435 dp->dccps_options_received.dccpor_timestamp_echo));
89560b53 436
410e27a4
GR
437 if (dccp_msk(sk)->dccpms_send_ack_vector &&
438 dccp_ackvec_add(dp->dccps_hc_rx_ackvec, sk,
439 DCCP_SKB_CB(skb)->dccpd_seq,
440 DCCP_ACKVEC_STATE_RECEIVED))
441 goto out_invalid_packet; /* FIXME: change error code */
442
d28934ad
GR
443 /* Stop the REQUEST timer */
444 inet_csk_clear_xmit_timer(sk, ICSK_TIME_RETRANS);
445 WARN_ON(sk->sk_send_head == NULL);
446 kfree_skb(sk->sk_send_head);
447 sk->sk_send_head = NULL;
448
410e27a4
GR
449 dp->dccps_isr = DCCP_SKB_CB(skb)->dccpd_seq;
450 dccp_update_gsr(sk, dp->dccps_isr);
03ace394 451 /*
410e27a4
GR
452 * SWL and AWL are initially adjusted so that they are not less than
453 * the initial Sequence Numbers received and sent, respectively:
454 * SWL := max(GSR + 1 - floor(W/4), ISR),
455 * AWL := max(GSS - W' + 1, ISS).
456 * These adjustments MUST be applied only at the beginning of the
457 * connection.
458 *
459 * AWL was adjusted in dccp_v4_connect -acme
03ace394 460 */
410e27a4
GR
461 dccp_set_seqno(&dp->dccps_swl,
462 max48(dp->dccps_swl, dp->dccps_isr));
7c657876 463
d83d8461 464 dccp_sync_mss(sk, icsk->icsk_pmtu_cookie);
7c657876
ACM
465
466 /*
467 * Step 10: Process REQUEST state (second part)
468 * If S.state == REQUEST,
7690af3f
ACM
469 * / * If we get here, P is a valid Response from the
470 * server (see Step 4), and we should move to
471 * PARTOPEN state. PARTOPEN means send an Ack,
472 * don't send Data packets, retransmit Acks
473 * periodically, and always include any Init Cookie
474 * from the Response * /
7c657876
ACM
475 * S.state := PARTOPEN
476 * Set PARTOPEN timer
8109b02b 477 * Continue with S.state == PARTOPEN
7690af3f
ACM
478 * / * Step 12 will send the Ack completing the
479 * three-way handshake * /
7c657876
ACM
480 */
481 dccp_set_state(sk, DCCP_PARTOPEN);
482
991d927c
GR
483 /*
484 * If feature negotiation was successful, activate features now;
485 * an activation failure means that this host could not activate
486 * one ore more features (e.g. insufficient memory), which would
487 * leave at least one feature in an undefined state.
488 */
489 if (dccp_feat_activate_values(sk, &dp->dccps_featneg))
490 goto unable_to_proceed;
491
7c657876 492 /* Make sure socket is routed, for correct metrics. */
57cca05a 493 icsk->icsk_af_ops->rebuild_header(sk);
7c657876
ACM
494
495 if (!sock_flag(sk, SOCK_DEAD)) {
496 sk->sk_state_change(sk);
8d8ad9d7 497 sk_wake_async(sk, SOCK_WAKE_IO, POLL_OUT);
7c657876
ACM
498 }
499
500 if (sk->sk_write_pending || icsk->icsk_ack.pingpong ||
501 icsk->icsk_accept_queue.rskq_defer_accept) {
502 /* Save one ACK. Data will be ready after
503 * several ticks, if write_pending is set.
504 *
505 * It may be deleted, but with this feature tcpdumps
506 * look so _wonderfully_ clever, that I was not able
507 * to stand against the temptation 8) --ANK
508 */
509 /*
510 * OK, in DCCP we can as well do a similar trick, its
511 * even in the draft, but there is no need for us to
512 * schedule an ack here, as dccp_sendmsg does this for
513 * us, also stated in the draft. -acme
514 */
515 __kfree_skb(skb);
516 return 0;
8109b02b 517 }
7c657876
ACM
518 dccp_send_ack(sk);
519 return -1;
520 }
521
522out_invalid_packet:
0c10c5d9
ACM
523 /* dccp_v4_do_rcv will send a reset */
524 DCCP_SKB_CB(skb)->dccpd_reset_code = DCCP_RESET_CODE_PACKET_ERROR;
8109b02b 525 return 1;
991d927c
GR
526
527unable_to_proceed:
528 DCCP_SKB_CB(skb)->dccpd_reset_code = DCCP_RESET_CODE_ABORTED;
529 /*
530 * We mark this socket as no longer usable, so that the loop in
531 * dccp_sendmsg() terminates and the application gets notified.
532 */
533 dccp_set_state(sk, DCCP_CLOSED);
534 sk->sk_err = ECOMM;
535 return 1;
7c657876
ACM
536}
537
538static int dccp_rcv_respond_partopen_state_process(struct sock *sk,
539 struct sk_buff *skb,
540 const struct dccp_hdr *dh,
541 const unsigned len)
542{
543 int queued = 0;
544
545 switch (dh->dccph_type) {
546 case DCCP_PKT_RESET:
547 inet_csk_clear_xmit_timer(sk, ICSK_TIME_DACK);
548 break;
2a9bc9bb
ACM
549 case DCCP_PKT_DATA:
550 if (sk->sk_state == DCCP_RESPOND)
551 break;
7c657876
ACM
552 case DCCP_PKT_DATAACK:
553 case DCCP_PKT_ACK:
554 /*
7690af3f
ACM
555 * FIXME: we should be reseting the PARTOPEN (DELACK) timer
556 * here but only if we haven't used the DELACK timer for
557 * something else, like sending a delayed ack for a TIMESTAMP
558 * echo, etc, for now were not clearing it, sending an extra
559 * ACK when there is nothing else to do in DELACK is not a big
560 * deal after all.
7c657876
ACM
561 */
562
563 /* Stop the PARTOPEN timer */
564 if (sk->sk_state == DCCP_PARTOPEN)
565 inet_csk_clear_xmit_timer(sk, ICSK_TIME_DACK);
566
567 dccp_sk(sk)->dccps_osr = DCCP_SKB_CB(skb)->dccpd_seq;
568 dccp_set_state(sk, DCCP_OPEN);
569
2a9bc9bb
ACM
570 if (dh->dccph_type == DCCP_PKT_DATAACK ||
571 dh->dccph_type == DCCP_PKT_DATA) {
709dd3aa 572 __dccp_rcv_established(sk, skb, dh, len);
7690af3f 573 queued = 1; /* packet was queued
709dd3aa 574 (by __dccp_rcv_established) */
7c657876
ACM
575 }
576 break;
577 }
578
579 return queued;
580}
581
582int dccp_rcv_state_process(struct sock *sk, struct sk_buff *skb,
583 struct dccp_hdr *dh, unsigned len)
584{
585 struct dccp_sock *dp = dccp_sk(sk);
0c10c5d9 586 struct dccp_skb_cb *dcb = DCCP_SKB_CB(skb);
7c657876
ACM
587 const int old_state = sk->sk_state;
588 int queued = 0;
589
8649b0d4
ACM
590 /*
591 * Step 3: Process LISTEN state
8649b0d4
ACM
592 *
593 * If S.state == LISTEN,
d83ca5ac
GR
594 * If P.type == Request or P contains a valid Init Cookie option,
595 * (* Must scan the packet's options to check for Init
596 * Cookies. Only Init Cookies are processed here,
597 * however; other options are processed in Step 8. This
598 * scan need only be performed if the endpoint uses Init
599 * Cookies *)
600 * (* Generate a new socket and switch to that socket *)
601 * Set S := new socket for this port pair
602 * S.state = RESPOND
603 * Choose S.ISS (initial seqno) or set from Init Cookies
604 * Initialize S.GAR := S.ISS
605 * Set S.ISR, S.GSR, S.SWL, S.SWH from packet or Init
606 * Cookies Continue with S.state == RESPOND
607 * (* A Response packet will be generated in Step 11 *)
608 * Otherwise,
609 * Generate Reset(No Connection) unless P.type == Reset
610 * Drop packet and return
8649b0d4
ACM
611 */
612 if (sk->sk_state == DCCP_LISTEN) {
613 if (dh->dccph_type == DCCP_PKT_REQUEST) {
57cca05a
ACM
614 if (inet_csk(sk)->icsk_af_ops->conn_request(sk,
615 skb) < 0)
8649b0d4 616 return 1;
8649b0d4
ACM
617 goto discard;
618 }
619 if (dh->dccph_type == DCCP_PKT_RESET)
620 goto discard;
621
0c10c5d9
ACM
622 /* Caller (dccp_v4_do_rcv) will send Reset */
623 dcb->dccpd_reset_code = DCCP_RESET_CODE_NO_CONNECTION;
8649b0d4
ACM
624 return 1;
625 }
626
991d927c 627 if (sk->sk_state != DCCP_REQUESTING && sk->sk_state != DCCP_RESPOND) {
410e27a4
GR
628 if (dccp_check_seqno(sk, skb))
629 goto discard;
7c657876 630
410e27a4
GR
631 /*
632 * Step 8: Process options and mark acknowledgeable
633 */
634 if (dccp_parse_options(sk, NULL, skb))
635 return 1;
7c657876 636
410e27a4
GR
637 if (dcb->dccpd_ack_seq != DCCP_PKT_WITHOUT_ACK_SEQ)
638 dccp_event_ack_recv(sk, skb);
639
640 if (dccp_msk(sk)->dccpms_send_ack_vector &&
641 dccp_ackvec_add(dp->dccps_hc_rx_ackvec, sk,
642 DCCP_SKB_CB(skb)->dccpd_seq,
643 DCCP_ACKVEC_STATE_RECEIVED))
644 goto discard;
645
646 dccp_deliver_input_to_ccids(sk, skb);
647 }
ddab0556 648
7c657876
ACM
649 /*
650 * Step 9: Process Reset
651 * If P.type == Reset,
652 * Tear down connection
653 * S.state := TIMEWAIT
654 * Set TIMEWAIT timer
655 * Drop packet and return
410e27a4 656 */
7c657876 657 if (dh->dccph_type == DCCP_PKT_RESET) {
d8ef2c29 658 dccp_rcv_reset(sk, skb);
7c657876 659 return 0;
410e27a4
GR
660 /*
661 * Step 7: Check for unexpected packet types
662 * If (S.is_server and P.type == Response)
663 * or (S.is_client and P.type == Request)
664 * or (S.state == RESPOND and P.type == Data),
665 * Send Sync packet acknowledging P.seqno
666 * Drop packet and return
667 */
668 } else if ((dp->dccps_role != DCCP_ROLE_CLIENT &&
669 dh->dccph_type == DCCP_PKT_RESPONSE) ||
670 (dp->dccps_role == DCCP_ROLE_CLIENT &&
671 dh->dccph_type == DCCP_PKT_REQUEST) ||
672 (sk->sk_state == DCCP_RESPOND &&
673 dh->dccph_type == DCCP_PKT_DATA)) {
674 dccp_send_sync(sk, dcb->dccpd_seq, DCCP_PKT_SYNC);
675 goto discard;
676 } else if (dh->dccph_type == DCCP_PKT_CLOSEREQ) {
0c869620
GR
677 if (dccp_rcv_closereq(sk, skb))
678 return 0;
7ad07e7c 679 goto discard;
410e27a4 680 } else if (dh->dccph_type == DCCP_PKT_CLOSE) {
0c869620
GR
681 if (dccp_rcv_close(sk, skb))
682 return 0;
683 goto discard;
7c657876
ACM
684 }
685
686 switch (sk->sk_state) {
410e27a4
GR
687 case DCCP_CLOSED:
688 dcb->dccpd_reset_code = DCCP_RESET_CODE_NO_CONNECTION;
689 return 1;
690
7c657876 691 case DCCP_REQUESTING:
7c657876
ACM
692 queued = dccp_rcv_request_sent_state_process(sk, skb, dh, len);
693 if (queued >= 0)
694 return queued;
695
696 __kfree_skb(skb);
697 return 0;
698
ddab0556 699 case DCCP_RESPOND:
410e27a4 700 case DCCP_PARTOPEN:
7690af3f
ACM
701 queued = dccp_rcv_respond_partopen_state_process(sk, skb,
702 dh, len);
7c657876
ACM
703 break;
704 }
705
7690af3f
ACM
706 if (dh->dccph_type == DCCP_PKT_ACK ||
707 dh->dccph_type == DCCP_PKT_DATAACK) {
7c657876
ACM
708 switch (old_state) {
709 case DCCP_PARTOPEN:
710 sk->sk_state_change(sk);
8d8ad9d7 711 sk_wake_async(sk, SOCK_WAKE_IO, POLL_OUT);
7c657876
ACM
712 break;
713 }
08831700
GR
714 } else if (unlikely(dh->dccph_type == DCCP_PKT_SYNC)) {
715 dccp_send_sync(sk, dcb->dccpd_seq, DCCP_PKT_SYNCACK);
716 goto discard;
7c657876
ACM
717 }
718
8109b02b 719 if (!queued) {
7c657876
ACM
720discard:
721 __kfree_skb(skb);
722 }
723 return 0;
724}
f21e68ca
ACM
725
726EXPORT_SYMBOL_GPL(dccp_rcv_state_process);
4712a792
GR
727
728/**
3393da82
GR
729 * dccp_sample_rtt - Validate and finalise computation of RTT sample
730 * @delta: number of microseconds between packet and acknowledgment
731 * The routine is kept generic to work in different contexts. It should be
732 * called immediately when the ACK used for the RTT sample arrives.
4712a792 733 */
3393da82 734u32 dccp_sample_rtt(struct sock *sk, long delta)
4712a792 735{
3393da82
GR
736 /* dccpor_elapsed_time is either zeroed out or set and > 0 */
737 delta -= dccp_sk(sk)->dccps_options_received.dccpor_elapsed_time * 10;
4712a792 738
410e27a4
GR
739 if (unlikely(delta <= 0)) {
740 DCCP_WARN("unusable RTT sample %ld, using min\n", delta);
741 return DCCP_SANE_RTT_MIN;
742 }
743 if (unlikely(delta > DCCP_SANE_RTT_MAX)) {
744 DCCP_WARN("RTT sample %ld too large, using max\n", delta);
745 return DCCP_SANE_RTT_MAX;
746 }
747
748 return delta;
4712a792
GR
749}
750
751EXPORT_SYMBOL_GPL(dccp_sample_rtt);