Commit | Line | Data |
---|---|---|
2874c5fd | 1 | /* SPDX-License-Identifier: GPL-2.0-or-later */ |
304a1618 ACM |
2 | /* |
3 | * INET An implementation of the TCP/IP protocol suite for the LINUX | |
4 | * operating system. INET is implemented using the BSD Socket | |
5 | * interface as the means of communication with the user level. | |
6 | * | |
7 | * Authors: Lotsa people, from code originally in tcp | |
304a1618 ACM |
8 | */ |
9 | ||
10 | #ifndef _INET_HASHTABLES_H | |
11 | #define _INET_HASHTABLES_H | |
12 | ||
8feaf0c0 | 13 | |
2d8c4ce5 | 14 | #include <linux/interrupt.h> |
9a1f27c4 | 15 | #include <linux/ip.h> |
33b62231 | 16 | #include <linux/ipv6.h> |
77d8bf9c ACM |
17 | #include <linux/list.h> |
18 | #include <linux/slab.h> | |
33b62231 | 19 | #include <linux/socket.h> |
77d8bf9c | 20 | #include <linux/spinlock.h> |
304a1618 | 21 | #include <linux/types.h> |
f3f05f70 | 22 | #include <linux/wait.h> |
304a1618 | 23 | |
463c84b9 | 24 | #include <net/inet_connection_sock.h> |
14c85021 | 25 | #include <net/inet_sock.h> |
28044fc1 | 26 | #include <net/ip.h> |
2d8c4ce5 | 27 | #include <net/sock.h> |
9a1f27c4 | 28 | #include <net/route.h> |
c752f073 | 29 | #include <net/tcp_states.h> |
0b441916 | 30 | #include <net/netns/hash.h> |
2d8c4ce5 | 31 | |
41c6d650 | 32 | #include <linux/refcount.h> |
e48c414e | 33 | #include <asm/byteorder.h> |
f3f05f70 | 34 | |
77d8bf9c | 35 | /* This is for all connections with a full identity, no wildcards. |
05dbc7b5 ED |
36 | * The 'e' prefix stands for Establish, but we really put all sockets |
37 | * but LISTEN ones. | |
77d8bf9c ACM |
38 | */ |
39 | struct inet_ehash_bucket { | |
3ab5aee7 | 40 | struct hlist_nulls_head chain; |
6d255361 | 41 | }; |
77d8bf9c ACM |
42 | |
43 | /* There are a few simple rules, which allow for local port reuse by | |
44 | * an application. In essence: | |
45 | * | |
46 | * 1) Sockets bound to different interfaces may share a local port. | |
47 | * Failing that, goto test 2. | |
48 | * 2) If all sockets have sk->sk_reuse set, and none of them are in | |
49 | * TCP_LISTEN state, the port may be shared. | |
50 | * Failing that, goto test 3. | |
51 | * 3) If all sockets are bound to a specific inet_sk(sk)->rcv_saddr local | |
52 | * address, and none of them are the same, the port may be | |
53 | * shared. | |
54 | * Failing this, the port cannot be shared. | |
55 | * | |
56 | * The interesting point, is test #2. This is what an FTP server does | |
57 | * all day. To optimize this case we use a specific flag bit defined | |
58 | * below. As we add sockets to a bind bucket list, we perform a | |
59 | * check of: (newsk->sk_reuse && (newsk->sk_state != TCP_LISTEN)) | |
60 | * As long as all sockets added to a bind bucket pass this test, | |
61 | * the flag bit will be set. | |
62 | * The resulting situation is that tcp_v[46]_verify_bind() can just check | |
63 | * for this flag bit, if it is set and the socket trying to bind has | |
64 | * sk->sk_reuse set, we don't even have to walk the owners list at all, | |
65 | * we return that it is ok to bind this socket to the requested local port. | |
66 | * | |
67 | * Sounds like a lot of work, but it is worth it. In a more naive | |
68 | * implementation (ie. current FreeBSD etc.) the entire list of ports | |
69 | * must be walked for each data port opened by an ftp server. Needless | |
70 | * to say, this does not scale at all. With a couple thousand FTP | |
71 | * users logged onto your box, isn't it nice to know that new data | |
72 | * ports are created in O(1) time? I thought so. ;-) -DaveM | |
73 | */ | |
637bc8bb JB |
74 | #define FASTREUSEPORT_ANY 1 |
75 | #define FASTREUSEPORT_STRICT 2 | |
76 | ||
77d8bf9c | 77 | struct inet_bind_bucket { |
0c5c9fb5 | 78 | possible_net_t ib_net; |
3c82a21f | 79 | int l3mdev; |
77d8bf9c | 80 | unsigned short port; |
da5e3630 TH |
81 | signed char fastreuse; |
82 | signed char fastreuseport; | |
83 | kuid_t fastuid; | |
637bc8bb JB |
84 | #if IS_ENABLED(CONFIG_IPV6) |
85 | struct in6_addr fast_v6_rcv_saddr; | |
86 | #endif | |
87 | __be32 fast_rcv_saddr; | |
88 | unsigned short fast_sk_family; | |
89 | bool fast_ipv6_only; | |
77d8bf9c ACM |
90 | struct hlist_node node; |
91 | struct hlist_head owners; | |
92 | }; | |
93 | ||
28044fc1 JK |
94 | struct inet_bind2_bucket { |
95 | possible_net_t ib_net; | |
96 | int l3mdev; | |
97 | unsigned short port; | |
5456262d MKL |
98 | #if IS_ENABLED(CONFIG_IPV6) |
99 | unsigned short family; | |
100 | #endif | |
28044fc1 JK |
101 | union { |
102 | #if IS_ENABLED(CONFIG_IPV6) | |
103 | struct in6_addr v6_rcv_saddr; | |
104 | #endif | |
105 | __be32 rcv_saddr; | |
106 | }; | |
107 | /* Node in the bhash2 inet_bind_hashbucket chain */ | |
108 | struct hlist_node node; | |
109 | /* List of sockets hashed to this bucket */ | |
110 | struct hlist_head owners; | |
936a192f KI |
111 | /* bhash has twsk in owners, but bhash2 has twsk in |
112 | * deathrow not to add a member in struct sock_common. | |
113 | */ | |
114 | struct hlist_head deathrow; | |
28044fc1 JK |
115 | }; |
116 | ||
117 | static inline struct net *ib_net(const struct inet_bind_bucket *ib) | |
118 | { | |
119 | return read_pnet(&ib->ib_net); | |
120 | } | |
121 | ||
122 | static inline struct net *ib2_net(const struct inet_bind2_bucket *ib) | |
7a9546ee ED |
123 | { |
124 | return read_pnet(&ib->ib_net); | |
125 | } | |
126 | ||
b67bfe0d SL |
127 | #define inet_bind_bucket_for_each(tb, head) \ |
128 | hlist_for_each_entry(tb, head, node) | |
77d8bf9c ACM |
129 | |
130 | struct inet_bind_hashbucket { | |
131 | spinlock_t lock; | |
132 | struct hlist_head chain; | |
133 | }; | |
134 | ||
8dbd76e7 ED |
135 | /* Sockets can be hashed in established or listening table. |
136 | * We must use different 'nulls' end-of-chain value for all hash buckets : | |
137 | * A socket might transition from ESTABLISH to LISTEN state without | |
138 | * RCU grace period. A lookup in ehash table needs to handle this case. | |
c25eb3bf | 139 | */ |
8dbd76e7 | 140 | #define LISTENING_NULLS_BASE (1U << 29) |
5caea4ea ED |
141 | struct inet_listen_hashbucket { |
142 | spinlock_t lock; | |
cae3873c | 143 | struct hlist_nulls_head nulls_head; |
5caea4ea ED |
144 | }; |
145 | ||
77d8bf9c ACM |
146 | /* This is for listening sockets, thus all sockets which possess wildcards. */ |
147 | #define INET_LHTABLE_SIZE 32 /* Yes, really, this is all you need. */ | |
148 | ||
149 | struct inet_hashinfo { | |
150 | /* This is for sockets with full identity only. Sockets here will | |
151 | * always be without wildcards and will have the following invariant: | |
152 | * | |
153 | * TCP_ESTABLISHED <= sk->sk_state < TCP_CLOSE | |
154 | * | |
77d8bf9c ACM |
155 | */ |
156 | struct inet_ehash_bucket *ehash; | |
9db66bdc | 157 | spinlock_t *ehash_locks; |
f373b53b | 158 | unsigned int ehash_mask; |
230140cf | 159 | unsigned int ehash_locks_mask; |
77d8bf9c ACM |
160 | |
161 | /* Ok, let's try this, I give up, we do need a local binding | |
162 | * TCP hash as well as the others for fast bind/connect. | |
163 | */ | |
61b7c691 | 164 | struct kmem_cache *bind_bucket_cachep; |
28044fc1 | 165 | /* This bind table is hashed by local port */ |
77d8bf9c | 166 | struct inet_bind_hashbucket *bhash; |
28044fc1 JK |
167 | struct kmem_cache *bind2_bucket_cachep; |
168 | /* This bind table is hashed by local port and sk->sk_rcv_saddr (ipv4) | |
169 | * or sk->sk_v6_rcv_saddr (ipv6). This 2nd bind table is used | |
170 | * primarily for expediting bind conflict resolution. | |
171 | */ | |
172 | struct inet_bind_hashbucket *bhash2; | |
cfcabdcc | 173 | unsigned int bhash_size; |
77d8bf9c | 174 | |
61b7c691 MKL |
175 | /* The 2nd listener table hashed by local port and address */ |
176 | unsigned int lhash2_mask; | |
177 | struct inet_listen_hashbucket *lhash2; | |
d1e5e640 KI |
178 | |
179 | bool pernet; | |
77d8bf9c ACM |
180 | }; |
181 | ||
429e42c1 KI |
182 | static inline struct inet_hashinfo *tcp_or_dccp_get_hashinfo(const struct sock *sk) |
183 | { | |
184 | #if IS_ENABLED(CONFIG_IP_DCCP) | |
185 | return sk->sk_prot->h.hashinfo ? : | |
186 | sock_net(sk)->ipv4.tcp_death_row.hashinfo; | |
187 | #else | |
188 | return sock_net(sk)->ipv4.tcp_death_row.hashinfo; | |
189 | #endif | |
190 | } | |
191 | ||
61b7c691 MKL |
192 | static inline struct inet_listen_hashbucket * |
193 | inet_lhash2_bucket(struct inet_hashinfo *h, u32 hash) | |
194 | { | |
195 | return &h->lhash2[hash & h->lhash2_mask]; | |
196 | } | |
197 | ||
81c3d547 ED |
198 | static inline struct inet_ehash_bucket *inet_ehash_bucket( |
199 | struct inet_hashinfo *hashinfo, | |
200 | unsigned int hash) | |
201 | { | |
f373b53b | 202 | return &hashinfo->ehash[hash & hashinfo->ehash_mask]; |
304a1618 ACM |
203 | } |
204 | ||
9db66bdc | 205 | static inline spinlock_t *inet_ehash_lockp( |
230140cf ED |
206 | struct inet_hashinfo *hashinfo, |
207 | unsigned int hash) | |
208 | { | |
209 | return &hashinfo->ehash_locks[hash & hashinfo->ehash_locks_mask]; | |
210 | } | |
211 | ||
095dc8e0 | 212 | int inet_ehash_locks_alloc(struct inet_hashinfo *hashinfo); |
230140cf | 213 | |
c96b6acc WH |
214 | static inline void inet_hashinfo2_free_mod(struct inet_hashinfo *h) |
215 | { | |
216 | kfree(h->lhash2); | |
217 | h->lhash2 = NULL; | |
218 | } | |
219 | ||
230140cf ED |
220 | static inline void inet_ehash_locks_free(struct inet_hashinfo *hashinfo) |
221 | { | |
095dc8e0 ED |
222 | kvfree(hashinfo->ehash_locks); |
223 | hashinfo->ehash_locks = NULL; | |
230140cf ED |
224 | } |
225 | ||
d1e5e640 KI |
226 | struct inet_hashinfo *inet_pernet_hashinfo_alloc(struct inet_hashinfo *hashinfo, |
227 | unsigned int ehash_entries); | |
228 | void inet_pernet_hashinfo_free(struct inet_hashinfo *hashinfo); | |
229 | ||
1fd51155 JP |
230 | struct inet_bind_bucket * |
231 | inet_bind_bucket_create(struct kmem_cache *cachep, struct net *net, | |
232 | struct inet_bind_hashbucket *head, | |
3c82a21f | 233 | const unsigned short snum, int l3mdev); |
1fd51155 JP |
234 | void inet_bind_bucket_destroy(struct kmem_cache *cachep, |
235 | struct inet_bind_bucket *tb); | |
236 | ||
28044fc1 JK |
237 | bool inet_bind_bucket_match(const struct inet_bind_bucket *tb, |
238 | const struct net *net, unsigned short port, | |
239 | int l3mdev); | |
240 | ||
241 | struct inet_bind2_bucket * | |
242 | inet_bind2_bucket_create(struct kmem_cache *cachep, struct net *net, | |
243 | struct inet_bind_hashbucket *head, | |
244 | unsigned short port, int l3mdev, | |
245 | const struct sock *sk); | |
246 | ||
247 | void inet_bind2_bucket_destroy(struct kmem_cache *cachep, | |
248 | struct inet_bind2_bucket *tb); | |
249 | ||
250 | struct inet_bind2_bucket * | |
251 | inet_bind2_bucket_find(const struct inet_bind_hashbucket *head, | |
252 | const struct net *net, | |
253 | unsigned short port, int l3mdev, | |
254 | const struct sock *sk); | |
255 | ||
256 | bool inet_bind2_bucket_match_addr_any(const struct inet_bind2_bucket *tb, | |
257 | const struct net *net, unsigned short port, | |
258 | int l3mdev, const struct sock *sk); | |
259 | ||
6eada011 ED |
260 | static inline u32 inet_bhashfn(const struct net *net, const __u16 lport, |
261 | const u32 bhash_size) | |
77d8bf9c | 262 | { |
0b441916 | 263 | return (lport + net_hash_mix(net)) & (bhash_size - 1); |
77d8bf9c ACM |
264 | } |
265 | ||
28044fc1 JK |
266 | static inline struct inet_bind_hashbucket * |
267 | inet_bhashfn_portaddr(const struct inet_hashinfo *hinfo, const struct sock *sk, | |
268 | const struct net *net, unsigned short port) | |
269 | { | |
270 | u32 hash; | |
271 | ||
272 | #if IS_ENABLED(CONFIG_IPV6) | |
273 | if (sk->sk_family == AF_INET6) | |
274 | hash = ipv6_portaddr_hash(net, &sk->sk_v6_rcv_saddr, port); | |
275 | else | |
276 | #endif | |
277 | hash = ipv4_portaddr_hash(net, sk->sk_rcv_saddr, port); | |
278 | return &hinfo->bhash2[hash & (hinfo->bhash_size - 1)]; | |
279 | } | |
280 | ||
281 | struct inet_bind_hashbucket * | |
282 | inet_bhash2_addr_any_hashbucket(const struct sock *sk, const struct net *net, int port); | |
283 | ||
284 | /* This should be called whenever a socket's sk_rcv_saddr (ipv4) or | |
285 | * sk_v6_rcv_saddr (ipv6) changes after it has been binded. The socket's | |
286 | * rcv_saddr field should already have been updated when this is called. | |
287 | */ | |
8c5dae4c | 288 | int inet_bhash2_update_saddr(struct sock *sk, void *saddr, int family); |
e0833d1f | 289 | void inet_bhash2_reset_saddr(struct sock *sk); |
28044fc1 | 290 | |
1fd51155 | 291 | void inet_bind_hash(struct sock *sk, struct inet_bind_bucket *tb, |
28044fc1 | 292 | struct inet_bind2_bucket *tb2, unsigned short port); |
2d8c4ce5 | 293 | |
2d8c4ce5 | 294 | /* Caller must disable local BH processing. */ |
1ce31c9e | 295 | int __inet_inherit_port(const struct sock *sk, struct sock *child); |
2d8c4ce5 | 296 | |
1fd51155 | 297 | void inet_put_port(struct sock *sk); |
2d8c4ce5 | 298 | |
61b7c691 MKL |
299 | void inet_hashinfo2_init(struct inet_hashinfo *h, const char *name, |
300 | unsigned long numentries, int scale, | |
301 | unsigned long low_limit, | |
302 | unsigned long high_limit); | |
c92c81df | 303 | int inet_hashinfo2_init_mod(struct inet_hashinfo *h); |
f3f05f70 | 304 | |
01770a16 RD |
305 | bool inet_ehash_insert(struct sock *sk, struct sock *osk, bool *found_dup_sk); |
306 | bool inet_ehash_nolisten(struct sock *sk, struct sock *osk, | |
307 | bool *found_dup_sk); | |
fe38d2a1 | 308 | int __inet_hash(struct sock *sk, struct sock *osk); |
086c653f | 309 | int inet_hash(struct sock *sk); |
1fd51155 | 310 | void inet_unhash(struct sock *sk); |
33b62231 | 311 | |
1fd51155 JP |
312 | struct sock *__inet_lookup_listener(struct net *net, |
313 | struct inet_hashinfo *hashinfo, | |
a583636a | 314 | struct sk_buff *skb, int doff, |
1fd51155 JP |
315 | const __be32 saddr, const __be16 sport, |
316 | const __be32 daddr, | |
317 | const unsigned short hnum, | |
3fa6f616 | 318 | const int dif, const int sdif); |
8f491069 | 319 | |
c67499c0 PE |
320 | static inline struct sock *inet_lookup_listener(struct net *net, |
321 | struct inet_hashinfo *hashinfo, | |
a583636a | 322 | struct sk_buff *skb, int doff, |
da5e3630 | 323 | __be32 saddr, __be16 sport, |
3fa6f616 | 324 | __be32 daddr, __be16 dport, int dif, int sdif) |
8f491069 | 325 | { |
a583636a | 326 | return __inet_lookup_listener(net, hashinfo, skb, doff, saddr, sport, |
3fa6f616 | 327 | daddr, ntohs(dport), dif, sdif); |
8f491069 | 328 | } |
8feaf0c0 ACM |
329 | |
330 | /* Socket demux engine toys. */ | |
4f765d84 AV |
331 | /* What happens here is ugly; there's a pair of adjacent fields in |
332 | struct inet_sock; __be16 dport followed by __u16 num. We want to | |
333 | search by pair, so we combine the keys into a single 32bit value | |
334 | and compare with 32bit value read from &...->dport. Let's at least | |
335 | make sure that it's not mixed with anything else... | |
336 | On 64bit targets we combine comparisons with pair of adjacent __be32 | |
337 | fields in the same way. | |
338 | */ | |
8feaf0c0 ACM |
339 | #ifdef __BIG_ENDIAN |
340 | #define INET_COMBINED_PORTS(__sport, __dport) \ | |
4f765d84 | 341 | ((__force __portpair)(((__force __u32)(__be16)(__sport) << 16) | (__u32)(__dport))) |
8feaf0c0 ACM |
342 | #else /* __LITTLE_ENDIAN */ |
343 | #define INET_COMBINED_PORTS(__sport, __dport) \ | |
4f765d84 | 344 | ((__force __portpair)(((__u32)(__dport) << 16) | (__force __u32)(__be16)(__sport))) |
8feaf0c0 ACM |
345 | #endif |
346 | ||
8feaf0c0 ACM |
347 | #ifdef __BIG_ENDIAN |
348 | #define INET_ADDR_COOKIE(__name, __saddr, __daddr) \ | |
4f765d84 AV |
349 | const __addrpair __name = (__force __addrpair) ( \ |
350 | (((__force __u64)(__be32)(__saddr)) << 32) | \ | |
c7228317 | 351 | ((__force __u64)(__be32)(__daddr))) |
8feaf0c0 ACM |
352 | #else /* __LITTLE_ENDIAN */ |
353 | #define INET_ADDR_COOKIE(__name, __saddr, __daddr) \ | |
4f765d84 AV |
354 | const __addrpair __name = (__force __addrpair) ( \ |
355 | (((__force __u64)(__be32)(__daddr)) << 32) | \ | |
c7228317 | 356 | ((__force __u64)(__be32)(__saddr))) |
8feaf0c0 | 357 | #endif /* __BIG_ENDIAN */ |
4915d50e | 358 | |
eda090c3 | 359 | static inline bool inet_match(struct net *net, const struct sock *sk, |
4915d50e ED |
360 | const __addrpair cookie, const __portpair ports, |
361 | int dif, int sdif) | |
362 | { | |
4915d50e ED |
363 | if (!net_eq(sock_net(sk), net) || |
364 | sk->sk_portpair != ports || | |
365 | sk->sk_addrpair != cookie) | |
366 | return false; | |
367 | ||
944fd1ae MM |
368 | /* READ_ONCE() paired with WRITE_ONCE() in sock_bindtoindex_locked() */ |
369 | return inet_sk_bound_dev_eq(net, READ_ONCE(sk->sk_bound_dev_if), dif, | |
370 | sdif); | |
4915d50e | 371 | } |
e48c414e | 372 | |
ee3cf32a | 373 | /* Sockets in TCP_CLOSE state are _always_ taken out of the hash, so we need |
e48c414e | 374 | * not check it for lookups anymore, thanks Alexey. -DaveM |
e48c414e | 375 | */ |
1fd51155 JP |
376 | struct sock *__inet_lookup_established(struct net *net, |
377 | struct inet_hashinfo *hashinfo, | |
378 | const __be32 saddr, const __be16 sport, | |
379 | const __be32 daddr, const u16 hnum, | |
3fa6f616 | 380 | const int dif, const int sdif); |
e48c414e | 381 | |
8f491069 | 382 | static inline struct sock * |
c67499c0 | 383 | inet_lookup_established(struct net *net, struct inet_hashinfo *hashinfo, |
fb99c848 AV |
384 | const __be32 saddr, const __be16 sport, |
385 | const __be32 daddr, const __be16 dport, | |
8f491069 HX |
386 | const int dif) |
387 | { | |
c67499c0 | 388 | return __inet_lookup_established(net, hashinfo, saddr, sport, daddr, |
3fa6f616 | 389 | ntohs(dport), dif, 0); |
8f491069 HX |
390 | } |
391 | ||
c67499c0 PE |
392 | static inline struct sock *__inet_lookup(struct net *net, |
393 | struct inet_hashinfo *hashinfo, | |
a583636a | 394 | struct sk_buff *skb, int doff, |
fb99c848 AV |
395 | const __be32 saddr, const __be16 sport, |
396 | const __be32 daddr, const __be16 dport, | |
3fa6f616 | 397 | const int dif, const int sdif, |
3b24d854 | 398 | bool *refcounted) |
e48c414e | 399 | { |
8f491069 | 400 | u16 hnum = ntohs(dport); |
3b24d854 | 401 | struct sock *sk; |
c67499c0 | 402 | |
3b24d854 | 403 | sk = __inet_lookup_established(net, hashinfo, saddr, sport, |
3fa6f616 | 404 | daddr, hnum, dif, sdif); |
3b24d854 ED |
405 | *refcounted = true; |
406 | if (sk) | |
407 | return sk; | |
408 | *refcounted = false; | |
409 | return __inet_lookup_listener(net, hashinfo, skb, doff, saddr, | |
3fa6f616 | 410 | sport, daddr, hnum, dif, sdif); |
e48c414e ACM |
411 | } |
412 | ||
c67499c0 PE |
413 | static inline struct sock *inet_lookup(struct net *net, |
414 | struct inet_hashinfo *hashinfo, | |
a583636a | 415 | struct sk_buff *skb, int doff, |
fb99c848 AV |
416 | const __be32 saddr, const __be16 sport, |
417 | const __be32 daddr, const __be16 dport, | |
e48c414e ACM |
418 | const int dif) |
419 | { | |
420 | struct sock *sk; | |
3b24d854 | 421 | bool refcounted; |
e48c414e | 422 | |
a583636a | 423 | sk = __inet_lookup(net, hashinfo, skb, doff, saddr, sport, daddr, |
3fa6f616 | 424 | dport, dif, 0, &refcounted); |
e48c414e | 425 | |
41c6d650 | 426 | if (sk && !refcounted && !refcount_inc_not_zero(&sk->sk_refcnt)) |
3b24d854 | 427 | sk = NULL; |
e48c414e ACM |
428 | return sk; |
429 | } | |
a7f5e7f1 | 430 | |
9a1f27c4 ACM |
431 | static inline struct sock *__inet_lookup_skb(struct inet_hashinfo *hashinfo, |
432 | struct sk_buff *skb, | |
a583636a | 433 | int doff, |
9a1f27c4 | 434 | const __be16 sport, |
3b24d854 | 435 | const __be16 dport, |
3fa6f616 | 436 | const int sdif, |
3b24d854 | 437 | bool *refcounted) |
9a1f27c4 | 438 | { |
71489e21 | 439 | struct sock *sk = skb_steal_sock(skb, refcounted); |
9a1f27c4 ACM |
440 | const struct iphdr *iph = ip_hdr(skb); |
441 | ||
41063e9d | 442 | if (sk) |
23542618 | 443 | return sk; |
3b24d854 ED |
444 | |
445 | return __inet_lookup(dev_net(skb_dst(skb)->dev), hashinfo, skb, | |
446 | doff, iph->saddr, sport, | |
3fa6f616 | 447 | iph->daddr, dport, inet_iif(skb), sdif, |
3b24d854 | 448 | refcounted); |
9a1f27c4 ACM |
449 | } |
450 | ||
d1e559d0 ED |
451 | u32 inet6_ehashfn(const struct net *net, |
452 | const struct in6_addr *laddr, const u16 lport, | |
453 | const struct in6_addr *faddr, const __be16 fport); | |
454 | ||
455 | static inline void sk_daddr_set(struct sock *sk, __be32 addr) | |
456 | { | |
457 | sk->sk_daddr = addr; /* alias of inet_daddr */ | |
458 | #if IS_ENABLED(CONFIG_IPV6) | |
459 | ipv6_addr_set_v4mapped(addr, &sk->sk_v6_daddr); | |
460 | #endif | |
461 | } | |
462 | ||
463 | static inline void sk_rcv_saddr_set(struct sock *sk, __be32 addr) | |
464 | { | |
465 | sk->sk_rcv_saddr = addr; /* alias of inet_rcv_saddr */ | |
466 | #if IS_ENABLED(CONFIG_IPV6) | |
467 | ipv6_addr_set_v4mapped(addr, &sk->sk_v6_rcv_saddr); | |
468 | #endif | |
469 | } | |
5b441f76 | 470 | |
1fd51155 | 471 | int __inet_hash_connect(struct inet_timewait_death_row *death_row, |
b2d05756 | 472 | struct sock *sk, u64 port_offset, |
1fd51155 JP |
473 | int (*check_established)(struct inet_timewait_death_row *, |
474 | struct sock *, __u16, | |
b4d6444e | 475 | struct inet_timewait_sock **)); |
9327f705 | 476 | |
1fd51155 JP |
477 | int inet_hash_connect(struct inet_timewait_death_row *death_row, |
478 | struct sock *sk); | |
304a1618 | 479 | #endif /* _INET_HASHTABLES_H */ |