You've already forked linux-apfs
mirror of
https://github.com/linux-apfs/linux-apfs.git
synced 2026-05-01 15:00:59 -07:00
[SK_BUFF]: Introduce tcp_hdr(), remove skb->h.th
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com> Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
committed by
David S. Miller
parent
ab6a5bb6b2
commit
aa8223c7bb
@@ -1352,8 +1352,8 @@ void ip_send_reply(struct sock *sk, struct sk_buff *skb, struct ip_reply_arg *ar
|
||||
.tos = RT_TOS(ip_hdr(skb)->tos) } },
|
||||
/* Not quite clean, but right. */
|
||||
.uli_u = { .ports =
|
||||
{ .sport = skb->h.th->dest,
|
||||
.dport = skb->h.th->source } },
|
||||
{ .sport = tcp_hdr(skb)->dest,
|
||||
.dport = tcp_hdr(skb)->source } },
|
||||
.proto = sk->sk_protocol };
|
||||
security_skb_classify_flow(skb, &fl);
|
||||
if (ip_route_output_key(&rt, &fl))
|
||||
|
||||
+18
-18
@@ -125,10 +125,11 @@ static __u16 const msstab[] = {
|
||||
__u32 cookie_v4_init_sequence(struct sock *sk, struct sk_buff *skb, __u16 *mssp)
|
||||
{
|
||||
struct tcp_sock *tp = tcp_sk(sk);
|
||||
const struct iphdr *iph = ip_hdr(skb);
|
||||
const struct tcphdr *th = tcp_hdr(skb);
|
||||
int mssind;
|
||||
const __u16 mss = *mssp;
|
||||
|
||||
|
||||
tp->last_synq_overflow = jiffies;
|
||||
|
||||
/* XXX sort msstab[] by probability? Binary search? */
|
||||
@@ -138,9 +139,8 @@ __u32 cookie_v4_init_sequence(struct sock *sk, struct sk_buff *skb, __u16 *mssp)
|
||||
|
||||
NET_INC_STATS_BH(LINUX_MIB_SYNCOOKIESSENT);
|
||||
|
||||
return secure_tcp_syn_cookie(ip_hdr(skb)->saddr, ip_hdr(skb)->daddr,
|
||||
skb->h.th->source, skb->h.th->dest,
|
||||
ntohl(skb->h.th->seq),
|
||||
return secure_tcp_syn_cookie(iph->saddr, iph->daddr,
|
||||
th->source, th->dest, ntohl(th->seq),
|
||||
jiffies / (HZ * 60), mssind);
|
||||
}
|
||||
|
||||
@@ -157,14 +157,13 @@ __u32 cookie_v4_init_sequence(struct sock *sk, struct sk_buff *skb, __u16 *mssp)
|
||||
*/
|
||||
static inline int cookie_check(struct sk_buff *skb, __u32 cookie)
|
||||
{
|
||||
__u32 seq;
|
||||
__u32 mssind;
|
||||
|
||||
seq = ntohl(skb->h.th->seq)-1;
|
||||
mssind = check_tcp_syn_cookie(cookie,
|
||||
ip_hdr(skb)->saddr, ip_hdr(skb)->daddr,
|
||||
skb->h.th->source, skb->h.th->dest,
|
||||
seq, jiffies / (HZ * 60), COUNTER_TRIES);
|
||||
const struct iphdr *iph = ip_hdr(skb);
|
||||
const struct tcphdr *th = tcp_hdr(skb);
|
||||
__u32 seq = ntohl(th->seq) - 1;
|
||||
__u32 mssind = check_tcp_syn_cookie(cookie, iph->saddr, iph->daddr,
|
||||
th->source, th->dest, seq,
|
||||
jiffies / (HZ * 60),
|
||||
COUNTER_TRIES);
|
||||
|
||||
return mssind < NUM_MSS ? msstab[mssind] + 1 : 0;
|
||||
}
|
||||
@@ -191,14 +190,15 @@ struct sock *cookie_v4_check(struct sock *sk, struct sk_buff *skb,
|
||||
struct inet_request_sock *ireq;
|
||||
struct tcp_request_sock *treq;
|
||||
struct tcp_sock *tp = tcp_sk(sk);
|
||||
__u32 cookie = ntohl(skb->h.th->ack_seq) - 1;
|
||||
const struct tcphdr *th = tcp_hdr(skb);
|
||||
__u32 cookie = ntohl(th->ack_seq) - 1;
|
||||
struct sock *ret = sk;
|
||||
struct request_sock *req;
|
||||
int mss;
|
||||
struct rtable *rt;
|
||||
__u8 rcv_wscale;
|
||||
|
||||
if (!sysctl_tcp_syncookies || !skb->h.th->ack)
|
||||
if (!sysctl_tcp_syncookies || !th->ack)
|
||||
goto out;
|
||||
|
||||
if (time_after(jiffies, tp->last_synq_overflow + TCP_TIMEOUT_INIT) ||
|
||||
@@ -220,10 +220,10 @@ struct sock *cookie_v4_check(struct sock *sk, struct sk_buff *skb,
|
||||
}
|
||||
ireq = inet_rsk(req);
|
||||
treq = tcp_rsk(req);
|
||||
treq->rcv_isn = ntohl(skb->h.th->seq) - 1;
|
||||
treq->rcv_isn = ntohl(th->seq) - 1;
|
||||
treq->snt_isn = cookie;
|
||||
req->mss = mss;
|
||||
ireq->rmt_port = skb->h.th->source;
|
||||
ireq->rmt_port = th->source;
|
||||
ireq->loc_addr = ip_hdr(skb)->daddr;
|
||||
ireq->rmt_addr = ip_hdr(skb)->saddr;
|
||||
ireq->opt = NULL;
|
||||
@@ -261,8 +261,8 @@ struct sock *cookie_v4_check(struct sock *sk, struct sk_buff *skb,
|
||||
.tos = RT_CONN_FLAGS(sk) } },
|
||||
.proto = IPPROTO_TCP,
|
||||
.uli_u = { .ports =
|
||||
{ .sport = skb->h.th->dest,
|
||||
.dport = skb->h.th->source } } };
|
||||
{ .sport = th->dest,
|
||||
.dport = th->source } } };
|
||||
security_req_classify_flow(req, &fl);
|
||||
if (ip_route_output_key(&rt, &fl)) {
|
||||
reqsk_free(req);
|
||||
|
||||
+11
-11
@@ -425,7 +425,7 @@ int tcp_ioctl(struct sock *sk, int cmd, unsigned long arg)
|
||||
/* Subtract 1, if FIN is in queue. */
|
||||
if (answ && !skb_queue_empty(&sk->sk_receive_queue))
|
||||
answ -=
|
||||
((struct sk_buff *)sk->sk_receive_queue.prev)->h.th->fin;
|
||||
tcp_hdr((struct sk_buff *)sk->sk_receive_queue.prev)->fin;
|
||||
} else
|
||||
answ = tp->urg_seq - tp->copied_seq;
|
||||
release_sock(sk);
|
||||
@@ -1016,9 +1016,9 @@ static inline struct sk_buff *tcp_recv_skb(struct sock *sk, u32 seq, u32 *off)
|
||||
|
||||
skb_queue_walk(&sk->sk_receive_queue, skb) {
|
||||
offset = seq - TCP_SKB_CB(skb)->seq;
|
||||
if (skb->h.th->syn)
|
||||
if (tcp_hdr(skb)->syn)
|
||||
offset--;
|
||||
if (offset < skb->len || skb->h.th->fin) {
|
||||
if (offset < skb->len || tcp_hdr(skb)->fin) {
|
||||
*off = offset;
|
||||
return skb;
|
||||
}
|
||||
@@ -1070,7 +1070,7 @@ int tcp_read_sock(struct sock *sk, read_descriptor_t *desc,
|
||||
if (offset != skb->len)
|
||||
break;
|
||||
}
|
||||
if (skb->h.th->fin) {
|
||||
if (tcp_hdr(skb)->fin) {
|
||||
sk_eat_skb(sk, skb, 0);
|
||||
++seq;
|
||||
break;
|
||||
@@ -1174,11 +1174,11 @@ int tcp_recvmsg(struct kiocb *iocb, struct sock *sk, struct msghdr *msg,
|
||||
break;
|
||||
}
|
||||
offset = *seq - TCP_SKB_CB(skb)->seq;
|
||||
if (skb->h.th->syn)
|
||||
if (tcp_hdr(skb)->syn)
|
||||
offset--;
|
||||
if (offset < skb->len)
|
||||
goto found_ok_skb;
|
||||
if (skb->h.th->fin)
|
||||
if (tcp_hdr(skb)->fin)
|
||||
goto found_fin_ok;
|
||||
BUG_TRAP(flags & MSG_PEEK);
|
||||
skb = skb->next;
|
||||
@@ -1394,7 +1394,7 @@ skip_copy:
|
||||
if (used + offset < skb->len)
|
||||
continue;
|
||||
|
||||
if (skb->h.th->fin)
|
||||
if (tcp_hdr(skb)->fin)
|
||||
goto found_fin_ok;
|
||||
if (!(flags & MSG_PEEK)) {
|
||||
sk_eat_skb(sk, skb, copied_early);
|
||||
@@ -1563,7 +1563,7 @@ void tcp_close(struct sock *sk, long timeout)
|
||||
*/
|
||||
while ((skb = __skb_dequeue(&sk->sk_receive_queue)) != NULL) {
|
||||
u32 len = TCP_SKB_CB(skb)->end_seq - TCP_SKB_CB(skb)->seq -
|
||||
skb->h.th->fin;
|
||||
tcp_hdr(skb)->fin;
|
||||
data_was_unread += len;
|
||||
__kfree_skb(skb);
|
||||
}
|
||||
@@ -2170,7 +2170,7 @@ struct sk_buff *tcp_tso_segment(struct sk_buff *skb, int features)
|
||||
if (!pskb_may_pull(skb, sizeof(*th)))
|
||||
goto out;
|
||||
|
||||
th = skb->h.th;
|
||||
th = tcp_hdr(skb);
|
||||
thlen = th->doff * 4;
|
||||
if (thlen < sizeof(*th))
|
||||
goto out;
|
||||
@@ -2210,7 +2210,7 @@ struct sk_buff *tcp_tso_segment(struct sk_buff *skb, int features)
|
||||
delta = htonl(oldlen + (thlen + len));
|
||||
|
||||
skb = segs;
|
||||
th = skb->h.th;
|
||||
th = tcp_hdr(skb);
|
||||
seq = ntohl(th->seq);
|
||||
|
||||
do {
|
||||
@@ -2224,7 +2224,7 @@ struct sk_buff *tcp_tso_segment(struct sk_buff *skb, int features)
|
||||
|
||||
seq += len;
|
||||
skb = skb->next;
|
||||
th = skb->h.th;
|
||||
th = tcp_hdr(skb);
|
||||
|
||||
th->seq = htonl(seq);
|
||||
th->cwr = 0;
|
||||
|
||||
+15
-13
@@ -148,7 +148,7 @@ static void tcp_measure_rcv_mss(struct sock *sk,
|
||||
* to handle super-low mtu links fairly.
|
||||
*/
|
||||
(len >= TCP_MIN_MSS + sizeof(struct tcphdr) &&
|
||||
!(tcp_flag_word(skb->h.th)&TCP_REMNANT))) {
|
||||
!(tcp_flag_word(tcp_hdr(skb)) & TCP_REMNANT))) {
|
||||
/* Subtract also invariant (if peer is RFC compliant),
|
||||
* tcp header plus fixed timestamp option length.
|
||||
* Resulting "len" is MSS free of SACK jitter.
|
||||
@@ -2559,9 +2559,9 @@ static int tcp_ack_update_window(struct sock *sk, struct tcp_sock *tp,
|
||||
struct sk_buff *skb, u32 ack, u32 ack_seq)
|
||||
{
|
||||
int flag = 0;
|
||||
u32 nwin = ntohs(skb->h.th->window);
|
||||
u32 nwin = ntohs(tcp_hdr(skb)->window);
|
||||
|
||||
if (likely(!skb->h.th->syn))
|
||||
if (likely(!tcp_hdr(skb)->syn))
|
||||
nwin <<= tp->rx_opt.snd_wscale;
|
||||
|
||||
if (tcp_may_update_window(tp, ack, ack_seq, nwin)) {
|
||||
@@ -2766,7 +2766,7 @@ static int tcp_ack(struct sock *sk, struct sk_buff *skb, int flag)
|
||||
if (TCP_SKB_CB(skb)->sacked)
|
||||
flag |= tcp_sacktag_write_queue(sk, skb, prior_snd_una);
|
||||
|
||||
if (TCP_ECN_rcv_ecn_echo(tp, skb->h.th))
|
||||
if (TCP_ECN_rcv_ecn_echo(tp, tcp_hdr(skb)))
|
||||
flag |= FLAG_ECE;
|
||||
|
||||
tcp_ca_event(sk, CA_EVENT_SLOW_ACK);
|
||||
@@ -2833,7 +2833,7 @@ uninteresting_ack:
|
||||
void tcp_parse_options(struct sk_buff *skb, struct tcp_options_received *opt_rx, int estab)
|
||||
{
|
||||
unsigned char *ptr;
|
||||
struct tcphdr *th = skb->h.th;
|
||||
struct tcphdr *th = tcp_hdr(skb);
|
||||
int length=(th->doff*4)-sizeof(struct tcphdr);
|
||||
|
||||
ptr = (unsigned char *)(th + 1);
|
||||
@@ -2995,7 +2995,7 @@ static inline void tcp_replace_ts_recent(struct tcp_sock *tp, u32 seq)
|
||||
static int tcp_disordered_ack(const struct sock *sk, const struct sk_buff *skb)
|
||||
{
|
||||
struct tcp_sock *tp = tcp_sk(sk);
|
||||
struct tcphdr *th = skb->h.th;
|
||||
struct tcphdr *th = tcp_hdr(skb);
|
||||
u32 seq = TCP_SKB_CB(skb)->seq;
|
||||
u32 ack = TCP_SKB_CB(skb)->ack_seq;
|
||||
|
||||
@@ -3357,8 +3357,8 @@ static void tcp_ofo_queue(struct sock *sk)
|
||||
__skb_unlink(skb, &tp->out_of_order_queue);
|
||||
__skb_queue_tail(&sk->sk_receive_queue, skb);
|
||||
tp->rcv_nxt = TCP_SKB_CB(skb)->end_seq;
|
||||
if (skb->h.th->fin)
|
||||
tcp_fin(skb, sk, skb->h.th);
|
||||
if (tcp_hdr(skb)->fin)
|
||||
tcp_fin(skb, sk, tcp_hdr(skb));
|
||||
}
|
||||
}
|
||||
|
||||
@@ -3366,7 +3366,7 @@ static int tcp_prune_queue(struct sock *sk);
|
||||
|
||||
static void tcp_data_queue(struct sock *sk, struct sk_buff *skb)
|
||||
{
|
||||
struct tcphdr *th = skb->h.th;
|
||||
struct tcphdr *th = tcp_hdr(skb);
|
||||
struct tcp_sock *tp = tcp_sk(sk);
|
||||
int eaten = -1;
|
||||
|
||||
@@ -3605,7 +3605,7 @@ tcp_collapse(struct sock *sk, struct sk_buff_head *list,
|
||||
* - bloated or contains data before "start" or
|
||||
* overlaps to the next one.
|
||||
*/
|
||||
if (!skb->h.th->syn && !skb->h.th->fin &&
|
||||
if (!tcp_hdr(skb)->syn && !tcp_hdr(skb)->fin &&
|
||||
(tcp_win_from_space(skb->truesize) > skb->len ||
|
||||
before(TCP_SKB_CB(skb)->seq, start) ||
|
||||
(skb->next != tail &&
|
||||
@@ -3616,7 +3616,7 @@ tcp_collapse(struct sock *sk, struct sk_buff_head *list,
|
||||
start = TCP_SKB_CB(skb)->end_seq;
|
||||
skb = skb->next;
|
||||
}
|
||||
if (skb == tail || skb->h.th->syn || skb->h.th->fin)
|
||||
if (skb == tail || tcp_hdr(skb)->syn || tcp_hdr(skb)->fin)
|
||||
return;
|
||||
|
||||
while (before(start, end)) {
|
||||
@@ -3665,7 +3665,9 @@ tcp_collapse(struct sock *sk, struct sk_buff_head *list,
|
||||
__kfree_skb(skb);
|
||||
NET_INC_STATS_BH(LINUX_MIB_TCPRCVCOLLAPSED);
|
||||
skb = next;
|
||||
if (skb == tail || skb->h.th->syn || skb->h.th->fin)
|
||||
if (skb == tail ||
|
||||
tcp_hdr(skb)->syn ||
|
||||
tcp_hdr(skb)->fin)
|
||||
return;
|
||||
}
|
||||
}
|
||||
@@ -4072,7 +4074,7 @@ static int tcp_dma_try_early_copy(struct sock *sk, struct sk_buff *skb, int hlen
|
||||
tcp_rcv_space_adjust(sk);
|
||||
|
||||
if ((tp->ucopy.len == 0) ||
|
||||
(tcp_flag_word(skb->h.th) & TCP_FLAG_PSH) ||
|
||||
(tcp_flag_word(tcp_hdr(skb)) & TCP_FLAG_PSH) ||
|
||||
(atomic_read(&sk->sk_rmem_alloc) > (sk->sk_rcvbuf >> 1))) {
|
||||
tp->ucopy.wakeup = 1;
|
||||
sk->sk_data_ready(sk, 0);
|
||||
|
||||
+16
-16
@@ -127,8 +127,8 @@ static inline __u32 tcp_v4_init_sequence(struct sk_buff *skb)
|
||||
{
|
||||
return secure_tcp_sequence_number(ip_hdr(skb)->daddr,
|
||||
ip_hdr(skb)->saddr,
|
||||
skb->h.th->dest,
|
||||
skb->h.th->source);
|
||||
tcp_hdr(skb)->dest,
|
||||
tcp_hdr(skb)->source);
|
||||
}
|
||||
|
||||
int tcp_twsk_unique(struct sock *sk, struct sock *sktw, void *twp)
|
||||
@@ -499,7 +499,7 @@ out:
|
||||
void tcp_v4_send_check(struct sock *sk, int len, struct sk_buff *skb)
|
||||
{
|
||||
struct inet_sock *inet = inet_sk(sk);
|
||||
struct tcphdr *th = skb->h.th;
|
||||
struct tcphdr *th = tcp_hdr(skb);
|
||||
|
||||
if (skb->ip_summed == CHECKSUM_PARTIAL) {
|
||||
th->check = ~tcp_v4_check(len, inet->saddr,
|
||||
@@ -522,7 +522,7 @@ int tcp_v4_gso_send_check(struct sk_buff *skb)
|
||||
return -EINVAL;
|
||||
|
||||
iph = ip_hdr(skb);
|
||||
th = skb->h.th;
|
||||
th = tcp_hdr(skb);
|
||||
|
||||
th->check = 0;
|
||||
th->check = ~tcp_v4_check(skb->len, iph->saddr, iph->daddr, 0);
|
||||
@@ -546,7 +546,7 @@ int tcp_v4_gso_send_check(struct sk_buff *skb)
|
||||
|
||||
static void tcp_v4_send_reset(struct sock *sk, struct sk_buff *skb)
|
||||
{
|
||||
struct tcphdr *th = skb->h.th;
|
||||
struct tcphdr *th = tcp_hdr(skb);
|
||||
struct {
|
||||
struct tcphdr th;
|
||||
#ifdef CONFIG_TCP_MD5SIG
|
||||
@@ -622,7 +622,7 @@ static void tcp_v4_send_ack(struct tcp_timewait_sock *twsk,
|
||||
struct sk_buff *skb, u32 seq, u32 ack,
|
||||
u32 win, u32 ts)
|
||||
{
|
||||
struct tcphdr *th = skb->h.th;
|
||||
struct tcphdr *th = tcp_hdr(skb);
|
||||
struct {
|
||||
struct tcphdr th;
|
||||
__be32 opt[(TCPOLEN_TSTAMP_ALIGNED >> 2)
|
||||
@@ -745,7 +745,7 @@ static int tcp_v4_send_synack(struct sock *sk, struct request_sock *req,
|
||||
skb = tcp_make_synack(sk, dst, req);
|
||||
|
||||
if (skb) {
|
||||
struct tcphdr *th = skb->h.th;
|
||||
struct tcphdr *th = tcp_hdr(skb);
|
||||
|
||||
th->check = tcp_v4_check(skb->len,
|
||||
ireq->loc_addr,
|
||||
@@ -781,7 +781,7 @@ static void syn_flood_warning(struct sk_buff *skb)
|
||||
warntime = jiffies;
|
||||
printk(KERN_INFO
|
||||
"possible SYN flooding on port %d. Sending cookies.\n",
|
||||
ntohs(skb->h.th->dest));
|
||||
ntohs(tcp_hdr(skb)->dest));
|
||||
}
|
||||
}
|
||||
#endif
|
||||
@@ -1134,7 +1134,7 @@ static int tcp_v4_inbound_md5_hash(struct sock *sk, struct sk_buff *skb)
|
||||
__u8 *hash_location = NULL;
|
||||
struct tcp_md5sig_key *hash_expected;
|
||||
const struct iphdr *iph = ip_hdr(skb);
|
||||
struct tcphdr *th = skb->h.th;
|
||||
struct tcphdr *th = tcp_hdr(skb);
|
||||
int length = (th->doff << 2) - sizeof(struct tcphdr);
|
||||
int genhash;
|
||||
unsigned char *ptr;
|
||||
@@ -1327,7 +1327,7 @@ int tcp_v4_conn_request(struct sock *sk, struct sk_buff *skb)
|
||||
ireq->rmt_addr = saddr;
|
||||
ireq->opt = tcp_v4_save_options(sk, skb);
|
||||
if (!want_cookie)
|
||||
TCP_ECN_create_request(req, skb->h.th);
|
||||
TCP_ECN_create_request(req, tcp_hdr(skb));
|
||||
|
||||
if (want_cookie) {
|
||||
#ifdef CONFIG_SYN_COOKIES
|
||||
@@ -1375,7 +1375,7 @@ int tcp_v4_conn_request(struct sock *sk, struct sk_buff *skb)
|
||||
LIMIT_NETDEBUG(KERN_DEBUG "TCP: drop open "
|
||||
"request from %u.%u.%u.%u/%u\n",
|
||||
NIPQUAD(saddr),
|
||||
ntohs(skb->h.th->source));
|
||||
ntohs(tcp_hdr(skb)->source));
|
||||
dst_release(dst);
|
||||
goto drop_and_free;
|
||||
}
|
||||
@@ -1481,7 +1481,7 @@ exit:
|
||||
|
||||
static struct sock *tcp_v4_hnd_req(struct sock *sk, struct sk_buff *skb)
|
||||
{
|
||||
struct tcphdr *th = skb->h.th;
|
||||
struct tcphdr *th = tcp_hdr(skb);
|
||||
const struct iphdr *iph = ip_hdr(skb);
|
||||
struct sock *nsk;
|
||||
struct request_sock **prev;
|
||||
@@ -1556,7 +1556,7 @@ int tcp_v4_do_rcv(struct sock *sk, struct sk_buff *skb)
|
||||
|
||||
if (sk->sk_state == TCP_ESTABLISHED) { /* Fast path */
|
||||
TCP_CHECK_TIMER(sk);
|
||||
if (tcp_rcv_established(sk, skb, skb->h.th, skb->len)) {
|
||||
if (tcp_rcv_established(sk, skb, tcp_hdr(skb), skb->len)) {
|
||||
rsk = sk;
|
||||
goto reset;
|
||||
}
|
||||
@@ -1582,7 +1582,7 @@ int tcp_v4_do_rcv(struct sock *sk, struct sk_buff *skb)
|
||||
}
|
||||
|
||||
TCP_CHECK_TIMER(sk);
|
||||
if (tcp_rcv_state_process(sk, skb, skb->h.th, skb->len)) {
|
||||
if (tcp_rcv_state_process(sk, skb, tcp_hdr(skb), skb->len)) {
|
||||
rsk = sk;
|
||||
goto reset;
|
||||
}
|
||||
@@ -1625,7 +1625,7 @@ int tcp_v4_rcv(struct sk_buff *skb)
|
||||
if (!pskb_may_pull(skb, sizeof(struct tcphdr)))
|
||||
goto discard_it;
|
||||
|
||||
th = skb->h.th;
|
||||
th = tcp_hdr(skb);
|
||||
|
||||
if (th->doff < sizeof(struct tcphdr) / 4)
|
||||
goto bad_packet;
|
||||
@@ -1640,7 +1640,7 @@ int tcp_v4_rcv(struct sk_buff *skb)
|
||||
tcp_v4_checksum_init(skb)))
|
||||
goto bad_packet;
|
||||
|
||||
th = skb->h.th;
|
||||
th = tcp_hdr(skb);
|
||||
iph = ip_hdr(skb);
|
||||
TCP_SKB_CB(skb)->seq = ntohl(th->seq);
|
||||
TCP_SKB_CB(skb)->end_seq = (TCP_SKB_CB(skb)->seq + th->syn + th->fin +
|
||||
|
||||
@@ -453,7 +453,8 @@ struct sock *tcp_create_openreq_child(struct sock *sk, struct request_sock *req,
|
||||
newtp->rx_opt.snd_wscale = newtp->rx_opt.rcv_wscale = 0;
|
||||
newtp->window_clamp = min(newtp->window_clamp, 65535U);
|
||||
}
|
||||
newtp->snd_wnd = ntohs(skb->h.th->window) << newtp->rx_opt.snd_wscale;
|
||||
newtp->snd_wnd = (ntohs(tcp_hdr(skb)->window) <<
|
||||
newtp->rx_opt.snd_wscale);
|
||||
newtp->max_window = newtp->snd_wnd;
|
||||
|
||||
if (newtp->rx_opt.tstamp_ok) {
|
||||
@@ -488,7 +489,7 @@ struct sock *tcp_check_req(struct sock *sk,struct sk_buff *skb,
|
||||
struct request_sock *req,
|
||||
struct request_sock **prev)
|
||||
{
|
||||
struct tcphdr *th = skb->h.th;
|
||||
const struct tcphdr *th = tcp_hdr(skb);
|
||||
__be32 flg = tcp_flag_word(th) & (TCP_FLAG_RST|TCP_FLAG_SYN|TCP_FLAG_ACK);
|
||||
int paws_reject = 0;
|
||||
struct tcp_options_received tmp_opt;
|
||||
@@ -710,8 +711,8 @@ int tcp_child_process(struct sock *parent, struct sock *child,
|
||||
int state = child->sk_state;
|
||||
|
||||
if (!sock_owned_by_user(child)) {
|
||||
ret = tcp_rcv_state_process(child, skb, skb->h.th, skb->len);
|
||||
|
||||
ret = tcp_rcv_state_process(child, skb, tcp_hdr(skb),
|
||||
skb->len);
|
||||
/* Wakeup parent, send SIGIO */
|
||||
if (state == TCP_SYN_RECV && child->sk_state != state)
|
||||
parent->sk_data_ready(parent, 0);
|
||||
|
||||
@@ -465,11 +465,12 @@ static int tcp_transmit_skb(struct sock *sk, struct sk_buff *skb, int clone_it,
|
||||
tcp_header_size += TCPOLEN_MD5SIG_ALIGNED;
|
||||
#endif
|
||||
|
||||
th = (struct tcphdr *) skb_push(skb, tcp_header_size);
|
||||
skb->h.th = th;
|
||||
skb_push(skb, tcp_header_size);
|
||||
skb_reset_transport_header(skb);
|
||||
skb_set_owner_w(skb, sk);
|
||||
|
||||
/* Build TCP header and checksum it. */
|
||||
th = tcp_hdr(skb);
|
||||
th->source = inet->sport;
|
||||
th->dest = inet->dport;
|
||||
th->seq = htonl(tcb->seq);
|
||||
@@ -524,7 +525,7 @@ static int tcp_transmit_skb(struct sock *sk, struct sk_buff *skb, int clone_it,
|
||||
tp->af_specific->calc_md5_hash(md5_hash_location,
|
||||
md5,
|
||||
sk, NULL, NULL,
|
||||
skb->h.th,
|
||||
tcp_hdr(skb),
|
||||
sk->sk_protocol,
|
||||
skb->len);
|
||||
}
|
||||
@@ -2128,8 +2129,10 @@ struct sk_buff * tcp_make_synack(struct sock *sk, struct dst_entry *dst,
|
||||
if (md5)
|
||||
tcp_header_size += TCPOLEN_MD5SIG_ALIGNED;
|
||||
#endif
|
||||
skb->h.th = th = (struct tcphdr *) skb_push(skb, tcp_header_size);
|
||||
skb_push(skb, tcp_header_size);
|
||||
skb_reset_transport_header(skb);
|
||||
|
||||
th = tcp_hdr(skb);
|
||||
memset(th, 0, sizeof(struct tcphdr));
|
||||
th->syn = 1;
|
||||
th->ack = 1;
|
||||
@@ -2183,7 +2186,7 @@ struct sk_buff * tcp_make_synack(struct sock *sk, struct dst_entry *dst,
|
||||
tp->af_specific->calc_md5_hash(md5_hash_location,
|
||||
md5,
|
||||
NULL, dst, req,
|
||||
skb->h.th, sk->sk_protocol,
|
||||
tcp_hdr(skb), sk->sk_protocol,
|
||||
skb->len);
|
||||
}
|
||||
#endif
|
||||
|
||||
+16
-16
@@ -117,8 +117,8 @@ static __u32 tcp_v6_init_sequence(struct sk_buff *skb)
|
||||
{
|
||||
return secure_tcpv6_sequence_number(ipv6_hdr(skb)->daddr.s6_addr32,
|
||||
ipv6_hdr(skb)->saddr.s6_addr32,
|
||||
skb->h.th->dest,
|
||||
skb->h.th->source);
|
||||
tcp_hdr(skb)->dest,
|
||||
tcp_hdr(skb)->source);
|
||||
}
|
||||
|
||||
static int tcp_v6_connect(struct sock *sk, struct sockaddr *uaddr,
|
||||
@@ -509,7 +509,7 @@ static int tcp_v6_send_synack(struct sock *sk, struct request_sock *req,
|
||||
|
||||
skb = tcp_make_synack(sk, dst, req);
|
||||
if (skb) {
|
||||
struct tcphdr *th = skb->h.th;
|
||||
struct tcphdr *th = tcp_hdr(skb);
|
||||
|
||||
th->check = tcp_v6_check(th, skb->len,
|
||||
&treq->loc_addr, &treq->rmt_addr,
|
||||
@@ -838,7 +838,7 @@ static int tcp_v6_inbound_md5_hash (struct sock *sk, struct sk_buff *skb)
|
||||
__u8 *hash_location = NULL;
|
||||
struct tcp_md5sig_key *hash_expected;
|
||||
struct ipv6hdr *ip6h = ipv6_hdr(skb);
|
||||
struct tcphdr *th = skb->h.th;
|
||||
struct tcphdr *th = tcp_hdr(skb);
|
||||
int length = (th->doff << 2) - sizeof (*th);
|
||||
int genhash;
|
||||
u8 *ptr;
|
||||
@@ -946,7 +946,7 @@ static struct timewait_sock_ops tcp6_timewait_sock_ops = {
|
||||
static void tcp_v6_send_check(struct sock *sk, int len, struct sk_buff *skb)
|
||||
{
|
||||
struct ipv6_pinfo *np = inet6_sk(sk);
|
||||
struct tcphdr *th = skb->h.th;
|
||||
struct tcphdr *th = tcp_hdr(skb);
|
||||
|
||||
if (skb->ip_summed == CHECKSUM_PARTIAL) {
|
||||
th->check = ~csum_ipv6_magic(&np->saddr, &np->daddr, len, IPPROTO_TCP, 0);
|
||||
@@ -967,7 +967,7 @@ static int tcp_v6_gso_send_check(struct sk_buff *skb)
|
||||
return -EINVAL;
|
||||
|
||||
ipv6h = ipv6_hdr(skb);
|
||||
th = skb->h.th;
|
||||
th = tcp_hdr(skb);
|
||||
|
||||
th->check = 0;
|
||||
th->check = ~csum_ipv6_magic(&ipv6h->saddr, &ipv6h->daddr, skb->len,
|
||||
@@ -979,7 +979,7 @@ static int tcp_v6_gso_send_check(struct sk_buff *skb)
|
||||
|
||||
static void tcp_v6_send_reset(struct sock *sk, struct sk_buff *skb)
|
||||
{
|
||||
struct tcphdr *th = skb->h.th, *t1;
|
||||
struct tcphdr *th = tcp_hdr(skb), *t1;
|
||||
struct sk_buff *buff;
|
||||
struct flowi fl;
|
||||
int tot_len = sizeof(*th);
|
||||
@@ -1079,7 +1079,7 @@ static void tcp_v6_send_reset(struct sock *sk, struct sk_buff *skb)
|
||||
static void tcp_v6_send_ack(struct tcp_timewait_sock *tw,
|
||||
struct sk_buff *skb, u32 seq, u32 ack, u32 win, u32 ts)
|
||||
{
|
||||
struct tcphdr *th = skb->h.th, *t1;
|
||||
struct tcphdr *th = tcp_hdr(skb), *t1;
|
||||
struct sk_buff *buff;
|
||||
struct flowi fl;
|
||||
int tot_len = sizeof(struct tcphdr);
|
||||
@@ -1195,7 +1195,7 @@ static void tcp_v6_reqsk_send_ack(struct sk_buff *skb, struct request_sock *req)
|
||||
static struct sock *tcp_v6_hnd_req(struct sock *sk,struct sk_buff *skb)
|
||||
{
|
||||
struct request_sock *req, **prev;
|
||||
const struct tcphdr *th = skb->h.th;
|
||||
const struct tcphdr *th = tcp_hdr(skb);
|
||||
struct sock *nsk;
|
||||
|
||||
/* Find possible connection requests. */
|
||||
@@ -1275,7 +1275,7 @@ static int tcp_v6_conn_request(struct sock *sk, struct sk_buff *skb)
|
||||
treq = inet6_rsk(req);
|
||||
ipv6_addr_copy(&treq->rmt_addr, &ipv6_hdr(skb)->saddr);
|
||||
ipv6_addr_copy(&treq->loc_addr, &ipv6_hdr(skb)->daddr);
|
||||
TCP_ECN_create_request(req, skb->h.th);
|
||||
TCP_ECN_create_request(req, tcp_hdr(skb));
|
||||
treq->pktopts = NULL;
|
||||
if (ipv6_opt_accepted(sk, skb) ||
|
||||
np->rxopt.bits.rxinfo || np->rxopt.bits.rxoinfo ||
|
||||
@@ -1528,14 +1528,14 @@ out:
|
||||
static __sum16 tcp_v6_checksum_init(struct sk_buff *skb)
|
||||
{
|
||||
if (skb->ip_summed == CHECKSUM_COMPLETE) {
|
||||
if (!tcp_v6_check(skb->h.th, skb->len, &ipv6_hdr(skb)->saddr,
|
||||
if (!tcp_v6_check(tcp_hdr(skb), skb->len, &ipv6_hdr(skb)->saddr,
|
||||
&ipv6_hdr(skb)->daddr, skb->csum)) {
|
||||
skb->ip_summed = CHECKSUM_UNNECESSARY;
|
||||
return 0;
|
||||
}
|
||||
}
|
||||
|
||||
skb->csum = ~csum_unfold(tcp_v6_check(skb->h.th, skb->len,
|
||||
skb->csum = ~csum_unfold(tcp_v6_check(tcp_hdr(skb), skb->len,
|
||||
&ipv6_hdr(skb)->saddr,
|
||||
&ipv6_hdr(skb)->daddr, 0));
|
||||
|
||||
@@ -1601,7 +1601,7 @@ static int tcp_v6_do_rcv(struct sock *sk, struct sk_buff *skb)
|
||||
|
||||
if (sk->sk_state == TCP_ESTABLISHED) { /* Fast path */
|
||||
TCP_CHECK_TIMER(sk);
|
||||
if (tcp_rcv_established(sk, skb, skb->h.th, skb->len))
|
||||
if (tcp_rcv_established(sk, skb, tcp_hdr(skb), skb->len))
|
||||
goto reset;
|
||||
TCP_CHECK_TIMER(sk);
|
||||
if (opt_skb)
|
||||
@@ -1632,7 +1632,7 @@ static int tcp_v6_do_rcv(struct sock *sk, struct sk_buff *skb)
|
||||
}
|
||||
|
||||
TCP_CHECK_TIMER(sk);
|
||||
if (tcp_rcv_state_process(sk, skb, skb->h.th, skb->len))
|
||||
if (tcp_rcv_state_process(sk, skb, tcp_hdr(skb), skb->len))
|
||||
goto reset;
|
||||
TCP_CHECK_TIMER(sk);
|
||||
if (opt_skb)
|
||||
@@ -1698,7 +1698,7 @@ static int tcp_v6_rcv(struct sk_buff **pskb)
|
||||
if (!pskb_may_pull(skb, sizeof(struct tcphdr)))
|
||||
goto discard_it;
|
||||
|
||||
th = skb->h.th;
|
||||
th = tcp_hdr(skb);
|
||||
|
||||
if (th->doff < sizeof(struct tcphdr)/4)
|
||||
goto bad_packet;
|
||||
@@ -1709,7 +1709,7 @@ static int tcp_v6_rcv(struct sk_buff **pskb)
|
||||
tcp_v6_checksum_init(skb)))
|
||||
goto bad_packet;
|
||||
|
||||
th = skb->h.th;
|
||||
th = tcp_hdr(skb);
|
||||
TCP_SKB_CB(skb)->seq = ntohl(th->seq);
|
||||
TCP_SKB_CB(skb)->end_seq = (TCP_SKB_CB(skb)->seq + th->syn + th->fin +
|
||||
skb->len - th->doff*4);
|
||||
|
||||
Reference in New Issue
Block a user