| /* |
| * QEMU RX packets abstractions |
| * |
| * Copyright (c) 2012 Ravello Systems LTD (http://ravellosystems.com) |
| * |
| * Developed by Daynix Computing LTD (http://www.daynix.com) |
| * |
| * Authors: |
| * Dmitry Fleytman <dmitry@daynix.com> |
| * Tamir Shomer <tamirs@daynix.com> |
| * Yan Vugenfirer <yan@daynix.com> |
| * |
| * This work is licensed under the terms of the GNU GPL, version 2 or later. |
| * See the COPYING file in the top-level directory. |
| * |
| */ |
| |
| #include "qemu/osdep.h" |
| #include "qemu/crc32c.h" |
| #include "trace.h" |
| #include "net_rx_pkt.h" |
| #include "net/checksum.h" |
| #include "net/tap.h" |
| |
| struct NetRxPkt { |
| struct virtio_net_hdr virt_hdr; |
| struct { |
| struct eth_header eth; |
| struct vlan_header vlan; |
| } ehdr_buf; |
| struct iovec *vec; |
| uint16_t vec_len_total; |
| uint16_t vec_len; |
| uint32_t tot_len; |
| uint16_t tci; |
| size_t ehdr_buf_len; |
| eth_pkt_types_e packet_type; |
| |
| /* Analysis results */ |
| bool hasip4; |
| bool hasip6; |
| |
| size_t l3hdr_off; |
| size_t l4hdr_off; |
| size_t l5hdr_off; |
| |
| eth_ip6_hdr_info ip6hdr_info; |
| eth_ip4_hdr_info ip4hdr_info; |
| eth_l4_hdr_info l4hdr_info; |
| }; |
| |
| void net_rx_pkt_init(struct NetRxPkt **pkt) |
| { |
| struct NetRxPkt *p = g_malloc0(sizeof *p); |
| p->vec = NULL; |
| p->vec_len_total = 0; |
| *pkt = p; |
| } |
| |
| void net_rx_pkt_uninit(struct NetRxPkt *pkt) |
| { |
| if (pkt->vec_len_total != 0) { |
| g_free(pkt->vec); |
| } |
| |
| g_free(pkt); |
| } |
| |
| struct virtio_net_hdr *net_rx_pkt_get_vhdr(struct NetRxPkt *pkt) |
| { |
| assert(pkt); |
| return &pkt->virt_hdr; |
| } |
| |
| static inline void |
| net_rx_pkt_iovec_realloc(struct NetRxPkt *pkt, |
| int new_iov_len) |
| { |
| if (pkt->vec_len_total < new_iov_len) { |
| g_free(pkt->vec); |
| pkt->vec = g_malloc(sizeof(*pkt->vec) * new_iov_len); |
| pkt->vec_len_total = new_iov_len; |
| } |
| } |
| |
| static void |
| net_rx_pkt_pull_data(struct NetRxPkt *pkt, |
| const struct iovec *iov, int iovcnt, |
| size_t ploff) |
| { |
| uint32_t pllen = iov_size(iov, iovcnt) - ploff; |
| |
| if (pkt->ehdr_buf_len) { |
| net_rx_pkt_iovec_realloc(pkt, iovcnt + 1); |
| |
| pkt->vec[0].iov_base = &pkt->ehdr_buf; |
| pkt->vec[0].iov_len = pkt->ehdr_buf_len; |
| |
| pkt->tot_len = pllen + pkt->ehdr_buf_len; |
| pkt->vec_len = iov_copy(pkt->vec + 1, pkt->vec_len_total - 1, |
| iov, iovcnt, ploff, pllen) + 1; |
| } else { |
| net_rx_pkt_iovec_realloc(pkt, iovcnt); |
| |
| pkt->tot_len = pllen; |
| pkt->vec_len = iov_copy(pkt->vec, pkt->vec_len_total, |
| iov, iovcnt, ploff, pkt->tot_len); |
| } |
| |
| eth_get_protocols(pkt->vec, pkt->vec_len, 0, &pkt->hasip4, &pkt->hasip6, |
| &pkt->l3hdr_off, &pkt->l4hdr_off, &pkt->l5hdr_off, |
| &pkt->ip6hdr_info, &pkt->ip4hdr_info, &pkt->l4hdr_info); |
| |
| trace_net_rx_pkt_parsed(pkt->hasip4, pkt->hasip6, pkt->l4hdr_info.proto, |
| pkt->l3hdr_off, pkt->l4hdr_off, pkt->l5hdr_off); |
| } |
| |
| void net_rx_pkt_attach_iovec(struct NetRxPkt *pkt, |
| const struct iovec *iov, int iovcnt, |
| size_t iovoff, bool strip_vlan) |
| { |
| uint16_t tci = 0; |
| uint16_t ploff = iovoff; |
| assert(pkt); |
| |
| if (strip_vlan) { |
| pkt->ehdr_buf_len = eth_strip_vlan(iov, iovcnt, iovoff, &pkt->ehdr_buf, |
| &ploff, &tci); |
| } else { |
| pkt->ehdr_buf_len = 0; |
| } |
| |
| pkt->tci = tci; |
| |
| net_rx_pkt_pull_data(pkt, iov, iovcnt, ploff); |
| } |
| |
| void net_rx_pkt_attach_iovec_ex(struct NetRxPkt *pkt, |
| const struct iovec *iov, int iovcnt, |
| size_t iovoff, int strip_vlan_index, |
| uint16_t vet, uint16_t vet_ext) |
| { |
| uint16_t tci = 0; |
| uint16_t ploff = iovoff; |
| assert(pkt); |
| |
| pkt->ehdr_buf_len = eth_strip_vlan_ex(iov, iovcnt, iovoff, |
| strip_vlan_index, vet, vet_ext, |
| &pkt->ehdr_buf, |
| &ploff, &tci); |
| |
| pkt->tci = tci; |
| |
| net_rx_pkt_pull_data(pkt, iov, iovcnt, ploff); |
| } |
| |
| void net_rx_pkt_dump(struct NetRxPkt *pkt) |
| { |
| #ifdef NET_RX_PKT_DEBUG |
| assert(pkt); |
| |
| printf("RX PKT: tot_len: %d, ehdr_buf_len: %lu, vlan_tag: %d\n", |
| pkt->tot_len, pkt->ehdr_buf_len, pkt->tci); |
| #endif |
| } |
| |
| void net_rx_pkt_set_packet_type(struct NetRxPkt *pkt, |
| eth_pkt_types_e packet_type) |
| { |
| assert(pkt); |
| |
| pkt->packet_type = packet_type; |
| |
| } |
| |
| eth_pkt_types_e net_rx_pkt_get_packet_type(struct NetRxPkt *pkt) |
| { |
| assert(pkt); |
| |
| return pkt->packet_type; |
| } |
| |
| size_t net_rx_pkt_get_total_len(struct NetRxPkt *pkt) |
| { |
| assert(pkt); |
| |
| return pkt->tot_len; |
| } |
| |
| void net_rx_pkt_set_protocols(struct NetRxPkt *pkt, |
| const struct iovec *iov, size_t iovcnt, |
| size_t iovoff) |
| { |
| assert(pkt); |
| |
| eth_get_protocols(iov, iovcnt, iovoff, &pkt->hasip4, &pkt->hasip6, |
| &pkt->l3hdr_off, &pkt->l4hdr_off, &pkt->l5hdr_off, |
| &pkt->ip6hdr_info, &pkt->ip4hdr_info, &pkt->l4hdr_info); |
| } |
| |
| void net_rx_pkt_get_protocols(struct NetRxPkt *pkt, |
| bool *hasip4, bool *hasip6, |
| EthL4HdrProto *l4hdr_proto) |
| { |
| assert(pkt); |
| |
| *hasip4 = pkt->hasip4; |
| *hasip6 = pkt->hasip6; |
| *l4hdr_proto = pkt->l4hdr_info.proto; |
| } |
| |
| size_t net_rx_pkt_get_l3_hdr_offset(struct NetRxPkt *pkt) |
| { |
| assert(pkt); |
| return pkt->l3hdr_off; |
| } |
| |
| size_t net_rx_pkt_get_l4_hdr_offset(struct NetRxPkt *pkt) |
| { |
| assert(pkt); |
| return pkt->l4hdr_off; |
| } |
| |
| size_t net_rx_pkt_get_l5_hdr_offset(struct NetRxPkt *pkt) |
| { |
| assert(pkt); |
| return pkt->l5hdr_off; |
| } |
| |
| eth_ip6_hdr_info *net_rx_pkt_get_ip6_info(struct NetRxPkt *pkt) |
| { |
| return &pkt->ip6hdr_info; |
| } |
| |
| eth_ip4_hdr_info *net_rx_pkt_get_ip4_info(struct NetRxPkt *pkt) |
| { |
| return &pkt->ip4hdr_info; |
| } |
| |
| static inline void |
| _net_rx_rss_add_chunk(uint8_t *rss_input, size_t *bytes_written, |
| void *ptr, size_t size) |
| { |
| memcpy(&rss_input[*bytes_written], ptr, size); |
| trace_net_rx_pkt_rss_add_chunk(ptr, size, *bytes_written); |
| *bytes_written += size; |
| } |
| |
| static inline void |
| _net_rx_rss_prepare_ip4(uint8_t *rss_input, |
| struct NetRxPkt *pkt, |
| size_t *bytes_written) |
| { |
| struct ip_header *ip4_hdr = &pkt->ip4hdr_info.ip4_hdr; |
| |
| _net_rx_rss_add_chunk(rss_input, bytes_written, |
| &ip4_hdr->ip_src, sizeof(uint32_t)); |
| |
| _net_rx_rss_add_chunk(rss_input, bytes_written, |
| &ip4_hdr->ip_dst, sizeof(uint32_t)); |
| } |
| |
| static inline void |
| _net_rx_rss_prepare_ip6(uint8_t *rss_input, |
| struct NetRxPkt *pkt, |
| bool ipv6ex, size_t *bytes_written) |
| { |
| eth_ip6_hdr_info *ip6info = &pkt->ip6hdr_info; |
| |
| _net_rx_rss_add_chunk(rss_input, bytes_written, |
| (ipv6ex && ip6info->rss_ex_src_valid) ? &ip6info->rss_ex_src |
| : &ip6info->ip6_hdr.ip6_src, |
| sizeof(struct in6_address)); |
| |
| _net_rx_rss_add_chunk(rss_input, bytes_written, |
| (ipv6ex && ip6info->rss_ex_dst_valid) ? &ip6info->rss_ex_dst |
| : &ip6info->ip6_hdr.ip6_dst, |
| sizeof(struct in6_address)); |
| } |
| |
| static inline void |
| _net_rx_rss_prepare_tcp(uint8_t *rss_input, |
| struct NetRxPkt *pkt, |
| size_t *bytes_written) |
| { |
| struct tcp_header *tcphdr = &pkt->l4hdr_info.hdr.tcp; |
| |
| _net_rx_rss_add_chunk(rss_input, bytes_written, |
| &tcphdr->th_sport, sizeof(uint16_t)); |
| |
| _net_rx_rss_add_chunk(rss_input, bytes_written, |
| &tcphdr->th_dport, sizeof(uint16_t)); |
| } |
| |
| static inline void |
| _net_rx_rss_prepare_udp(uint8_t *rss_input, |
| struct NetRxPkt *pkt, |
| size_t *bytes_written) |
| { |
| struct udp_header *udphdr = &pkt->l4hdr_info.hdr.udp; |
| |
| _net_rx_rss_add_chunk(rss_input, bytes_written, |
| &udphdr->uh_sport, sizeof(uint16_t)); |
| |
| _net_rx_rss_add_chunk(rss_input, bytes_written, |
| &udphdr->uh_dport, sizeof(uint16_t)); |
| } |
| |
| uint32_t |
| net_rx_pkt_calc_rss_hash(struct NetRxPkt *pkt, |
| NetRxPktRssType type, |
| uint8_t *key) |
| { |
| uint8_t rss_input[36]; |
| size_t rss_length = 0; |
| uint32_t rss_hash = 0; |
| net_toeplitz_key key_data; |
| |
| switch (type) { |
| case NetPktRssIpV4: |
| assert(pkt->hasip4); |
| trace_net_rx_pkt_rss_ip4(); |
| _net_rx_rss_prepare_ip4(&rss_input[0], pkt, &rss_length); |
| break; |
| case NetPktRssIpV4Tcp: |
| assert(pkt->hasip4); |
| assert(pkt->l4hdr_info.proto == ETH_L4_HDR_PROTO_TCP); |
| trace_net_rx_pkt_rss_ip4_tcp(); |
| _net_rx_rss_prepare_ip4(&rss_input[0], pkt, &rss_length); |
| _net_rx_rss_prepare_tcp(&rss_input[0], pkt, &rss_length); |
| break; |
| case NetPktRssIpV6Tcp: |
| assert(pkt->hasip6); |
| assert(pkt->l4hdr_info.proto == ETH_L4_HDR_PROTO_TCP); |
| trace_net_rx_pkt_rss_ip6_tcp(); |
| _net_rx_rss_prepare_ip6(&rss_input[0], pkt, false, &rss_length); |
| _net_rx_rss_prepare_tcp(&rss_input[0], pkt, &rss_length); |
| break; |
| case NetPktRssIpV6: |
| assert(pkt->hasip6); |
| trace_net_rx_pkt_rss_ip6(); |
| _net_rx_rss_prepare_ip6(&rss_input[0], pkt, false, &rss_length); |
| break; |
| case NetPktRssIpV6Ex: |
| assert(pkt->hasip6); |
| trace_net_rx_pkt_rss_ip6_ex(); |
| _net_rx_rss_prepare_ip6(&rss_input[0], pkt, true, &rss_length); |
| break; |
| case NetPktRssIpV6TcpEx: |
| assert(pkt->hasip6); |
| assert(pkt->l4hdr_info.proto == ETH_L4_HDR_PROTO_TCP); |
| trace_net_rx_pkt_rss_ip6_ex_tcp(); |
| _net_rx_rss_prepare_ip6(&rss_input[0], pkt, true, &rss_length); |
| _net_rx_rss_prepare_tcp(&rss_input[0], pkt, &rss_length); |
| break; |
| case NetPktRssIpV4Udp: |
| assert(pkt->hasip4); |
| assert(pkt->l4hdr_info.proto == ETH_L4_HDR_PROTO_UDP); |
| trace_net_rx_pkt_rss_ip4_udp(); |
| _net_rx_rss_prepare_ip4(&rss_input[0], pkt, &rss_length); |
| _net_rx_rss_prepare_udp(&rss_input[0], pkt, &rss_length); |
| break; |
| case NetPktRssIpV6Udp: |
| assert(pkt->hasip6); |
| assert(pkt->l4hdr_info.proto == ETH_L4_HDR_PROTO_UDP); |
| trace_net_rx_pkt_rss_ip6_udp(); |
| _net_rx_rss_prepare_ip6(&rss_input[0], pkt, false, &rss_length); |
| _net_rx_rss_prepare_udp(&rss_input[0], pkt, &rss_length); |
| break; |
| case NetPktRssIpV6UdpEx: |
| assert(pkt->hasip6); |
| assert(pkt->l4hdr_info.proto == ETH_L4_HDR_PROTO_UDP); |
| trace_net_rx_pkt_rss_ip6_ex_udp(); |
| _net_rx_rss_prepare_ip6(&rss_input[0], pkt, true, &rss_length); |
| _net_rx_rss_prepare_udp(&rss_input[0], pkt, &rss_length); |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| |
| net_toeplitz_key_init(&key_data, key); |
| net_toeplitz_add(&rss_hash, rss_input, rss_length, &key_data); |
| |
| trace_net_rx_pkt_rss_hash(rss_length, rss_hash); |
| |
| return rss_hash; |
| } |
| |
| uint16_t net_rx_pkt_get_ip_id(struct NetRxPkt *pkt) |
| { |
| assert(pkt); |
| |
| if (pkt->hasip4) { |
| return be16_to_cpu(pkt->ip4hdr_info.ip4_hdr.ip_id); |
| } |
| |
| return 0; |
| } |
| |
| bool net_rx_pkt_is_tcp_ack(struct NetRxPkt *pkt) |
| { |
| assert(pkt); |
| |
| if (pkt->l4hdr_info.proto == ETH_L4_HDR_PROTO_TCP) { |
| return TCP_HEADER_FLAGS(&pkt->l4hdr_info.hdr.tcp) & TCP_FLAG_ACK; |
| } |
| |
| return false; |
| } |
| |
| bool net_rx_pkt_has_tcp_data(struct NetRxPkt *pkt) |
| { |
| assert(pkt); |
| |
| if (pkt->l4hdr_info.proto == ETH_L4_HDR_PROTO_TCP) { |
| return pkt->l4hdr_info.has_tcp_data; |
| } |
| |
| return false; |
| } |
| |
| struct iovec *net_rx_pkt_get_iovec(struct NetRxPkt *pkt) |
| { |
| assert(pkt); |
| |
| return pkt->vec; |
| } |
| |
| uint16_t net_rx_pkt_get_iovec_len(struct NetRxPkt *pkt) |
| { |
| assert(pkt); |
| |
| return pkt->vec_len; |
| } |
| |
| void net_rx_pkt_set_vhdr(struct NetRxPkt *pkt, |
| struct virtio_net_hdr *vhdr) |
| { |
| assert(pkt); |
| |
| memcpy(&pkt->virt_hdr, vhdr, sizeof pkt->virt_hdr); |
| } |
| |
| void net_rx_pkt_set_vhdr_iovec(struct NetRxPkt *pkt, |
| const struct iovec *iov, int iovcnt) |
| { |
| assert(pkt); |
| |
| iov_to_buf(iov, iovcnt, 0, &pkt->virt_hdr, sizeof pkt->virt_hdr); |
| } |
| |
| void net_rx_pkt_unset_vhdr(struct NetRxPkt *pkt) |
| { |
| assert(pkt); |
| |
| memset(&pkt->virt_hdr, 0, sizeof(pkt->virt_hdr)); |
| } |
| |
| bool net_rx_pkt_is_vlan_stripped(struct NetRxPkt *pkt) |
| { |
| assert(pkt); |
| |
| return pkt->ehdr_buf_len ? true : false; |
| } |
| |
| uint16_t net_rx_pkt_get_vlan_tag(struct NetRxPkt *pkt) |
| { |
| assert(pkt); |
| |
| return pkt->tci; |
| } |
| |
| bool net_rx_pkt_validate_l3_csum(struct NetRxPkt *pkt, bool *csum_valid) |
| { |
| uint32_t cntr; |
| uint16_t csum; |
| uint32_t csl; |
| |
| trace_net_rx_pkt_l3_csum_validate_entry(); |
| |
| if (!pkt->hasip4) { |
| trace_net_rx_pkt_l3_csum_validate_not_ip4(); |
| return false; |
| } |
| |
| csl = pkt->l4hdr_off - pkt->l3hdr_off; |
| |
| cntr = net_checksum_add_iov(pkt->vec, pkt->vec_len, |
| pkt->l3hdr_off, |
| csl, 0); |
| |
| csum = net_checksum_finish(cntr); |
| |
| *csum_valid = (csum == 0); |
| |
| trace_net_rx_pkt_l3_csum_validate_csum(pkt->l3hdr_off, csl, |
| cntr, csum, *csum_valid); |
| |
| return true; |
| } |
| |
| static uint16_t |
| _net_rx_pkt_calc_l4_csum(struct NetRxPkt *pkt) |
| { |
| uint32_t cntr; |
| uint16_t csum; |
| uint16_t csl; |
| uint32_t cso; |
| |
| trace_net_rx_pkt_l4_csum_calc_entry(); |
| |
| if (pkt->hasip4) { |
| if (pkt->l4hdr_info.proto == ETH_L4_HDR_PROTO_UDP) { |
| csl = be16_to_cpu(pkt->l4hdr_info.hdr.udp.uh_ulen); |
| trace_net_rx_pkt_l4_csum_calc_ip4_udp(); |
| } else { |
| csl = be16_to_cpu(pkt->ip4hdr_info.ip4_hdr.ip_len) - |
| IP_HDR_GET_LEN(&pkt->ip4hdr_info.ip4_hdr); |
| trace_net_rx_pkt_l4_csum_calc_ip4_tcp(); |
| } |
| |
| cntr = eth_calc_ip4_pseudo_hdr_csum(&pkt->ip4hdr_info.ip4_hdr, |
| csl, &cso); |
| trace_net_rx_pkt_l4_csum_calc_ph_csum(cntr, csl); |
| } else { |
| if (pkt->l4hdr_info.proto == ETH_L4_HDR_PROTO_UDP) { |
| csl = be16_to_cpu(pkt->l4hdr_info.hdr.udp.uh_ulen); |
| trace_net_rx_pkt_l4_csum_calc_ip6_udp(); |
| } else { |
| struct ip6_header *ip6hdr = &pkt->ip6hdr_info.ip6_hdr; |
| size_t full_ip6hdr_len = pkt->l4hdr_off - pkt->l3hdr_off; |
| size_t ip6opts_len = full_ip6hdr_len - sizeof(struct ip6_header); |
| |
| csl = be16_to_cpu(ip6hdr->ip6_ctlun.ip6_un1.ip6_un1_plen) - |
| ip6opts_len; |
| trace_net_rx_pkt_l4_csum_calc_ip6_tcp(); |
| } |
| |
| cntr = eth_calc_ip6_pseudo_hdr_csum(&pkt->ip6hdr_info.ip6_hdr, csl, |
| pkt->ip6hdr_info.l4proto, &cso); |
| trace_net_rx_pkt_l4_csum_calc_ph_csum(cntr, csl); |
| } |
| |
| cntr += net_checksum_add_iov(pkt->vec, pkt->vec_len, |
| pkt->l4hdr_off, csl, cso); |
| |
| csum = net_checksum_finish_nozero(cntr); |
| |
| trace_net_rx_pkt_l4_csum_calc_csum(pkt->l4hdr_off, csl, cntr, csum); |
| |
| return csum; |
| } |
| |
| static bool |
| _net_rx_pkt_validate_sctp_sum(struct NetRxPkt *pkt) |
| { |
| size_t csum_off; |
| size_t off = pkt->l4hdr_off; |
| size_t vec_len = pkt->vec_len; |
| struct iovec *vec; |
| uint32_t calculated = 0; |
| uint32_t original; |
| bool valid; |
| |
| for (vec = pkt->vec; vec->iov_len < off; vec++) { |
| off -= vec->iov_len; |
| vec_len--; |
| } |
| |
| csum_off = off + 8; |
| |
| if (!iov_to_buf(vec, vec_len, csum_off, &original, sizeof(original))) { |
| return false; |
| } |
| |
| if (!iov_from_buf(vec, vec_len, csum_off, |
| &calculated, sizeof(calculated))) { |
| return false; |
| } |
| |
| calculated = crc32c(0xffffffff, |
| (uint8_t *)vec->iov_base + off, vec->iov_len - off); |
| calculated = iov_crc32c(calculated ^ 0xffffffff, vec + 1, vec_len - 1); |
| valid = calculated == le32_to_cpu(original); |
| iov_from_buf(vec, vec_len, csum_off, &original, sizeof(original)); |
| |
| return valid; |
| } |
| |
| bool net_rx_pkt_validate_l4_csum(struct NetRxPkt *pkt, bool *csum_valid) |
| { |
| uint32_t csum; |
| |
| trace_net_rx_pkt_l4_csum_validate_entry(); |
| |
| if (pkt->hasip4 && pkt->ip4hdr_info.fragment) { |
| trace_net_rx_pkt_l4_csum_validate_ip4_fragment(); |
| return false; |
| } |
| |
| switch (pkt->l4hdr_info.proto) { |
| case ETH_L4_HDR_PROTO_UDP: |
| if (pkt->l4hdr_info.hdr.udp.uh_sum == 0) { |
| trace_net_rx_pkt_l4_csum_validate_udp_with_no_checksum(); |
| return false; |
| } |
| /* fall through */ |
| case ETH_L4_HDR_PROTO_TCP: |
| csum = _net_rx_pkt_calc_l4_csum(pkt); |
| *csum_valid = ((csum == 0) || (csum == 0xFFFF)); |
| break; |
| |
| case ETH_L4_HDR_PROTO_SCTP: |
| *csum_valid = _net_rx_pkt_validate_sctp_sum(pkt); |
| break; |
| |
| default: |
| trace_net_rx_pkt_l4_csum_validate_not_xxp(); |
| return false; |
| } |
| |
| trace_net_rx_pkt_l4_csum_validate_csum(*csum_valid); |
| |
| return true; |
| } |
| |
| bool net_rx_pkt_fix_l4_csum(struct NetRxPkt *pkt) |
| { |
| uint16_t csum = 0; |
| uint32_t l4_cso; |
| |
| trace_net_rx_pkt_l4_csum_fix_entry(); |
| |
| switch (pkt->l4hdr_info.proto) { |
| case ETH_L4_HDR_PROTO_TCP: |
| l4_cso = offsetof(struct tcp_header, th_sum); |
| trace_net_rx_pkt_l4_csum_fix_tcp(l4_cso); |
| break; |
| |
| case ETH_L4_HDR_PROTO_UDP: |
| if (pkt->l4hdr_info.hdr.udp.uh_sum == 0) { |
| trace_net_rx_pkt_l4_csum_fix_udp_with_no_checksum(); |
| return false; |
| } |
| l4_cso = offsetof(struct udp_header, uh_sum); |
| trace_net_rx_pkt_l4_csum_fix_udp(l4_cso); |
| break; |
| |
| default: |
| trace_net_rx_pkt_l4_csum_fix_not_xxp(); |
| return false; |
| } |
| |
| if (pkt->hasip4 && pkt->ip4hdr_info.fragment) { |
| trace_net_rx_pkt_l4_csum_fix_ip4_fragment(); |
| return false; |
| } |
| |
| /* Set zero to checksum word */ |
| iov_from_buf(pkt->vec, pkt->vec_len, |
| pkt->l4hdr_off + l4_cso, |
| &csum, sizeof(csum)); |
| |
| /* Calculate L4 checksum */ |
| csum = cpu_to_be16(_net_rx_pkt_calc_l4_csum(pkt)); |
| |
| /* Set calculated checksum to checksum word */ |
| iov_from_buf(pkt->vec, pkt->vec_len, |
| pkt->l4hdr_off + l4_cso, |
| &csum, sizeof(csum)); |
| |
| trace_net_rx_pkt_l4_csum_fix_csum(pkt->l4hdr_off + l4_cso, csum); |
| |
| return true; |
| } |