@@ -974,7 +974,7 @@ dp_packet_ol_tx_l4_checksum(const struct dp_packet *a)
/* Returns 'true' if packet 'a' is marked for TCP segmentation offloading. */
static inline bool
-dp_packet_ol_is_tso(const struct dp_packet *a)
+dp_packet_ol_tcp_seg(const struct dp_packet *a)
{
return !!(*dp_packet_ol_flags_ptr(a) & DP_PACKET_OL_TX_TCP_SEG);
}
@@ -6682,7 +6682,7 @@ netdev_linux_prepend_vnet_hdr(struct dp_packet *p, int mtu)
{
struct virtio_net_hdr *vnet = dp_packet_push_zeros(p, sizeof *vnet);
- if (dp_packet_ol_is_tso(p)) {
+ if (dp_packet_ol_tcp_seg(p)) {
uint16_t hdr_len = ((char *) dp_packet_l4(p)
- (char *) dp_packet_eth(p)) + TCP_HEADER_LEN;
@@ -794,7 +794,7 @@ netdev_send_prepare_packet(const uint64_t netdev_flags,
{
uint64_t l4_mask;
- if (dp_packet_ol_is_tso(packet)
+ if (dp_packet_ol_tcp_seg(packet)
&& !(netdev_flags & NETDEV_OFFLOAD_TX_TCP_TSO)) {
/* Fall back to GSO in software. */
VLOG_ERR_BUF(errormsg, "No TSO support");
@@ -960,7 +960,7 @@ netdev_push_header(const struct netdev *netdev,
size_t i, size = dp_packet_batch_size(batch);
DP_PACKET_BATCH_REFILL_FOR_EACH (i, size, packet, batch) {
- if (OVS_UNLIKELY(dp_packet_ol_is_tso(packet)
+ if (OVS_UNLIKELY(dp_packet_ol_tcp_seg(packet)
|| dp_packet_ol_l4_mask(packet))) {
COVERAGE_INC(netdev_push_header_drops);
dp_packet_delete(packet);