Loading...
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 | /* SPDX-License-Identifier: (GPL-2.0 OR MIT) * Google virtual Ethernet (gve) driver * * Copyright (C) 2015-2019 Google, Inc. */ /* GVE Transmit Descriptor formats */ #ifndef _GVE_DESC_H_ #define _GVE_DESC_H_ #include <linux/build_bug.h> /* A note on seg_addrs * * Base addresses encoded in seg_addr are not assumed to be physical * addresses. The ring format assumes these come from some linear address * space. This could be physical memory, kernel virtual memory, user virtual * memory. * If raw dma addressing is not supported then gVNIC uses lists of registered * pages. Each queue is assumed to be associated with a single such linear * address space to ensure a consistent meaning for seg_addrs posted to its * rings. */ struct gve_tx_pkt_desc { u8 type_flags; /* desc type is lower 4 bits, flags upper */ u8 l4_csum_offset; /* relative offset of L4 csum word */ u8 l4_hdr_offset; /* Offset of start of L4 headers in packet */ u8 desc_cnt; /* Total descriptors for this packet */ __be16 len; /* Total length of this packet (in bytes) */ __be16 seg_len; /* Length of this descriptor's segment */ __be64 seg_addr; /* Base address (see note) of this segment */ } __packed; struct gve_tx_mtd_desc { u8 type_flags; /* type is lower 4 bits, subtype upper */ u8 path_state; /* state is lower 4 bits, hash type upper */ __be16 reserved0; __be32 path_hash; __be64 reserved1; } __packed; struct gve_tx_seg_desc { u8 type_flags; /* type is lower 4 bits, flags upper */ u8 l3_offset; /* TSO: 2 byte units to start of IPH */ __be16 reserved; __be16 mss; /* TSO MSS */ __be16 seg_len; __be64 seg_addr; } __packed; /* GVE Transmit Descriptor Types */ #define GVE_TXD_STD (0x0 << 4) /* Std with Host Address */ #define GVE_TXD_TSO (0x1 << 4) /* TSO with Host Address */ #define GVE_TXD_SEG (0x2 << 4) /* Seg with Host Address */ #define GVE_TXD_MTD (0x3 << 4) /* Metadata */ /* GVE Transmit Descriptor Flags for Std Pkts */ #define GVE_TXF_L4CSUM BIT(0) /* Need csum offload */ #define GVE_TXF_TSTAMP BIT(2) /* Timestamp required */ /* GVE Transmit Descriptor Flags for TSO Segs */ #define GVE_TXSF_IPV6 BIT(1) /* IPv6 TSO */ /* GVE Transmit Descriptor Options for MTD Segs */ #define GVE_MTD_SUBTYPE_PATH 0 #define GVE_MTD_PATH_STATE_DEFAULT 0 #define GVE_MTD_PATH_STATE_TIMEOUT 1 #define GVE_MTD_PATH_STATE_CONGESTION 2 #define GVE_MTD_PATH_STATE_RETRANSMIT 3 #define GVE_MTD_PATH_HASH_NONE (0x0 << 4) #define GVE_MTD_PATH_HASH_L4 (0x1 << 4) /* GVE Receive Packet Descriptor */ /* The start of an ethernet packet comes 2 bytes into the rx buffer. * gVNIC adds this padding so that both the DMA and the L3/4 protocol header * access is aligned. */ #define GVE_RX_PAD 2 struct gve_rx_desc { u8 padding[48]; __be32 rss_hash; /* Receive-side scaling hash (Toeplitz for gVNIC) */ __be16 mss; __be16 reserved; /* Reserved to zero */ u8 hdr_len; /* Header length (L2-L4) including padding */ u8 hdr_off; /* 64-byte-scaled offset into RX_DATA entry */ __sum16 csum; /* 1's-complement partial checksum of L3+ bytes */ __be16 len; /* Length of the received packet */ __be16 flags_seq; /* Flags [15:3] and sequence number [2:0] (1-7) */ } __packed; static_assert(sizeof(struct gve_rx_desc) == 64); /* If the device supports raw dma addressing then the addr in data slot is * the dma address of the buffer. * If the device only supports registered segments then the addr is a byte * offset into the registered segment (an ordered list of pages) where the * buffer is. */ union gve_rx_data_slot { __be64 qpl_offset; __be64 addr; }; /* GVE Recive Packet Descriptor Seq No */ #define GVE_SEQNO(x) (be16_to_cpu(x) & 0x7) /* GVE Recive Packet Descriptor Flags */ #define GVE_RXFLG(x) cpu_to_be16(1 << (3 + (x))) #define GVE_RXF_FRAG GVE_RXFLG(3) /* IP Fragment */ #define GVE_RXF_IPV4 GVE_RXFLG(4) /* IPv4 */ #define GVE_RXF_IPV6 GVE_RXFLG(5) /* IPv6 */ #define GVE_RXF_TCP GVE_RXFLG(6) /* TCP Packet */ #define GVE_RXF_UDP GVE_RXFLG(7) /* UDP Packet */ #define GVE_RXF_ERR GVE_RXFLG(8) /* Packet Error Detected */ #define GVE_RXF_PKT_CONT GVE_RXFLG(10) /* Multi Fragment RX packet */ /* GVE IRQ */ #define GVE_IRQ_ACK BIT(31) #define GVE_IRQ_MASK BIT(30) #define GVE_IRQ_EVENT BIT(29) static inline bool gve_needs_rss(__be16 flag) { if (flag & GVE_RXF_FRAG) return false; if (flag & (GVE_RXF_IPV4 | GVE_RXF_IPV6)) return true; return false; } static inline u8 gve_next_seqno(u8 seq) { return (seq + 1) == 8 ? 1 : seq + 1; } #endif /* _GVE_DESC_H_ */ |