Loading...
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 | /* SPDX-License-Identifier: GPL-2.0 or BSD-3-Clause */ /* * Copyright(c) 2017 - 2020 Intel Corporation. */ #ifndef _HFI1_VNIC_H #define _HFI1_VNIC_H #include <rdma/opa_vnic.h> #include "hfi.h" #include "sdma.h" #define HFI1_VNIC_MAX_TXQ 16 #define HFI1_VNIC_MAX_PAD 12 /* L4 header definitions */ #define HFI1_VNIC_L4_HDR_OFFSET OPA_VNIC_L2_HDR_LEN #define HFI1_VNIC_GET_L4_HDR(data) \ (*((u16 *)((u8 *)(data) + HFI1_VNIC_L4_HDR_OFFSET))) #define HFI1_VNIC_GET_VESWID(data) \ (HFI1_VNIC_GET_L4_HDR(data) & 0xFFF) /* Service class */ #define HFI1_VNIC_SC_OFFSET_LOW 6 #define HFI1_VNIC_SC_OFFSET_HI 7 #define HFI1_VNIC_SC_SHIFT 4 #define HFI1_VNIC_MAX_QUEUE 16 #define HFI1_NUM_VNIC_CTXT 8 /** * struct hfi1_vnic_sdma - VNIC per Tx ring SDMA information * @dd - device data pointer * @sde - sdma engine * @vinfo - vnic info pointer * @wait - iowait structure * @stx - sdma tx request * @state - vnic Tx ring SDMA state * @q_idx - vnic Tx queue index */ struct hfi1_vnic_sdma { struct hfi1_devdata *dd; struct sdma_engine *sde; struct hfi1_vnic_vport_info *vinfo; struct iowait wait; struct sdma_txreq stx; unsigned int state; u8 q_idx; bool pkts_sent; }; /** * struct hfi1_vnic_rx_queue - HFI1 VNIC receive queue * @idx: queue index * @vinfo: pointer to vport information * @netdev: network device * @napi: netdev napi structure * @skbq: queue of received socket buffers */ struct hfi1_vnic_rx_queue { u8 idx; struct hfi1_vnic_vport_info *vinfo; struct net_device *netdev; struct napi_struct napi; }; /** * struct hfi1_vnic_vport_info - HFI1 VNIC virtual port information * @dd: device data pointer * @netdev: net device pointer * @flags: state flags * @lock: vport lock * @num_tx_q: number of transmit queues * @num_rx_q: number of receive queues * @vesw_id: virtual switch id * @rxq: Array of receive queues * @stats: per queue stats * @sdma: VNIC SDMA structure per TXQ */ struct hfi1_vnic_vport_info { struct hfi1_devdata *dd; struct net_device *netdev; unsigned long flags; /* Lock used around state updates */ struct mutex lock; u8 num_tx_q; u8 num_rx_q; u16 vesw_id; struct hfi1_vnic_rx_queue rxq[HFI1_NUM_VNIC_CTXT]; struct opa_vnic_stats stats[HFI1_VNIC_MAX_QUEUE]; struct hfi1_vnic_sdma sdma[HFI1_VNIC_MAX_TXQ]; }; #define v_dbg(format, arg...) \ netdev_dbg(vinfo->netdev, format, ## arg) #define v_err(format, arg...) \ netdev_err(vinfo->netdev, format, ## arg) #define v_info(format, arg...) \ netdev_info(vinfo->netdev, format, ## arg) /* vnic hfi1 internal functions */ void hfi1_vnic_setup(struct hfi1_devdata *dd); int hfi1_vnic_txreq_init(struct hfi1_devdata *dd); void hfi1_vnic_txreq_deinit(struct hfi1_devdata *dd); void hfi1_vnic_bypass_rcv(struct hfi1_packet *packet); void hfi1_vnic_sdma_init(struct hfi1_vnic_vport_info *vinfo); bool hfi1_vnic_sdma_write_avail(struct hfi1_vnic_vport_info *vinfo, u8 q_idx); /* vnic rdma netdev operations */ struct net_device *hfi1_vnic_alloc_rn(struct ib_device *device, u32 port_num, enum rdma_netdev_t type, const char *name, unsigned char name_assign_type, void (*setup)(struct net_device *)); int hfi1_vnic_send_dma(struct hfi1_devdata *dd, u8 q_idx, struct hfi1_vnic_vport_info *vinfo, struct sk_buff *skb, u64 pbc, u8 plen); #endif /* _HFI1_VNIC_H */ |