|
33 | 33 | #include <linux/bpf_trace.h> |
34 | 34 | #include <linux/pm_runtime.h> |
35 | 35 | #include <linux/etherdevice.h> |
36 | | -#include <linux/lockdep.h> |
37 | 36 | #ifdef CONFIG_IGB_DCA |
38 | 37 | #include <linux/dca.h> |
39 | 38 | #endif |
@@ -116,8 +115,6 @@ static void igb_configure_tx(struct igb_adapter *); |
116 | 115 | static void igb_configure_rx(struct igb_adapter *); |
117 | 116 | static void igb_clean_all_tx_rings(struct igb_adapter *); |
118 | 117 | static void igb_clean_all_rx_rings(struct igb_adapter *); |
119 | | -static void igb_clean_tx_ring(struct igb_ring *); |
120 | | -static void igb_clean_rx_ring(struct igb_ring *); |
121 | 118 | static void igb_set_rx_mode(struct net_device *); |
122 | 119 | static void igb_update_phy_info(struct timer_list *); |
123 | 120 | static void igb_watchdog(struct timer_list *); |
@@ -2919,29 +2916,7 @@ static int igb_xdp(struct net_device *dev, struct netdev_bpf *xdp) |
2919 | 2916 | } |
2920 | 2917 | } |
2921 | 2918 |
|
2922 | | -/* This function assumes __netif_tx_lock is held by the caller. */ |
2923 | | -static void igb_xdp_ring_update_tail(struct igb_ring *ring) |
2924 | | -{ |
2925 | | - lockdep_assert_held(&txring_txq(ring)->_xmit_lock); |
2926 | | - |
2927 | | - /* Force memory writes to complete before letting h/w know there |
2928 | | - * are new descriptors to fetch. |
2929 | | - */ |
2930 | | - wmb(); |
2931 | | - writel(ring->next_to_use, ring->tail); |
2932 | | -} |
2933 | | - |
2934 | | -static struct igb_ring *igb_xdp_tx_queue_mapping(struct igb_adapter *adapter) |
2935 | | -{ |
2936 | | - unsigned int r_idx = smp_processor_id(); |
2937 | | - |
2938 | | - if (r_idx >= adapter->num_tx_queues) |
2939 | | - r_idx = r_idx % adapter->num_tx_queues; |
2940 | | - |
2941 | | - return adapter->tx_ring[r_idx]; |
2942 | | -} |
2943 | | - |
2944 | | -static int igb_xdp_xmit_back(struct igb_adapter *adapter, struct xdp_buff *xdp) |
| 2919 | +int igb_xdp_xmit_back(struct igb_adapter *adapter, struct xdp_buff *xdp) |
2945 | 2920 | { |
2946 | 2921 | struct xdp_frame *xdpf = xdp_convert_buff_to_frame(xdp); |
2947 | 2922 | int cpu = smp_processor_id(); |
@@ -4888,7 +4863,7 @@ static void igb_free_all_tx_resources(struct igb_adapter *adapter) |
4888 | 4863 | * igb_clean_tx_ring - Free Tx Buffers |
4889 | 4864 | * @tx_ring: ring to be cleaned |
4890 | 4865 | **/ |
4891 | | -static void igb_clean_tx_ring(struct igb_ring *tx_ring) |
| 4866 | +void igb_clean_tx_ring(struct igb_ring *tx_ring) |
4892 | 4867 | { |
4893 | 4868 | u16 i = tx_ring->next_to_clean; |
4894 | 4869 | struct igb_tx_buffer *tx_buffer = &tx_ring->tx_buffer_info[i]; |
@@ -5007,7 +4982,7 @@ static void igb_free_all_rx_resources(struct igb_adapter *adapter) |
5007 | 4982 | * igb_clean_rx_ring - Free Rx Buffers per Queue |
5008 | 4983 | * @rx_ring: ring to free buffers from |
5009 | 4984 | **/ |
5010 | | -static void igb_clean_rx_ring(struct igb_ring *rx_ring) |
| 4985 | +void igb_clean_rx_ring(struct igb_ring *rx_ring) |
5011 | 4986 | { |
5012 | 4987 | u16 i = rx_ring->next_to_clean; |
5013 | 4988 |
|
@@ -8786,9 +8761,9 @@ static bool igb_cleanup_headers(struct igb_ring *rx_ring, |
8786 | 8761 | * order to populate the hash, checksum, VLAN, timestamp, protocol, and |
8787 | 8762 | * other fields within the skb. |
8788 | 8763 | **/ |
8789 | | -static void igb_process_skb_fields(struct igb_ring *rx_ring, |
8790 | | - union e1000_adv_rx_desc *rx_desc, |
8791 | | - struct sk_buff *skb) |
| 8764 | +void igb_process_skb_fields(struct igb_ring *rx_ring, |
| 8765 | + union e1000_adv_rx_desc *rx_desc, |
| 8766 | + struct sk_buff *skb) |
8792 | 8767 | { |
8793 | 8768 | struct net_device *dev = rx_ring->netdev; |
8794 | 8769 |
|
|
0 commit comments