Loading include/linux/if_vlan.h +2 −3 Original line number Diff line number Diff line Loading @@ -119,7 +119,7 @@ extern u16 vlan_dev_vlan_id(const struct net_device *dev); extern int __vlan_hwaccel_rx(struct sk_buff *skb, struct vlan_group *grp, u16 vlan_tci, int polling); extern int vlan_hwaccel_do_receive(struct sk_buff *skb); extern void vlan_hwaccel_do_receive(struct sk_buff *skb); extern gro_result_t vlan_gro_receive(struct napi_struct *napi, struct vlan_group *grp, unsigned int vlan_tci, struct sk_buff *skb); Loading Loading @@ -147,9 +147,8 @@ static inline int __vlan_hwaccel_rx(struct sk_buff *skb, struct vlan_group *grp, return NET_XMIT_SUCCESS; } static inline int vlan_hwaccel_do_receive(struct sk_buff *skb) static inline void vlan_hwaccel_do_receive(struct sk_buff *skb) { return 0; } static inline gro_result_t Loading net/8021q/vlan_core.c +1 −2 Original line number Diff line number Diff line Loading @@ -35,7 +35,7 @@ int __vlan_hwaccel_rx(struct sk_buff *skb, struct vlan_group *grp, } EXPORT_SYMBOL(__vlan_hwaccel_rx); int vlan_hwaccel_do_receive(struct sk_buff *skb) void vlan_hwaccel_do_receive(struct sk_buff *skb) { struct net_device *dev = skb->dev; struct vlan_rx_stats *rx_stats; Loading Loading @@ -69,7 +69,6 @@ int vlan_hwaccel_do_receive(struct sk_buff *skb) break; } u64_stats_update_end(&rx_stats->syncp); return 0; } struct net_device *vlan_dev_real_dev(const struct net_device *dev) Loading net/core/dev.c +2 −2 Original line number Diff line number Diff line Loading @@ -2841,8 +2841,8 @@ static int __netif_receive_skb(struct sk_buff *skb) if (!netdev_tstamp_prequeue) net_timestamp_check(skb); if (vlan_tx_tag_present(skb) && vlan_hwaccel_do_receive(skb)) return NET_RX_SUCCESS; if (vlan_tx_tag_present(skb)) vlan_hwaccel_do_receive(skb); /* if we've gotten here through NAPI, check netpoll */ if (netpoll_receive_skb(skb)) Loading Loading
include/linux/if_vlan.h +2 −3 Original line number Diff line number Diff line Loading @@ -119,7 +119,7 @@ extern u16 vlan_dev_vlan_id(const struct net_device *dev); extern int __vlan_hwaccel_rx(struct sk_buff *skb, struct vlan_group *grp, u16 vlan_tci, int polling); extern int vlan_hwaccel_do_receive(struct sk_buff *skb); extern void vlan_hwaccel_do_receive(struct sk_buff *skb); extern gro_result_t vlan_gro_receive(struct napi_struct *napi, struct vlan_group *grp, unsigned int vlan_tci, struct sk_buff *skb); Loading Loading @@ -147,9 +147,8 @@ static inline int __vlan_hwaccel_rx(struct sk_buff *skb, struct vlan_group *grp, return NET_XMIT_SUCCESS; } static inline int vlan_hwaccel_do_receive(struct sk_buff *skb) static inline void vlan_hwaccel_do_receive(struct sk_buff *skb) { return 0; } static inline gro_result_t Loading
net/8021q/vlan_core.c +1 −2 Original line number Diff line number Diff line Loading @@ -35,7 +35,7 @@ int __vlan_hwaccel_rx(struct sk_buff *skb, struct vlan_group *grp, } EXPORT_SYMBOL(__vlan_hwaccel_rx); int vlan_hwaccel_do_receive(struct sk_buff *skb) void vlan_hwaccel_do_receive(struct sk_buff *skb) { struct net_device *dev = skb->dev; struct vlan_rx_stats *rx_stats; Loading Loading @@ -69,7 +69,6 @@ int vlan_hwaccel_do_receive(struct sk_buff *skb) break; } u64_stats_update_end(&rx_stats->syncp); return 0; } struct net_device *vlan_dev_real_dev(const struct net_device *dev) Loading
net/core/dev.c +2 −2 Original line number Diff line number Diff line Loading @@ -2841,8 +2841,8 @@ static int __netif_receive_skb(struct sk_buff *skb) if (!netdev_tstamp_prequeue) net_timestamp_check(skb); if (vlan_tx_tag_present(skb) && vlan_hwaccel_do_receive(skb)) return NET_RX_SUCCESS; if (vlan_tx_tag_present(skb)) vlan_hwaccel_do_receive(skb); /* if we've gotten here through NAPI, check netpoll */ if (netpoll_receive_skb(skb)) Loading