Loading drivers/net/forcedeth.c +5 −4 Original line number Diff line number Diff line Loading @@ -1569,12 +1569,11 @@ static int nv_start_xmit(struct sk_buff *skb, struct net_device *dev) ((skb_shinfo(skb)->frags[i].size & (NV_TX2_TSO_MAX_SIZE-1)) ? 1 : 0); } spin_lock_irq(&np->lock); empty_slots = nv_get_empty_tx_slots(np); if ((empty_slots - np->tx_limit_stop) <= entries) { spin_unlock_irq(&np->lock); spin_lock_irq(&np->lock); netif_stop_queue(dev); spin_unlock_irq(&np->lock); return NETDEV_TX_BUSY; } Loading Loading @@ -1669,6 +1668,8 @@ static int nv_start_xmit(struct sk_buff *skb, struct net_device *dev) tx_flags_vlan = NV_TX3_VLAN_TAG_PRESENT | vlan_tx_tag_get(skb); } spin_lock_irq(&np->lock); /* set tx flags */ if (np->desc_ver == DESC_VER_1 || np->desc_ver == DESC_VER_2) { start_tx.orig->flaglen |= cpu_to_le32(tx_flags | tx_flags_extra); Loading @@ -1679,6 +1680,7 @@ static int nv_start_xmit(struct sk_buff *skb, struct net_device *dev) np->put_tx.ex = put_tx.ex; } spin_unlock_irq(&np->lock); dprintk(KERN_DEBUG "%s: nv_start_xmit: entries %d queued for transmission. tx_flags_extra: %x\n", dev->name, entries, tx_flags_extra); Loading @@ -1693,7 +1695,6 @@ static int nv_start_xmit(struct sk_buff *skb, struct net_device *dev) } dev->trans_start = jiffies; spin_unlock_irq(&np->lock); writel(NVREG_TXRXCTL_KICK|np->txrxctl_bits, get_hwbase(dev) + NvRegTxRxControl); pci_push(get_hwbase(dev)); return NETDEV_TX_OK; Loading Loading
drivers/net/forcedeth.c +5 −4 Original line number Diff line number Diff line Loading @@ -1569,12 +1569,11 @@ static int nv_start_xmit(struct sk_buff *skb, struct net_device *dev) ((skb_shinfo(skb)->frags[i].size & (NV_TX2_TSO_MAX_SIZE-1)) ? 1 : 0); } spin_lock_irq(&np->lock); empty_slots = nv_get_empty_tx_slots(np); if ((empty_slots - np->tx_limit_stop) <= entries) { spin_unlock_irq(&np->lock); spin_lock_irq(&np->lock); netif_stop_queue(dev); spin_unlock_irq(&np->lock); return NETDEV_TX_BUSY; } Loading Loading @@ -1669,6 +1668,8 @@ static int nv_start_xmit(struct sk_buff *skb, struct net_device *dev) tx_flags_vlan = NV_TX3_VLAN_TAG_PRESENT | vlan_tx_tag_get(skb); } spin_lock_irq(&np->lock); /* set tx flags */ if (np->desc_ver == DESC_VER_1 || np->desc_ver == DESC_VER_2) { start_tx.orig->flaglen |= cpu_to_le32(tx_flags | tx_flags_extra); Loading @@ -1679,6 +1680,7 @@ static int nv_start_xmit(struct sk_buff *skb, struct net_device *dev) np->put_tx.ex = put_tx.ex; } spin_unlock_irq(&np->lock); dprintk(KERN_DEBUG "%s: nv_start_xmit: entries %d queued for transmission. tx_flags_extra: %x\n", dev->name, entries, tx_flags_extra); Loading @@ -1693,7 +1695,6 @@ static int nv_start_xmit(struct sk_buff *skb, struct net_device *dev) } dev->trans_start = jiffies; spin_unlock_irq(&np->lock); writel(NVREG_TXRXCTL_KICK|np->txrxctl_bits, get_hwbase(dev) + NvRegTxRxControl); pci_push(get_hwbase(dev)); return NETDEV_TX_OK; Loading