Loading drivers/net/tun.c +12 −18 Original line number Diff line number Diff line Loading @@ -103,13 +103,10 @@ struct tun_struct { uid_t owner; gid_t group; struct sk_buff_head readq; struct net_device *dev; struct fasync_struct *fasync; struct tap_filter txflt; struct sock *sk; struct socket socket; #ifdef TUN_DEBUG Loading Loading @@ -155,7 +152,7 @@ static int tun_attach(struct tun_struct *tun, struct file *file) tfile->tun = tun; tun->tfile = tfile; dev_hold(tun->dev); sock_hold(tun->sk); sock_hold(tun->socket.sk); atomic_inc(&tfile->count); out: Loading @@ -171,7 +168,7 @@ static void __tun_detach(struct tun_struct *tun) netif_tx_unlock_bh(tun->dev); /* Drop read queue */ skb_queue_purge(&tun->readq); skb_queue_purge(&tun->socket.sk->sk_receive_queue); /* Drop the extra count on the net device */ dev_put(tun->dev); Loading Loading @@ -340,7 +337,7 @@ static void tun_free_netdev(struct net_device *dev) { struct tun_struct *tun = netdev_priv(dev); sock_put(tun->sk); sock_put(tun->socket.sk); } /* Net device open. */ Loading Loading @@ -374,7 +371,7 @@ static netdev_tx_t tun_net_xmit(struct sk_buff *skb, struct net_device *dev) if (!check_filter(&tun->txflt, skb)) goto drop; if (skb_queue_len(&tun->readq) >= dev->tx_queue_len) { if (skb_queue_len(&tun->socket.sk->sk_receive_queue) >= dev->tx_queue_len) { if (!(tun->flags & TUN_ONE_QUEUE)) { /* Normal queueing mode. */ /* Packet scheduler handles dropping of further packets. */ Loading @@ -391,7 +388,7 @@ static netdev_tx_t tun_net_xmit(struct sk_buff *skb, struct net_device *dev) } /* Enqueue packet */ skb_queue_tail(&tun->readq, skb); skb_queue_tail(&tun->socket.sk->sk_receive_queue, skb); dev->trans_start = jiffies; /* Notify and wake up reader process */ Loading Loading @@ -492,13 +489,13 @@ static unsigned int tun_chr_poll(struct file *file, poll_table * wait) if (!tun) return POLLERR; sk = tun->sk; sk = tun->socket.sk; DBG(KERN_INFO "%s: tun_chr_poll\n", tun->dev->name); poll_wait(file, &tun->socket.wait, wait); if (!skb_queue_empty(&tun->readq)) if (!skb_queue_empty(&sk->sk_receive_queue)) mask |= POLLIN | POLLRDNORM; if (sock_writeable(sk) || Loading @@ -519,7 +516,7 @@ static inline struct sk_buff *tun_alloc_skb(struct tun_struct *tun, size_t prepad, size_t len, size_t linear, int noblock) { struct sock *sk = tun->sk; struct sock *sk = tun->socket.sk; struct sk_buff *skb; int err; Loading Loading @@ -787,7 +784,7 @@ static ssize_t tun_chr_aio_read(struct kiocb *iocb, const struct iovec *iv, current->state = TASK_INTERRUPTIBLE; /* Read frames from the queue */ if (!(skb=skb_dequeue(&tun->readq))) { if (!(skb=skb_dequeue(&tun->socket.sk->sk_receive_queue))) { if (file->f_flags & O_NONBLOCK) { ret = -EAGAIN; break; Loading Loading @@ -824,8 +821,6 @@ static void tun_setup(struct net_device *dev) { struct tun_struct *tun = netdev_priv(dev); skb_queue_head_init(&tun->readq); tun->owner = -1; tun->group = -1; Loading Loading @@ -991,7 +986,6 @@ static int tun_set_iff(struct net *net, struct file *file, struct ifreq *ifr) sk->sk_write_space = tun_sock_write_space; sk->sk_sndbuf = INT_MAX; tun->sk = sk; container_of(sk, struct tun_sock, sk)->tun = tun; tun_net_init(dev); Loading Loading @@ -1249,7 +1243,7 @@ static long tun_chr_ioctl(struct file *file, unsigned int cmd, break; case TUNGETSNDBUF: sndbuf = tun->sk->sk_sndbuf; sndbuf = tun->socket.sk->sk_sndbuf; if (copy_to_user(argp, &sndbuf, sizeof(sndbuf))) ret = -EFAULT; break; Loading @@ -1260,7 +1254,7 @@ static long tun_chr_ioctl(struct file *file, unsigned int cmd, break; } tun->sk->sk_sndbuf = sndbuf; tun->socket.sk->sk_sndbuf = sndbuf; break; default: Loading Loading @@ -1343,7 +1337,7 @@ static int tun_chr_close(struct inode *inode, struct file *file) tun = tfile->tun; if (tun) sock_put(tun->sk); sock_put(tun->socket.sk); put_net(tfile->net); kfree(tfile); Loading Loading
drivers/net/tun.c +12 −18 Original line number Diff line number Diff line Loading @@ -103,13 +103,10 @@ struct tun_struct { uid_t owner; gid_t group; struct sk_buff_head readq; struct net_device *dev; struct fasync_struct *fasync; struct tap_filter txflt; struct sock *sk; struct socket socket; #ifdef TUN_DEBUG Loading Loading @@ -155,7 +152,7 @@ static int tun_attach(struct tun_struct *tun, struct file *file) tfile->tun = tun; tun->tfile = tfile; dev_hold(tun->dev); sock_hold(tun->sk); sock_hold(tun->socket.sk); atomic_inc(&tfile->count); out: Loading @@ -171,7 +168,7 @@ static void __tun_detach(struct tun_struct *tun) netif_tx_unlock_bh(tun->dev); /* Drop read queue */ skb_queue_purge(&tun->readq); skb_queue_purge(&tun->socket.sk->sk_receive_queue); /* Drop the extra count on the net device */ dev_put(tun->dev); Loading Loading @@ -340,7 +337,7 @@ static void tun_free_netdev(struct net_device *dev) { struct tun_struct *tun = netdev_priv(dev); sock_put(tun->sk); sock_put(tun->socket.sk); } /* Net device open. */ Loading Loading @@ -374,7 +371,7 @@ static netdev_tx_t tun_net_xmit(struct sk_buff *skb, struct net_device *dev) if (!check_filter(&tun->txflt, skb)) goto drop; if (skb_queue_len(&tun->readq) >= dev->tx_queue_len) { if (skb_queue_len(&tun->socket.sk->sk_receive_queue) >= dev->tx_queue_len) { if (!(tun->flags & TUN_ONE_QUEUE)) { /* Normal queueing mode. */ /* Packet scheduler handles dropping of further packets. */ Loading @@ -391,7 +388,7 @@ static netdev_tx_t tun_net_xmit(struct sk_buff *skb, struct net_device *dev) } /* Enqueue packet */ skb_queue_tail(&tun->readq, skb); skb_queue_tail(&tun->socket.sk->sk_receive_queue, skb); dev->trans_start = jiffies; /* Notify and wake up reader process */ Loading Loading @@ -492,13 +489,13 @@ static unsigned int tun_chr_poll(struct file *file, poll_table * wait) if (!tun) return POLLERR; sk = tun->sk; sk = tun->socket.sk; DBG(KERN_INFO "%s: tun_chr_poll\n", tun->dev->name); poll_wait(file, &tun->socket.wait, wait); if (!skb_queue_empty(&tun->readq)) if (!skb_queue_empty(&sk->sk_receive_queue)) mask |= POLLIN | POLLRDNORM; if (sock_writeable(sk) || Loading @@ -519,7 +516,7 @@ static inline struct sk_buff *tun_alloc_skb(struct tun_struct *tun, size_t prepad, size_t len, size_t linear, int noblock) { struct sock *sk = tun->sk; struct sock *sk = tun->socket.sk; struct sk_buff *skb; int err; Loading Loading @@ -787,7 +784,7 @@ static ssize_t tun_chr_aio_read(struct kiocb *iocb, const struct iovec *iv, current->state = TASK_INTERRUPTIBLE; /* Read frames from the queue */ if (!(skb=skb_dequeue(&tun->readq))) { if (!(skb=skb_dequeue(&tun->socket.sk->sk_receive_queue))) { if (file->f_flags & O_NONBLOCK) { ret = -EAGAIN; break; Loading Loading @@ -824,8 +821,6 @@ static void tun_setup(struct net_device *dev) { struct tun_struct *tun = netdev_priv(dev); skb_queue_head_init(&tun->readq); tun->owner = -1; tun->group = -1; Loading Loading @@ -991,7 +986,6 @@ static int tun_set_iff(struct net *net, struct file *file, struct ifreq *ifr) sk->sk_write_space = tun_sock_write_space; sk->sk_sndbuf = INT_MAX; tun->sk = sk; container_of(sk, struct tun_sock, sk)->tun = tun; tun_net_init(dev); Loading Loading @@ -1249,7 +1243,7 @@ static long tun_chr_ioctl(struct file *file, unsigned int cmd, break; case TUNGETSNDBUF: sndbuf = tun->sk->sk_sndbuf; sndbuf = tun->socket.sk->sk_sndbuf; if (copy_to_user(argp, &sndbuf, sizeof(sndbuf))) ret = -EFAULT; break; Loading @@ -1260,7 +1254,7 @@ static long tun_chr_ioctl(struct file *file, unsigned int cmd, break; } tun->sk->sk_sndbuf = sndbuf; tun->socket.sk->sk_sndbuf = sndbuf; break; default: Loading Loading @@ -1343,7 +1337,7 @@ static int tun_chr_close(struct inode *inode, struct file *file) tun = tfile->tun; if (tun) sock_put(tun->sk); sock_put(tun->socket.sk); put_net(tfile->net); kfree(tfile); Loading