Re: [PATCH V4 5/8]macvtap: macvtap TX zero-copy support

From: Michael S. Tsirkin
Date: Wed May 04 2011 - 12:12:48 EST


On Wed, May 04, 2011 at 08:37:29AM -0700, Shirley Ma wrote:
> On Wed, 2011-05-04 at 17:58 +0300, Michael S. Tsirkin wrote:
> > On Wed, May 04, 2011 at 01:14:53AM -0700, Shirley Ma wrote:
> > > Only when buffer size is greater than GOODCOPY_LEN (256), macvtap
> > > enables zero-copy.
> > >
> > > Signed-off-by: Shirley Ma <xma@xxxxxxxxxx>
> >
> >
> > Looks good. Some thoughts below.
> >
> > > ---
> > >
> > > drivers/net/macvtap.c | 126
> > ++++++++++++++++++++++++++++++++++++++++++++----
> > > 1 files changed, 115 insertions(+), 11 deletions(-)
> > >
> > > diff --git a/drivers/net/macvtap.c b/drivers/net/macvtap.c
> > > index 6696e56..e8bc5ff 100644
> > > --- a/drivers/net/macvtap.c
> > > +++ b/drivers/net/macvtap.c
> > > @@ -60,6 +60,7 @@ static struct proto macvtap_proto = {
> > > */
> > > static dev_t macvtap_major;
> > > #define MACVTAP_NUM_DEVS 65536
> > > +#define GOODCOPY_LEN 256
> >
> > Scope with MACVTAP_ please.
> Ok.
>
> > For small packets, is it better to copy in vhost
> > and skip all the back and forth with callbacks? If yes, does
> > it make sense to put the constant above in some header
> > shared with vhost-net?
>
> skb is created in macvtap, the small packet copy is in skb, so I don't
> think we can do it in vhost here.

One simple way is to pass NULL instead of the pend pointer
for when we want macvtap to copy unconditionally.
vhost-net will know that packet is copied and can notify user
unconditionally.

I think this will solve the small packet regression you see ... no?

> > > static struct class *macvtap_class;
> > > static struct cdev macvtap_cdev;
> > >
> > > @@ -340,6 +341,7 @@ static int macvtap_open(struct inode *inode,
> > struct file *file)
> > > {
> > > struct net *net = current->nsproxy->net_ns;
> > > struct net_device *dev = dev_get_by_index(net, iminor(inode));
> > > + struct macvlan_dev *vlan = netdev_priv(dev);
> > > struct macvtap_queue *q;
> > > int err;
> > >
> > > @@ -369,6 +371,16 @@ static int macvtap_open(struct inode *inode,
> > struct file *file)
> > > q->flags = IFF_VNET_HDR | IFF_NO_PI | IFF_TAP;
> > > q->vnet_hdr_sz = sizeof(struct virtio_net_hdr);
> > >
> > > + /*
> > > + * so far only VM uses macvtap, enable zero copy between guest
> > > + * kernel and host kernel when lower device supports high
> > memory
> > > + * DMA
> > > + */
> > > + if (vlan) {
> > > + if (vlan->lowerdev->features & NETIF_F_ZEROCOPY)
> > > + sock_set_flag(&q->sk, SOCK_ZEROCOPY);
> > > + }
> > > +
> > > err = macvtap_set_queue(dev, file, q);
> > > if (err)
> > > sock_put(&q->sk);
> > > @@ -433,6 +445,80 @@ static inline struct sk_buff
> > *macvtap_alloc_skb(struct sock *sk, size_t prepad,
> > > return skb;
> > > }
> > >
> > > +/* set skb frags from iovec, this can move to core network code for
> > reuse */
> > > +static int zerocopy_sg_from_iovec(struct sk_buff *skb, const struct
> > iovec *from,
> > > + int offset, size_t count)
> > > +{
> > > + int len = iov_length(from, count) - offset;
> > > + int copy = skb_headlen(skb);
> > > + int size, offset1 = 0;
> > > + int i = 0;
> > > + skb_frag_t *f;
> > > +
> > > + /* Skip over from offset */
> > > + while (offset >= from->iov_len) {
> > > + offset -= from->iov_len;
> > > + ++from;
> > > + --count;
> > > + }
> > > +
> > > + /* copy up to skb headlen */
> > > + while (copy > 0) {
> > > + size = min_t(unsigned int, copy, from->iov_len -
> > offset);
> > > + if (copy_from_user(skb->data + offset1, from->iov_base
> > + offset,
> > > + size))
> > > + return -EFAULT;
> > > + if (copy > size) {
> > > + ++from;
> > > + --count;
> > > + }
> > > + copy -= size;
> > > + offset1 += size;
> > > + offset = 0;
> > > + }
> > > +
> > > + if (len == offset1)
> > > + return 0;
> > > +
> > > + while (count--) {
> > > + struct page *page[MAX_SKB_FRAGS];
> > > + int num_pages;
> > > + unsigned long base;
> > > +
> > > + len = from->iov_len - offset1;
> > > + if (!len) {
> > > + offset1 = 0;
> > > + ++from;
> > > + continue;
> > > + }
> > > + base = (unsigned long)from->iov_base + offset1;
> > > + size = ((base & ~PAGE_MASK) + len + ~PAGE_MASK) >>
> > PAGE_SHIFT;
> > > + num_pages = get_user_pages_fast(base, size, 0,
> > &page[i]);
> > > + if ((num_pages != size) ||
> > > + (num_pages > MAX_SKB_FRAGS -
> > skb_shinfo(skb)->nr_frags))
> > > + /* put_page is in skb free */
> > > + return -EFAULT;
> > > + skb->data_len += len;
> > > + skb->len += len;
> > > + skb->truesize += len;
> > > + while (len) {
> > > + f = &skb_shinfo(skb)->frags[i];
> > > + f->page = page[i];
> > > + f->page_offset = base & ~PAGE_MASK;
> > > + f->size = min_t(int, len, PAGE_SIZE -
> > f->page_offset);
> > > + skb_shinfo(skb)->nr_frags++;
> > > + /* increase sk_wmem_alloc */
> > > + atomic_add(f->size, &skb->sk->sk_wmem_alloc);
> >
> > One thing that gave me pause that we only accound for part of the page
> > here. I think we should count the whole page, no?
>
> The whole page is pinned, but it might not be for this sock only.

Right, but worst-case it is, so I think we should stay on the safe side
and limit what the user can do.

> I think I should change to atomic_add to outside the loop to save cost.

Sure, good idea.

> > > + base += f->size;
> > > + len -= f->size;
> > > + i++;
> > > + }
> > > + offset1 = 0;
> > > + ++from;
> > > + }
> > > + return 0;
> > > +}
> > > +
> > > /*
> > > * macvtap_skb_from_vnet_hdr and macvtap_skb_to_vnet_hdr should
> > > * be shared with the tun/tap driver.
> > > @@ -515,17 +601,19 @@ static int macvtap_skb_to_vnet_hdr(const
> > struct sk_buff *skb,
> > >
> > >
> > > /* Get packet from user space buffer */
> > > -static ssize_t macvtap_get_user(struct macvtap_queue *q,
> > > - const struct iovec *iv, size_t count,
> > > - int noblock)
> > > +static ssize_t macvtap_get_user(struct macvtap_queue *q, struct
> > msghdr *m,
> > > + const struct iovec *iv, unsigned long
> > total_len,
> > > + size_t count, int noblock)
> > > {
> > > struct sk_buff *skb;
> > > struct macvlan_dev *vlan;
> > > - size_t len = count;
> > > + unsigned long len = total_len;
> > > int err;
> > > struct virtio_net_hdr vnet_hdr = { 0 };
> > > int vnet_hdr_len = 0;
> > > + int copylen, zerocopy;
> > >
> > > + zerocopy = sock_flag(&q->sk, SOCK_ZEROCOPY) && (len >
> > GOODCOPY_LEN);
> > > if (q->flags & IFF_VNET_HDR) {
> > > vnet_hdr_len = q->vnet_hdr_sz;
> > >
> > > @@ -552,12 +640,28 @@ static ssize_t macvtap_get_user(struct
> > macvtap_queue *q,
> > > if (unlikely(len < ETH_HLEN))
> > > goto err;
> > >
> > > - skb = macvtap_alloc_skb(&q->sk, NET_IP_ALIGN, len,
> > vnet_hdr.hdr_len,
> > > - noblock, &err);
> > > + if (zerocopy)
> > > + /* There are 256 bytes to be copied in skb, so there
> > is enough
> > > + * room for skb expand head in case it is used.
> > > + * The rest buffer is mapped from userspace.
> > > + */
> > > + copylen = GOODCOPY_LEN;
> >
> > Just curious: where does the number 256 come from?
> > Also, as long as we are copying, should we care about
> > alignment?
>
> 256 makes the size big enough for any skb head expanding.
>
> That's my concern before.

I'm not sure I understand.
Could you tell me how do we know 256 is big enough for any skb head
expanding please?

> But guest should alignment the buffer already,
> after moving the pointer 256 bytes. It should still alignment, right?

I mean in the host. But whatever, it's not that important at this point.

> > > + else
> > > + copylen = len;
> > > +
> > > + skb = macvtap_alloc_skb(&q->sk, NET_IP_ALIGN, copylen,
> > > + vnet_hdr.hdr_len, noblock, &err);
> > > if (!skb)
> > > goto err;
> > >
> > > - err = skb_copy_datagram_from_iovec(skb, 0, iv, vnet_hdr_len,
> > len);
> > > + if (zerocopy)
> > > + err = zerocopy_sg_from_iovec(skb, iv, vnet_hdr_len,
> > count);
> > > + else
> > > + err = skb_copy_datagram_from_iovec(skb, 0, iv,
> > vnet_hdr_len,
> > > + len);
> > > + if (sock_flag(&q->sk, SOCK_ZEROCOPY))
> > > + memcpy(&skb_shinfo(skb)->ubuf, m->msg_control,
> > > + sizeof(struct skb_ubuf_info));
> > > if (err)
> > > goto err_kfree;
> > >
> > > @@ -579,7 +683,7 @@ static ssize_t macvtap_get_user(struct
> > macvtap_queue *q,
> > > kfree_skb(skb);
> > > rcu_read_unlock_bh();
> > >
> > > - return count;
> > > + return total_len;
> > >
> > > err_kfree:
> > > kfree_skb(skb);
> > > @@ -601,8 +705,8 @@ static ssize_t macvtap_aio_write(struct kiocb
> > *iocb, const struct iovec *iv,
> > > ssize_t result = -ENOLINK;
> > > struct macvtap_queue *q = file->private_data;
> > >
> > > - result = macvtap_get_user(q, iv, iov_length(iv, count),
> > > - file->f_flags & O_NONBLOCK);
> > > + result = macvtap_get_user(q, NULL, iv, iov_length(iv, count),
> > count,
> > > + file->f_flags & O_NONBLOCK);
> > > return result;
> > > }
> > >
> > > @@ -815,7 +919,7 @@ static int macvtap_sendmsg(struct kiocb *iocb,
> > struct socket *sock,
> > > struct msghdr *m, size_t total_len)
> > > {
> > > struct macvtap_queue *q = container_of(sock, struct
> > macvtap_queue, sock);
> > > - return macvtap_get_user(q, m->msg_iov, total_len,
> > > + return macvtap_get_user(q, m, m->msg_iov, total_len,
> > m->msg_iovlen,
> > > m->msg_flags & MSG_DONTWAIT);
> > > }
> > >
> > >
> >
> >
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/