Re: Problems with /proc/net/tcp6 - possible bug - ipv6

From: Eric Dumazet
Date: Sat Jan 22 2011 - 10:15:55 EST


Le samedi 22 janvier 2011 Ã 09:59 +0100, Eric Dumazet a Ãcrit :
> Le vendredi 21 janvier 2011 Ã 22:30 -0800, PK a Ãcrit :
> > Creating many ipv6 connections hits a ceiling on connections/fds ; okay, fine.
> >
> > But in my case I'm seeing millions of entries spring up within a few seconds and
> > then vanish within a few minutes, in /proc/net/tcp6 (vanish due to garbage
> > collection?)
> >
> > Furthermore I can trigger this easily on vanilla kernels from 2.6.36 to
> > 2.6.38-rc1-next-20110121 inside a ubuntu 10.10 amd64 vm, causing the kernel to
> > spew warnings. There is also some corruption in the logs (see kernel-sample.log
> > line 296), but that may be unrelated.
> >
> > More explanation, kernel config of the primary machine I saw this on, sample
> > ruby script to reproduce (inside the ubuntu VMs I apt-get and use ruby-1.9.1),
> > are located at
> > https://github.com/runningdogx/net6-bug
> >
> > Seems to only affect 64-bit. So far I have not been able to reproduce on 32-bit
> > ubuntu VMs of any kernel version.
> > Seems to only affect IPv6. So far I have not been able to reproduce using IPv4
> > connections (and watching /proc/net/tcp of course).
> > Does not trigger the bug if the connections are made to ::1. Only externally
> > routable local and global IPv6 addresses seem to cause problems.
> > Seems to have been introduced between 2.6.35 and 2.6.36 (see README on github
> > for more kernels I've tried)
> >
> > All the tested Ubuntu VMs are stock 10.10 userland, with vanilla kernels (the
> > latest ubuntu kernel is 2.6.35-something, and my initial test didn't show it
> > suffering from this problem)
> >
> > Originally noticed on separate Gentoo 64-bit non-vm system when doing web
> > benchmarking.
> >
> > not subscribed, so please keep me in cc although I'll try to follow the thread
> >
> >


I had some incidents, after hours of testing...

After following patch, I could not reproduce it.

I cant believe this bug was not noticed before today.

Thanks !

[PATCH] tcp: fix bug in listening_get_next()

commit a8b690f98baf9fb19 (tcp: Fix slowness in read /proc/net/tcp)
introduced a bug in handling of SYN_RECV sockets.

st->offset represents number of sockets found since beginning of
listening_hash[st->bucket].

We should not reset st->offset when iterating through
syn_table[st->sbucket], or else if more than ~25 sockets (if
PAGE_SIZE=4096) are in SYN_RECV state, we exit from listening_get_next()
with a too small st->offset

Next time we enter tcp_seek_last_pos(), we are not able to seek past
already found sockets.

Reported-by: PK <runningdoglackey@xxxxxxxxx>
CC: Tom Herbert <therbert@xxxxxxxxxx>
Signed-off-by: Eric Dumazet <eric.dumazet@xxxxxxxxx>
---
net/ipv4/tcp_ipv4.c | 1 -
1 file changed, 1 deletion(-)

diff --git a/net/ipv4/tcp_ipv4.c b/net/ipv4/tcp_ipv4.c
index 856f684..02f583b 100644
--- a/net/ipv4/tcp_ipv4.c
+++ b/net/ipv4/tcp_ipv4.c
@@ -1994,7 +1994,6 @@ static void *listening_get_next(struct seq_file *seq, void *cur)
}
req = req->dl_next;
}
- st->offset = 0;
if (++st->sbucket >= icsk->icsk_accept_queue.listen_opt->nr_table_entries)
break;
get_req:


--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/