Re: [NFS] [PATCH 005 of 9] knfsd: Be more selective in whichsockets lockd listens on.
From: Neil Brown
Date: Thu Jul 27 2006 - 22:31:03 EST
On Wednesday July 26, bfields@xxxxxxxxxxxx wrote:
> On Tue, Jul 25, 2006 at 11:54:47AM +1000, NeilBrown wrote:
> > @@ -112,6 +114,7 @@ lockd(struct svc_rqst *rqstp)
> > * Let our maker know we're running.
> > */
> > nlmsvc_pid = current->pid;
> > + nlmsvc_serv = serv;
>
> Nitpick: any reason not to just get rid of the local variable "serv"
> after that?
Well... it is used two more times in that function.... but in both
those cases it isn't really needed because we pass it to a function that
also gets rqstp, serv is always rqstp->rq_server. So there is room
for cleaning up there ... patch to follow.
>
> > @@ -224,8 +259,10 @@ lockd_up(void)
> > /*
> > * Check whether we're already up and running.
> > */
> > - if (nlmsvc_pid)
> > + if (nlmsvc_pid) {
> > + error = make_socks(nlmsvc_serv, proto);
> > goto out;
>
> ...
>
> > + if ((error = make_socks(serv, proto)) < 0) {
> > if (warned++ == 0)
> > printk(KERN_WARNING
> > "lockd_up: makesock failed, error=%d\n", error);
>
> The warning is printk'ed a little inconsistently. (If we care, maybe it
> should just go inside make_socks?)
So. Do we care? That's a hard one... I guess we do. I'll move the
message into make_socks.
>
> By the way, why don't most callers use the error returned from
> lockd_up()?
Most? I could 2 out of 5.
One of those is just getting an extra ref-count so it cannot fail.
The other - in nfsctl - is simply carelessness on my part.
I should fix that.... patch to follow.
However...
One would expect that if lockd_up fails, a matching lockd_down
wouldn't be needed. However nlmsvc_users is unconditionally
increased by lockd_up. That's a worry.
- The nfs client will only call lockd_down if lockd_up succeeded.
- NFSD currently will call lockd_down either way.
- The lockd reclaimer ignores the return value and always calls
lockd_down.
So the most common usage is to always call lockd_down, but I cannot
help feeling that is wrong. And 'fixing' the client code to do it
that way would make it very ugly....
Patch to follow :-)
>
> > diff .prev/fs/nfsd/nfssvc.c ./fs/nfsd/nfssvc.c
> > --- .prev/fs/nfsd/nfssvc.c 2006-07-24 15:14:31.000000000 +1000
> > +++ ./fs/nfsd/nfssvc.c 2006-07-24 15:15:04.000000000 +1000
> > @@ -134,6 +134,9 @@ static int killsig = 0; /* signal that w
> > static void nfsd_last_thread(struct svc_serv *serv)
> > {
> > /* When last nfsd thread exits we need to do some clean-up */
> > + struct svc_sock *svsk;
> > + list_for_each_entry(svsk, &serv->sv_permsocks, sk_list)
> > + lockd_down();
>
> So I guess it's a minor point, but: we take the trouble to only open tcp
> or udp sockets as necessary, but then won't close them down till all the
> mounts and nfsd's go away at which point we close them all down.
I thought about making that cleaner but couldn't quite motivate myself
to do it.... and it can always be done later (?).
>
> Would it be that bad just to always listen on both?
Some people seem to want to only listen on one.
Why? Maybe a security thing?
So once you have opened a protocol once, the "horse has bolted" and
closing it is no big deal?
dunno... any else go opinions on this?
Steve?
NeilBrown
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/