Re: [PATCH] x86-32: Allocate irq stacks seperate from percpu area

From: Eric Dumazet
Date: Wed Oct 27 2010 - 09:58:00 EST


Le mercredi 27 octobre 2010 Ã 15:42 +0200, Tejun Heo a Ãcrit :
> Hello,
>
> On 10/27/2010 03:33 PM, Eric Dumazet wrote:
> > Le mercredi 27 octobre 2010 Ã 11:57 +0200, Peter Zijlstra a Ãcrit :
> >> On Wed, 2010-10-27 at 08:07 +0200, Eric Dumazet wrote:
> >>>> - irqctx = &per_cpu(hardirq_stack, cpu);
> >>>> + irqctx = (union irq_ctx *)__get_free_pages(THREAD_FLAGS, THREAD_ORDER);
> >>>
> >>> Hmm, then we lose NUMA affinity for stacks.
> >>
> >> I guess we could use:
> >>
> >> alloc_pages_node(cpu_to_node(cpu), THREAD_FLAGS, THREAD_ORDER);
> >>
> >>
> >
> > Anyway, I just discovered per_cpu data on my machine (NUMA capable) all
> > sit on a single node, if 32bit kernel used.
> >
> > # cat /proc/buddyinfo
> > Node 0, zone DMA 0 1 0 1 2 1 1 0 1 1 3
> > Node 0, zone Normal 94 251 81 16 3 2 1 2 1 2 187
> > Node 0, zone HighMem 113 88 47 36 18 5 4 3 2 0 268
> > Node 1, zone HighMem 154 97 43 16 9 4 3 2 3 2 482
> ...
> >
> > I presume node 1 having only HighMem could be the reason ?
>
> What does cpu_to_node() on each cpu say? Also, do you know why
> num_possible_cpus() is 32, not 16?
>

I dont know, machine is HP ProLiant BL460c G6
[ 0.000000] SMP: Allowing 32 CPUs, 16 hotplug CPUs

for_each_possible_cpu(cpu) {
pr_err("cpu=%d node=%d\n", cpu, cpu_to_node(cpu));
}

cpu=0 node=1
cpu=1 node=0
cpu=2 node=1
cpu=3 node=0
cpu=4 node=1
cpu=5 node=0
cpu=6 node=1
cpu=7 node=0
cpu=8 node=1
cpu=9 node=0
cpu=10 node=1
cpu=11 node=0
cpu=12 node=1
cpu=13 node=0
cpu=14 node=1
cpu=15 node=0
cpu=16 node=0
cpu=17 node=0
cpu=18 node=0
cpu=19 node=0
cpu=20 node=0
cpu=21 node=0
cpu=22 node=0
cpu=23 node=0
cpu=24 node=0
cpu=25 node=0
cpu=26 node=0
cpu=27 node=0
cpu=28 node=0
cpu=29 node=0
cpu=30 node=0
cpu=31 node=0




--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/