Re: [PATCH] [3/5] Mark complex bitops.h inlines as __always_inline

From: H. Peter Anvin
Date: Thu Jan 08 2009 - 03:04:59 EST


Ingo Molnar wrote:
> * Hugh Dickins <hugh@xxxxxxxxxxx> wrote:
>
>> Hugh Dickins noticed that released gcc versions building the kernel with
>> CONFIG_OPTIMIZE_INLINING=y don't inline some of the bitops - sometimes
>> generating very inefficient pageflag tests, and many instances of
>> constant_test_bit().
>
> Could you quantify that please?
>
> We really dont want to reintroduce __always_inline just for performance /
> code size reasons. If GCC messes up and makes a larger / more inefficient
> kernel, GCC will be fixed. CONFIG_OPTIMIZE_INLINING is default-off, so
> enable it only if it improves your kernel.
>

There is one condition under which gcc simply won't know, and that is
when an inline is composed primarily of asm code. gcc, I believe,
creates a worst-case estimate based on the number of semicolons or
newlines (something that works semi-okayish on RISC), and thus tend to
vastly overestimate the size of an asm() on x86, where statements are
highly variable length. Hence it is probably always going to need
hints, unless the whole handling of inline assembly is revamped (which
would be good for scheduling, but I doubt it will happen.)

-hpa
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/