problems with optimisation

David Brown david.brown@hesbynett.no
Sun Dec 30 11:25:00 GMT 2012


On 29/12/12 23:26, Ángel González wrote:
> On 29/12/12 17:26, David Brown wrote:
>> With -Os, the compiler will obey normal "inline" directives (at least,
>> that is my experience when compiling C on the avr - I have not tried
>> C++ much on it).  It won't do any automatic extra inlining, except for
>> static functions that are only used once - which are always inlined as
>> this saves space.  Again, I don't know how that plays with template
>> functions or other C++ features.
>>
>> As far as I know, gcc uses weighting heuristics to decide whether to
>> do something the rcall you mentioned above, compared to using the
>> inlined code directly.  It is certainly not impossible that the
>> weightings are not optimal here.
>>
>> There is currently very little use of C++ with avr-gcc.  The avr port
>> maintainers and the avrlibc developers have little experience with
>> C++, and feel they have enough to do with just the C support.  But
>> there are a few people on the avr-gcc mailing list that work with C++,
>> and it is certainly worth posting there too - they may be able to give
>> suggestions.
>>
>> <https://lists.nongnu.org/mailman/listinfo/avr-gcc-list>
>>
>> mvh.,
>>
>> David
>
> I got good results (code apparently better) using -O3 in avr instead of
> -Os. Just the skipped instructions in the prologue and epiloques may be
> worth it. It may that since on avr you have one cycle per instruction
> (except branches), when optimizing for speed, you indirectly also
> optimize the number of instructions. However, I was using C, not C++, so
> the different way of coding could lead to worse optimizations.

It is not always easy to guess the best choice of optimisation flags. 
You are right that on the avr, small often means fast - and 
optimisations that first appear to make code larger (such as inlining 
functions that are used more than once, or loop unrolling) can lead to 
smaller code by avoiding prologues/epilogues, function call overheads, 
and other "bookkeeping" code.  Theoretically, the compiler knows this 
and will pick the smaller code with -Os.  In practice, it is a very hard 
problem, and there is a limit to the complexity (and accuracy) that can 
be achieved here.

On the bright side, gcc seems to be getting steadily smarter about these 
things - gcc 4.7 does partial function inlining and function 
specialisation in some cases.

Personally, I would like to see the distinction between "optimise for 
speed" and "optimise for size" disappear - or at least be reduced to a 
specialised flag (meaning "I /really/ don't care about speed - just make 
the code as small as possible", and vice-versa).  There are several 
reasons for this:

On modern "big" cpus, small means fast because small fits the fastest 
cache levels (including branch target buffers, prefetch buffers, etc.) 
best.  On an old 386 cpu it might make sense to unroll a loop - on an i7 
the fastest code will have the loop intact (unless unrolling gives 
additional optimisations).  And now the 386 will be deprecated...

On small cpus (like the avr), fast means small because fast means 
running fewer instructions.

In cases where it makes sense to bias on the side of size or speed, 
programmers are notoriously bad at making such decisions themselves. 
Hands up all developers who always profile their code before deciding 
which bits need optimisations :-)  The compiler, on the other hand, can 
do a reasonable job in many cases (see the -fipa-profile flag for an 
example).


On big cpus, the normal optimisation choice should be "make this code as 
fast as possible on this processor, maintaining all standards".  Other 
sensible options are "as fast as possible disregarding the fine print of 
IEEE standards" (the "-Ofast" flag), and "as fast as possible but still 
easy to debug" (the "-Og" flag).

On small cpus, the ideal flag would be something like "as fast as 
possible, but fitting within 32K code memory" - but I don't see that 
coming in the next version or two of gcc...



> I recommend giving gcc as much information as possible, and watch the
> generated code. I got gcc to perform a few tricky optimizations, and in
> one case, I manually unrolled a loop for him (otherwise, it didn't
> notice it could be optimized). If you see a very bad instance of code
> generation, open a bug. :)
> What difference do you have from -Os to -O3 ?
>

The more information the compiler gets, the better.  In particular, you 
always get better results if you can make your functions (and data) 
static - if the compiler can see that the functions don't escape (by 
taking their addresses), it can do far more optimisations.

mvh.,

David



More information about the Gcc-help mailing list