This is the mail archive of the
fortran@gcc.gnu.org
mailing list for the GNU Fortran project.
Re: Gfortran vs Ifort when using MPI
- From: "Johan Seland" <johan dot seland at gmail dot com>
- To: IainS <developer at sandoe-acoustics dot co dot uk>
- Cc: "Fortran List" <fortran at gcc dot gnu dot org>
- Date: Wed, 9 Jul 2008 15:24:56 +0200
- Subject: Re: Gfortran vs Ifort when using MPI
- Dkim-signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=gamma; h=domainkey-signature:received:received:message-id:date:from:to :subject:cc:in-reply-to:mime-version:content-type :content-transfer-encoding:content-disposition:references; bh=5SDE+P8O6U+tmT+5YCpOArKzMjS7Xn+dzP50s/YmpnY=; b=NU2etmFgZzwT+B0wLxJmNwJwikC9/VOCCwb+doKWZYBZDJTLhEjOR50s5G04PEzYrV UQPalNrVyUbFU4ZpdipCA/wFciLQZxRe9z2IG5ixdpLFgGkV2Ul3WJCdR/0RfQIBotzn f0X7eXAabJ5qe4xRxo1HiFjg7/IWT6NSkH4rY=
- Domainkey-signature: a=rsa-sha1; c=nofws; d=gmail.com; s=gamma; h=message-id:date:from:to:subject:cc:in-reply-to:mime-version :content-type:content-transfer-encoding:content-disposition :references; b=IoA9dAjWyfVsHDD57F9ykaqKVq/8RYQSKQop6kjPLYHx2LMwuf+3xHSls7I/qqEO1k r4xcnjJ/OBGxuXrA6DAG2S2NsjAk7eYlACICwuMsL+YsW4Plx/wooYa3CtOutfmpS1kn RUDOye2LIkD6TDxa1aooCu8y/gsUjjpeMUcng=
- References: <c86493ec0807080618x4baa4b6ewb86bb922cb67d621@mail.gmail.com> <4873708A.2070404@sbcglobal.net> <4873B17E.10407@net-b.de> <51CFCBC5-8A40-4F7F-990A-3E95821266FE@gmail.com> <49289939-5D0F-4C49-855B-3F6502FED238@sandoe-acoustics.co.uk> <c86493ec0807090038o3799bf5kdc309e95bf095e3@mail.gmail.com>
Just to keep you all informed.
I have made some progress during my instrumentation today. It seems
like the vectorization or FP-handling is not blame.
The deciding factor seems to be which Fortran compiler that was used
in compiling the MPI library. In other words, when using ifort to
compile MPI, the simulator can be compiled using gfortran and the
speeds are on par with ifort compiled simulator.
I will continue to investigate.
On Wed, Jul 9, 2008 at 9:38 AM, Johan Seland <johan.seland@gmail.com> wrote:
> Thanks for all the tips.
>
> As a first test, there was no difference when using march=native and
> ftree-vectorize. Furthermore I suspect fvect-cost-model is a 4.3
> feature only?
> When executed through mpirun, my binary is not dumping profile
> information (when compiled with -pg), so I will now start to manually
> instrument my code.
>
> I keep you informed.
>
> On Wed, Jul 9, 2008 at 9:04 AM, IainS <developer@sandoe-acoustics.co.uk> wrote:
>>
>> On 8 Jul 2008, at 21:25, FX wrote:
>>
>>>> Depending on the system, using the right -march makes a big difference.
>>>> Try -march=native (not supported on all targets). (If the compiler thinks
>>>> that there is no SSE, -march=native can help a lot; if I recall correctly,
>>>> for x86 the settings are rather conservative.
>>>
>>> For a 50x-100x performance difference, I suppose there is something
>>> *awfully* wrong going on, like wrong alignment or NaNs all over the place. I
>>> suggest a) checking the results, b) roughly profiling the gfortran code
>>> ("gfortran -pg" and using gprof) to get an idea where it spends so much
>>> time. It'll be much easier than us shooting in the dark.
>>
>> One thing you might check for in your profiling is presence of, and
>> different handling of, denormals - that can make huge difference on x86.
>> Iain
>>
>
>
>
> --
> Regards Johan Seland
>
--
Regards Johan Seland