public inbox for gcc-help@gcc.gnu.org
 help / color / mirror / Atom feed
* Gnu C++ & Open MP
@ 2008-01-09  1:40 burlen
  2008-01-09  2:58 ` Tom St Denis
  0 siblings, 1 reply; 4+ messages in thread
From: burlen @ 2008-01-09  1:40 UTC (permalink / raw)
  To: gcc-help

Hi,
I notice some strange result when I use Gnu OpenMP, I have an existing 
C++ serial code which has some time consuming for loops where some 
calculations are made on a large arrays of doubles,(12 arrays of length 
1E6 for example) and also some of these arrays are copied to create a 
vector field(also an array of doubles). This particular for loop is 
where the code takes 90% of its time and typically can take 1 or 2 
minutes here.  What I notice is that when I run open MP its actually 
slower than the serial code! Not by much but still slower. Here I use 
the command 'time' to compare. Also I notice by watching top, that in 
both cases only 1 core is taxed, while in the OpenMP build memory usage 
doubles(actually a second core goes up from 0% to about 10%). I have a 
quad core system(dual cpu dual core, and 4G of ram), so this result is 
totally unexpected. In my code I explicitly set the number of threads to 
4. I was expecting to see all 4 cores running at 80-90%. I have verified 
that 4 threads are being launched by runing the OpenMP build in GDB. 
What is going on here? how can I figure what the problem is?
Thanks Burlen



^ permalink raw reply	[flat|nested] 4+ messages in thread
* Gnu C++ & Open MP
@ 2008-01-09  1:58 burlen
  0 siblings, 0 replies; 4+ messages in thread
From: burlen @ 2008-01-09  1:58 UTC (permalink / raw)
  To: gcc-help

Hi,
I notice some strange result when I use Gnu OpenMP, I have an existing
C++ serial code which has some time consuming for loops where some
calculations are made on a large arrays of doubles,(12 arrays of length
1E6 for example) and also some of these arrays are copied to create a
vector field(also an array of doubles). This particular for loop is
where the code takes 90% of its time and typically can take 1 or 2
minutes here.  What I notice is that when I run open MP its actually
slower than the serial code! Not by much but still slower. Here I use
the command 'time' to compare. Also I notice by watching top, that in
both cases only 1 core is taxed, while in the OpenMP build memory usage
doubles(actually a second core goes up from 0% to about 10%). I have a
quad core system(dual cpu dual core, and 4G of ram), so this result is
totally unexpected. In my code I explicitly set the number of threads to
4. I was expecting to see all 4 cores running at 80-90%. I have verified
that 4 threads are being launched by runing the OpenMP build in GDB.
What is going on here? how can I figure what the problem is?
Thanks Burlen




^ permalink raw reply	[flat|nested] 4+ messages in thread

end of thread, other threads:[~2008-01-08 17:55 UTC | newest]

Thread overview: 4+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2008-01-09  1:40 Gnu C++ & Open MP burlen
2008-01-09  2:58 ` Tom St Denis
2008-01-09  9:29   ` burlen
2008-01-09  1:58 burlen

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for read-only IMAP folder(s) and NNTP newsgroup(s).