Académique Documents
Professionnel Documents
Culture Documents
ISSN 2278-6856
Abstract
A decade ago CPUs speed could not be increased without
extra ordinary cooling and consequently hit a clock speed
barrier. Processors design then was switched to multi-core
architecture that minimizes energy consumption. The multicore architecture is introduced to improve computing
performance by providing hardware parallelism through
more CPU cores, each having restrained clock speed. This
has been a break through in High Performance Computing
(HPC). While more processor cores rendered effective
execution results, multi-core technology inaugurated an
extra layer of complexity for programming issues. To exploit
each core in a multi-core environment, application software
should be optimized by using multithreading. Multi-core
processors can even degrade the performance for single
threaded application due to reduction in clock speed. In this
paper we compare performance of multithreading finegrained and course-grained computational problems further
flavored as computation-intensive and data-intensive
problems
by using MPI and hybrid MPI+OpenMP
approach.
1. INTRODUCTION
Multi-core processors cluster is becoming more popular
than traditional Symmetric Multi Processors (SMP)
cluster. Commodity multi-core processors presents more
cost-effective solution to HPC community than expensive
SMP clusters. Moreover, multi-core processors deliver
better processing results as cores inside same CPU die
communicate with high speed interconnection whereas in
cluster of SMP nodes inter-processor communication takes
place through motherboard within a node, resulting a
comparatively slower communication. Both scientific &
business applications can be benefited from multi-core
processors [4]. Execution time can be minimized by
running multiple threads on multiple cores. Multiple cores
are effective for data parallel applications where same
code can run through multiple threads on different sets of
data as well as for functionally decomposed computation
intensive tasks where each task run in parallel on different
cores[6]. Prior to multi-core architecture, hyper threading
Page 64
Figure 1.
ISSN 2278-6856
2.RELATED WORK
MPI+OpenMP programming paradigm have been
reported in several published work that
mainly
experimented on SMP cluster. Jost & Jin [1] compare
MPI. OpenMP and hybrid approach by taking different
number of CPUs in SMP cluster. IBM SP systems are used
by Cappello & Daniel [3] to compare NAS parallel
benchmarks on SMP cluster. Authors also show a study of
communication and memory access patterns in the cluster.
Hits rates of L1 and L2 cache are studied by Wu & taylor
[2] on multi-core cluster by using NAS parallel
benchmarks SP and BT. Chen & Watson III [4] compare
results between Intel and AMD processors cluster by using
OpenMP directives and a locally developed threading
library.
Figure 2.
4.EXPERIMENTAL
RESULTS
PERFORMANCE COMPARISON
AND
ISSN 2278-6856
Problem Type
FGCI 1
FGCI 2
FGDI 1
FGDI 2
CGCI 1
CGCI 2
CGDI 1
CGDI 2
MPI
2824
2650
3911
4057
2189
2142
3358
3153
Hybrid
2477
2345
3523
3688
1855
1831
2895
2766
MPi
)
s
m
(
e
m
i
T
n
io
t
u
c
e
x
E
% Improvement
14 %
13 %
11 %
10 %
18 %
17 %
16 %
14 %
Hybrid
4500
4000
3500
3000
2500
2000
1500
1000
500
0
FGCI1 FGCI2 FGDI1 FGDI2 CGCI1 CGCI2 CGDI1 CGDI2
Problem Type
Figure 6.
Figure 3.
Page 66
Hybrid
C
G
D
I2
C
I2
C
I1
I2
I2
D
I1
C
G
C
G
C
G
FG
D
FG
D
FG
C
FG
C
I1
4000
3500
3000
2500
2000
1500
1000
500
0
I1
No. of Messages
Passing
MPI
Problem Type
Figure 8.
5.CONCLUSION
ISSN 2278-6856
REFERENCES
[1]. G. Jost, H. Jin, D. Mey, and F. Hatay, Comparing
the OpenMP, MPI, and Hybrid
Programming
Paradigms on an SMP Cluster, The Fifth European
Workshop on OpenMP (EWOMP03), Sep.2003
[2]. X. Wu and V. Taylor, Performance Characteristics
of Hybrid MPI/OpenMP Implementations of NAS
Parallel Benchmarks SP and BT on Large-scale
Multicore Clusters, International Workshop on
Performance
Modeling,
Benchmarking
and
Simulation of High Performance Computing Systems
, Vol. 38, No. 4, March 2011
[3]. F. Cappello and D. Etiemble, MPI versus
MPI+OpenMP on the IBM SP for the NAS
Benchmarks, IEEE conference on Supercomputing,
Nov. 2000, pp. 12-23
[4]. J. Chen, W. Watson III and W. Mao, MultiThreading Performance on Commodity Multi-core
Processors, 9th International Conference on High
Performance
Computing, March 2007, pp. 1-8
Page 67