Base Results
Optimized Results
Base and Optimized
Base Results
Optimized Results
Base and Optimized
Base Results
Optimized Results
Base and Optimized
Base Results
Optimized Results
Base and Optimized
Base Results
Optimized Results
Base and Optimized
Base Results
Optimized Results
Base and Optimized
Base Results
Optimized Results
Base and Optimized
Base Results
Optimized Results
Base and Optimized
Manufacturer/Processor Type, Speed, Count, Threads, Processes
Includes the manufacturer/processor type, processor speed, number of processors, threads, and number of processes.
Move mouse over this column for each row to display additional information, including; manufacturer, system name, interconnect, MPI, affiliation, and submission date.

Computer System
Name and version of Message Passing Interface (MPI) implementation.

Run Type

Run Type, indicates whether the benchmark was a base run or was optimized.

Processors

Processors, this is the number of processors used in the benchmark, entered in the form by the benchmark submitter.

G-HPL ( system performance )
HPL, solves a randomly generated dense linear system of equations in double floating-point precision (IEEE 64-bit) arithmetic using MPI. The linear system matrix is stored in a two-dimensional block-cyclic fashion and multiple variants of code are provided for computational kernels and communication patterns. The solution method is LU factorization through Gaussian elimination with partial row pivoting followed by a backward substitution. Unit: Tera Flops per Second
EP-DGEMM ( per process )
Embarrassingly Parallel DGEMM, benchmark measures the floating-point execution rate of double precision real matrix-matrix multiply performed by the DGEMM subroutine from the BLAS (Basic Linear Algebra Subprograms). It is run in embarrassingly parallel manner - all computational processes perform the benchmark at the same time, the arithmetic average rate is reported. Unit: Giga Flops per Second
EP-STREAM ( per process )
The Embarrassingly Parallel STREAM benchmark is a simple synthetic benchmark program that measures sustainable memory bandwidth and the corresponding computation rate for simple vector kernels. It is run in embarrassingly parallel manner - all computational processes perform the benchmark at the same time, the average is computed. Unit: Giga Bytes per Second
EP-RandomAccess ( per process )
Embarrassingly Parallel Random Access - (also called GUPs), measures the rate at which the computer can update pseudo-random locations of its memory. The embarrassingly parallel version runs the code locally on each processor. No explicit communication is performed (but shared-memory effects might occur). Unit: Giga Updates per Second
EP-FFT ( per process )
Embarrassingly Parallel FFT, performs the same test as FFT but in embarrassingly parallel fashion - the code is run locally on each processor. No explicit communication is performed (but shared-memory effects might occur). Unit: Giga Flops per Second
Description above

See the row above this column for this columns description

Description below

See the row below this column for this columns description









To highlight the HPC Challenge Class 1 Awards which will be presented at SC14 we are not displaying the new submissions until the Awards session at SC14. The Awards session is on November 18th at 12:15 CST in room 273.
( 17.3 more days... )

Embarrassingly Parallel Results - Optimized Runs Only - 23 Systems - Generated on Sat Nov 1 05:54:15 2014
System Information
System - Processor - Speed - Count - Threads - Processes
EP-STREAMEP-Random
Access
EP-DGEMMEP-FFT
CopyScaleAddTriad
MA/PT/PS/PC/TH/PR/CM/CS/IC/IA/SDGB/sGB/sGB/sGB/sGup/sGFlop/sGFlop/s
Manufacturer: IBM
Processor Type: IBM PowerPC 440
Processor Speed: 0.7GHz
Processor Count: 1024
Threads: 1
Processses: 1024
System Name: Blue Gene/L
Interconnect: Custom
MPI: MPICH 1.0 customized for Blue Gene/L
Affiliation: Blue Gene Computational Center at IBM T.J. Watson Research Center
Submission Date: 04-11-05
IBM Blue Gene/L PowerPC 440   0.7GHz   1024   1   1024
0.72
0.72
0.82
0.84
0.00
2.47
0.22
Manufacturer: Cray Inc.
Processor Type: AMD Opteron
Processor Speed: 2.4GHz
Processor Count: 12960
Threads: 1
Processses: 25920
System Name: Red Storm/XT3
Interconnect: Cray custom
MPI: MPICH 2 v1.0.2
Affiliation: NNSA/Sandia National Laboratories
Submission Date: 11-10-06
Cray Inc. Red Storm/XT3 AMD Opteron   2.4GHz   12960   1   25920
2.07
2.10
2.04
2.08
0.01
4.40
0.62
Manufacturer: IBM
Processor Type: IBM PowerPC 440
Processor Speed: 0.7GHz
Processor Count: 131072
Threads: 1
Processses: 65536
System Name: Blue Gene/L
Interconnect: Custom Torus / Tree
MPI: MPICH2 1.0.1
Affiliation: National Nuclear Security Administration
Submission Date: 11-02-05
IBM Blue Gene/L PowerPC 440   0.7GHz   131072   1   65536
1.64
1.31
2.05
2.44
0.01
2.07
0.24
Manufacturer: Cray Inc.
Processor Type: AMD Opteron
Processor Speed: 2.6GHz
Processor Count: 10404
Threads: 1
Processses: 10404
System Name: XT3 Dual-Core
Interconnect: Cray SeaStar
MPI: xt-mpt 1.5.25
Affiliation: Oak Ridge National Lab
Submission Date: 11-06-06
Cray Inc. XT3 Dual-Core AMD Opteron   2.6GHz   10404   1   10404
2.54
2.24
2.05
2.55
0.01
4.79
0.65
Manufacturer: IBM
Processor Type: IBM PowerPC 440
Processor Speed: 0.7GHz
Processor Count: 131072
Threads: 1
Processses: 65536
System Name: Blue Gene/L
Interconnect: Custom Torus / Tree
MPI: MPICH2 1.0.1
Affiliation: National Nuclear Security Administration
Submission Date: 11-02-05
IBM Blue Gene/L PowerPC 440   0.7GHz   131072   1   65536
1.64
1.31
2.05
2.44
0.01
2.31
0.21
Manufacturer: IBM
Processor Type: IBM PowerPC 440
Processor Speed: 0.7GHz
Processor Count: 32768
Threads: 1
Processses: 16384
System Name: Blue Gene/L
Interconnect: Blue Gene Custom Interconnect
MPI: MPICH 1.1
Affiliation: IBM T.J. Watson Research Center
Submission Date: 11-04-05
IBM Blue Gene/L PowerPC 440   0.7GHz   32768   1   16384
1.64
1.31
2.06
2.44
0.01
2.31
0.21
Manufacturer: Cray Inc.
Processor Type: AMD Opteron
Processor Speed: 2.4GHz
Processor Count: 12960
Threads: 1
Processses: 25920
System Name: Red Storm/XT3
Interconnect: Seastar
MPI: xt-mpt/1.5.39 based on MPICH 2.0
Affiliation: DOE/NNSA/Sandia National Laboratories
Submission Date: 11-06-07
Cray Inc. Red Storm/XT3 AMD Opteron   2.4GHz   12960   1   25920
2.51
2.61
2.53
2.69
0.01
4.40
0.63
Manufacturer: Cray Inc.
Processor Type: AMD Opteron
Processor Speed: 2.4GHz
Processor Count: 12800
Threads: 1
Processses: 25600
System Name: Red Storm/XT3
Interconnect: Seastar
MPI: xt-mpt/1.5.39 based on MPICH 2.0
Affiliation: DOE/NNSA/Sandia National Laboratories
Submission Date: 11-06-07
Cray Inc. Red Storm/XT3 AMD Opteron   2.4GHz   12800   1   25600
2.86
2.87
2.89
3.01
0.01
4.40
0.62
Manufacturer: Cray Inc.
Processor Type: AMD Opteron
Processor Speed: 2.4GHz
Processor Count: 5208
Threads: 1
Processses: 5208
System Name: XT3
Interconnect: Cray Seastar
MPI: xt-mpt/1.3.07
Affiliation: Oak Ridge National Laboratory, DOE Office of Science
Submission Date: 11-10-05
Cray Inc. XT3 AMD Opteron   2.4GHz   5208   1   5208
5.66
4.82
4.72
5.63
0.02
4.41
0.59
Manufacturer: Cray Inc.
Processor Type: AMD Opteron
Processor Speed: 2.4GHz
Processor Count: 5208
Threads: 1
Processses: 5208
System Name: XT3
Interconnect: Cray Seastar
MPI: xt-mpt/1.3.07
Affiliation: Oak Ridge National Laboratories - DOE Office of Science
Submission Date: 11-12-05
Cray Inc. XT3 AMD Opteron   2.4GHz   5208   1   5208
5.66
4.82
4.72
5.63
0.02
4.41
0.59
Manufacturer: Cray Inc.
Processor Type: AMD Opteron
Processor Speed: 2.4GHz
Processor Count: 5208
Threads: 1
Processses: 5208
System Name: XT3
Interconnect: Cray Seastar
MPI: xt-mpt/1.3.07
Affiliation: Oak Ridge National Lab - DOD Office of Science
Submission Date: 11-12-05
Cray Inc. XT3 AMD Opteron   2.4GHz   5208   1   5208
5.66
4.77
4.73
5.61
0.02
4.42
0.61
Manufacturer: IBM
Processor Type: IBM Power5+
Processor Speed: 2.2GHz
Processor Count: 64
Threads: 1
Processses: 64
System Name: P5 P575+
Interconnect: HPS
MPI: poe 4.2.2.3
Affiliation: IBM
Submission Date: 05-08-06
IBM P5 P575+ Power5+   2.2GHz   64   1   64
8.77
9.35
11.23
11.96
0.02
8.39
0.76
Manufacturer: IBM
Processor Type: IBM Power5+
Processor Speed: 2.2GHz
Processor Count: 128
Threads: 1
Processses: 128
System Name: P5 P575+
Interconnect: HPS
MPI: poe 4.2.2.3
Affiliation: IBM
Submission Date: 05-08-06
IBM P5 P575+ Power5+   2.2GHz   128   1   128
9.70
9.35
12.06
11.97
0.02
8.46
0.76
Manufacturer: Cray Inc.
Processor Type: Cray X1E
Processor Speed: 1.13GHz
Processor Count: 248
Threads: 1
Processses: 248
System Name: mfeg8
Interconnect: Modified 2D Torus
MPI: mpt 2.4
Affiliation: Cray
Submission Date: 06-15-05
Cray Inc. mfeg8 Cray X1E   1.13GHz   248   1   248
10.75
10.82
13.38
13.23
0.14
13.56
1.84
Manufacturer: Cray Inc.
Processor Type: Cray X1E
Processor Speed: 1.13GHz
Processor Count: 1008
Threads: 1
Processses: 1008
System Name: X1
Interconnect: Cray Modified 2D torus
MPI: MPT
Affiliation: DOE/Office of Science/ORNL
Submission Date: 11-02-05
Cray Inc. X1 Cray E   1.13GHz   1008   1   1008
10.95
10.88
13.45
12.59
0.16
14.18
1.47
Manufacturer: Cray Inc.
Processor Type: Cray X1 MSP
Processor Speed: 0.8GHz
Processor Count: 60
Threads: 1
Processses: 60
System Name: X1
Interconnect: Cray modified 2D torus
MPI: MPT 2.4
Affiliation: U.S. Army Engineer Research and Development Center Major Shared Resource Center
Submission Date: 04-26-04
Cray Inc. X1 Cray MSP   0.8GHz   60   1   60
19.44
19.45
20.58
21.77
0.21


System Information
System - Processor - Speed - Count - Threads - Processes
EP-STREAMEP-Random
Access
EP-DGEMMEP-FFT
CopyScaleAddTriad
MA/PT/PS/PC/TH/PR/CM/CS/IC/IA/SDGB/sGB/sGB/sGB/sGup/sGFlop/sGFlop/s
Manufacturer: Cray Inc.
Processor Type: Cray X1 MSP
Processor Speed: 0.8GHz
Processor Count: 252
Threads: 1
Processses: 252
System Name: X1
Interconnect: X1
MPI: MPT 2.4
Affiliation: Oak Ridge National Laboratory
Submission Date: 04-26-04
Cray Inc. X1 Cray MSP   0.8GHz   252   1   252
19.56
18.90
21.09
21.74
0.21


Manufacturer: Cray Inc.
Processor Type: Cray X1 MSP
Processor Speed: 0.8GHz
Processor Count: 124
Threads: 1
Processses: 124
System Name: X1
Interconnect: Cray modified 2D torus
MPI: MPT.2.3.0.3
Affiliation: Army High Performance Computing Research Center (AHPCRC)
Submission Date: 05-03-04
Cray Inc. X1 Cray MSP   0.8GHz   124   1   124
19.51
19.34
21.18
21.75
0.21


Manufacturer: Cray Inc.
Processor Type: Cray X1 MSP
Processor Speed: 0.8GHz
Processor Count: 124
Threads: 1
Processses: 124
System Name: X1
Interconnect: Cray modified 2D torus
MPI: MPT 2.3.0.3
Affiliation: Army High Performance Computing Research Center (AHPCRC)
Submission Date: 05-05-04
Cray Inc. X1 Cray MSP   0.8GHz   124   1   124
19.51
19.34
21.18
21.75
0.21


Manufacturer: NEC
Processor Type: NEC SX-7
Processor Speed: 0.552GHz
Processor Count: 32
Threads: 1
Processses: 32
System Name: NEC SX-7
Interconnect: non
MPI: MPI/SX 7.0.6
Affiliation: Tohoku University, Information Synergy Center
Submission Date: 03-24-06
NEC SX-7   0.552GHz   32   1   32
27.29
27.27
27.91
27.64
0.32
8.62
1.49
Manufacturer: NEC
Processor Type: NEC SX-8
Processor Speed: 2GHz
Processor Count: 40
Threads: 1
Processses: 40
System Name: NEC SX-7C
Interconnect: IXS
MPI: MPI/SX 7.1.3
Affiliation: Tohoku University, Information Synergy Center
Submission Date: 03-24-06
NEC SX-7C SX-8   2GHz   40   1   40
39.62
36.46
37.31
36.00
0.50
15.95
7.91
Manufacturer: NEC
Processor Type: NEC SX-8
Processor Speed: 2GHz
Processor Count: 40
Threads: 8
Processses: 5
System Name: NEC SX-7C
Interconnect: IXS
MPI: MPI/SX 7.1.3
Affiliation: Tohoku University, Information Synergy Center
Submission Date: 03-24-06
NEC SX-7C SX-8   2GHz   40   8   5
319.80
283.92
266.47
288.60
0.59
114.67
11.43
Manufacturer: NEC
Processor Type: NEC SX-7
Processor Speed: 0.552GHz
Processor Count: 32
Threads: 16
Processses: 2
System Name: NEC SX-7
Interconnect: non
MPI: MPI/SX 7.0.6
Affiliation: Tohoku University, Information Synergy Center
Submission Date: 03-24-06
NEC SX-7   0.552GHz   32   16   2
424.26
422.58
448.82
452.36
0.39
140.94
32.47



 

Note:
Blank fields in the table above are from early benchmark runs that did not include that individual benchmark,
in particular G-RandomAccess, FFT and DGEMM.



Column Definitions
EP-STREAM ( per process )
The Embarrassingly Parallel STREAM benchmark is a simple synthetic benchmark program that measures sustainable memory bandwidth and the corresponding computation rate for simple numerical vector kernels. It is run in embarrassingly parallel manner - all computational processes perform the benchmark at the same time, the arithmetic average rate is reported. Unit: Giga Bytes per Second
EP-RandomAccess ( per process )
Embarrassingly Parallel RandomAccess, also called GUPs, measures the rate at which the computer can update pseudo-random locations of its memory - this rate is expressed in billions (giga) of updates per second (GUP/s). The embarrassingly parallel version runs the code locally on each process. No explicit communication is performed (but shared-memory effects might occur). Unit: Giga Updates per Second
EP-DGEMM ( per process )
The Embarrassingly Parallel DGEMM benchmark measures the floating-point execution rate of double precision real matrix-matrix multiply performed by the DGEMM subroutine from the BLAS (Basic Linear Algebra Subprograms). It is run in embarrassingly parallel manner - all computational processes perform the benchmark at the same time, the arithmetic average rate is reported. Unit: Giga Flops per Second
EP-FFT ( per process )
Embarrassingly Parallel FFT performs the same test as FFT but in embarrassingly parallel fashion - the code is run locally on each processor. No explicit communication is performed (but shared-memory effects might occur). Unit: Giga Flops per Second




Sat Nov 1 05:54:15 2014
0 seconds