SPEC(R) MPIL2007 Summary Intel Corporation Intel Server System R2208WFTZS (Intel Xeon Gold 6148, 2.40 GHz) Sun Jul 23 18:49:40 2017 MPI2007 License: 13 Test date: Jul-2017 Test sponsor: Intel Corporation Hardware availability: Jul-2017 Tested by: Intel Corporation Software availability: Sep-2017 Base Base Base Peak Peak Peak Benchmarks Ranks Run Time Ratio Ranks Run Time Ratio -------------- ------ --------- --------- ------ --------- --------- 121.pop2 160 411 9.46 S 121.pop2 160 407 9.57 S 121.pop2 160 409 9.51 * 122.tachyon 160 378 5.14 * 122.tachyon 160 378 5.14 S 122.tachyon 160 380 5.12 S 125.RAxML 160 360 8.10 * 125.RAxML 160 360 8.11 S 125.RAxML 160 360 8.10 S 126.lammps 160 374 6.58 S 126.lammps 160 374 6.57 * 126.lammps 160 375 6.56 S 128.GAPgeofem 160 413 14.4 S 128.GAPgeofem 160 409 14.5 S 128.GAPgeofem 160 410 14.5 * 129.tera_tf 160 203 5.41 S 129.tera_tf 160 215 5.12 S 129.tera_tf 160 205 5.35 * 132.zeusmp2 160 229 9.25 S 132.zeusmp2 160 229 9.25 * 132.zeusmp2 160 230 9.20 S 137.lu 160 334 12.6 S 137.lu 160 336 12.5 * 137.lu 160 338 12.4 S 142.dmilc 160 267 13.8 * 142.dmilc 160 267 13.8 S 142.dmilc 160 266 13.8 S 143.dleslie 160 229 13.5 S 143.dleslie 160 228 13.6 * 143.dleslie 160 227 13.7 S 145.lGemsFDTD 160 438 10.1 S 145.lGemsFDTD 160 437 10.1 * 145.lGemsFDTD 160 435 10.1 S 147.l2wrf2 160 743 11.0 S 147.l2wrf2 160 737 11.1 * 147.l2wrf2 160 735 11.2 S ============================================================================== 121.pop2 160 409 9.51 * 122.tachyon 160 378 5.14 * 125.RAxML 160 360 8.10 * 126.lammps 160 374 6.57 * 128.GAPgeofem 160 410 14.5 * 129.tera_tf 160 205 5.35 * 132.zeusmp2 160 229 9.25 * 137.lu 160 336 12.5 * 142.dmilc 160 267 13.8 * 143.dleslie 160 228 13.6 * 145.lGemsFDTD 160 437 10.1 * 147.l2wrf2 160 737 11.1 * SPECmpiL_base2007 9.43 SPECmpiL_peak2007 Not Run BENCHMARK DETAILS ----------------- Type of System: Homogeneous Total Compute Nodes: 4 Total Chips: 8 Total Cores: 160 Total Threads: 320 Total Memory: 768 GB Base Ranks Run: 160 Minimum Peak Ranks: -- Maximum Peak Ranks: -- C Compiler: Intel C++ Composer XE 2017 for Linux Version 17.0.4.196 Build 20170411 C++ Compiler: Intel C++ Composer XE 2017 for Linux Version 17.0.4.196 Build 20170411 Fortran Compiler: Intel Fortran Composer XE 2017 for Linux Version 17.0.4.196 Build 20170411 Base Pointers: 64-bit Peak Pointers: Not Applicable MPI Library: Intel MPI Library 17u4 for Linux Other MPI Info: None Pre-processors: No Other Software: None Node Description: Endeavor Node =============================== HARDWARE -------- Number of nodes: 4 Uses of the node: compute Vendor: Intel Model: Intel Server System R2208WFTZS (Intel Xeon Gold 6148, 2.4 GHz) CPU Name: Intel Xeon Gold 6148 CPU(s) orderable: 1-2 chips Chips enabled: 2 Cores enabled: 40 Cores per chip: 20 Threads per core: 2 CPU Characteristics: Intel Turbo Boost Technology up to 3.7 GHz CPU MHz: 2400 Primary Cache: 32 KB I + 32 KB D on chip per core Secondary Cache: 1 MB I+D on chip per core L3 Cache: 27.5 MB I+D on chip per chip Other Cache: None Memory: 192 GB (12 x 16 GB 2Rx4 DDR4-2666 ECC Registered) Disk Subsystem: 1 x 800 GB SSD (INTEL SSDSC2BA80) Other Hardware: None Adapter: Intel Omni-Path Fabric Adapter 100 series Number of Adapters: 1 Slot Type: PCI-Express x16 Data Rate: 12.5 GB/s Ports Used: 1 Interconnect Type: Intel Omni-Path Fabric Adapter 100 series Adapter: Intel Omni-Path Edge Switch 100 series Number of Adapters: 1 Slot Type: PCI-Express x16 Data Rate: 12.5 GB/s Ports Used: 1 Interconnect Type: Intel Omni-Path Fabric Adapter 100 series SOFTWARE -------- Adapter: Intel Omni-Path Fabric Adapter 100 series Adapter Driver: IFS 10.4 Adapter Firmware: 0.9-46 Adapter: Intel Omni-Path Edge Switch 100 series Adapter Driver: IFS 10.4 Adapter Firmware: 0.9-46 Operating System: Oracle Linux Server release 7.3, Kernel 3.10.0-514.6.2.0.1.el7.x86_64.knl1 Local File System: Linux/xfs Shared File System: LFS System State: Multi-User Other Software: IBM Platform LSF Standard 9.1.1.1 Node Description: Lustre FS =========================== HARDWARE -------- Number of nodes: 11 Uses of the node: fileserver Vendor: Intel Model: Intel Server System R2224GZ4GC4 CPU Name: Intel Xeon E5-2680 CPU(s) orderable: 1-2 chips Chips enabled: 2 Cores enabled: 16 Cores per chip: 8 Threads per core: 2 CPU Characteristics: Intel Turbo Boost Technology disabled CPU MHz: 2700 Primary Cache: 32 KB I + 32 KB D on chip per core Secondary Cache: 2 MB I+D on chip per chip L3 Cache: 20 MB I+D on chip per chip Other Cache: None Memory: 64 GB (8 x 8GB 1600MHz Reg ECC DDR3) Disk Subsystem: 2.1 TB Other Hardware: None Adapter: Intel Omni-Path Fabric Adapter 100 series Number of Adapters: 1 Slot Type: PCI-Express x16 Data Rate: 12.5 GB/s Ports Used: 1 Interconnect Type: Intel Omni-Path Fabric Adapter 100 series SOFTWARE -------- Adapter: Intel Omni-Path Fabric Adapter 100 series Adapter Driver: IFS 10.4 Adapter Firmware: 0.9-46 Operating System: Redhat* Enterprise Linux* Server Release 7.2, Kernel 3.10.0-514.6.2.0.1.el7.x86_64.knl1 Local File System: None Shared File System: Lustre FS System State: Multi-User Other Software: None Interconnect Description: Intel Omni-Path ========================================= HARDWARE -------- Vendor: Intel Model: Intel Omni-Path 100 series Switch Model: Intel Omni-Path Edge Switch 100 series Number of Switches: 24 Number of Ports: 48 Data Rate: 12.5 GB/s Firmware: 0.9-46 Topology: Fat tree Primary Use: MPI traffic Interconnect Description: Intel Omni-Path ========================================= HARDWARE -------- Vendor: Intel Corporation Model: Intel Omni-Path 100 series Switch Model: Intel Omni-Path Edge Switch 100 series Number of Switches: 1 Number of Ports: 48 Data Rate: 12.5 GB/s Firmware: 0.9-46 Topology: Fat tree Primary Use: Cluster File System Submit Notes ------------ The config file option 'submit' was used. General Notes ------------- MPI startup command: mpiexec.hydra command was used to start MPI jobs. Software environment: export I_MPI_COMPATIBILITY=3 export I_MPI_FABRICS=shm:tmi export I_MPI_HYDRA_PMI_CONNECT=alltoall Network: Endeavour Omni-Path fabric consists of 48-port switches = 24 core switches connected to each leaf of the rack switch. Job placement: Each MPI job was assigned to a topologically compact set of nodes, i.e. the minimal needed number of leaf switches was used for each job = 1 switch for 40/80/160/320/640 ranks, 2 switches for 1280 and 1980 ranks. IBM Platform LSF was used for job submission. It has no impact on performance. Information can be found at: http://www.ibm.com Base Compiler Invocation ------------------------ C benchmarks: mpiicc C++ benchmarks: 126.lammps: mpiicpc Fortran benchmarks: mpiifort Benchmarks using both Fortran and C: mpiicc mpiifort Base Portability Flags ---------------------- 121.pop2: -DSPEC_MPI_CASE_FLAG 126.lammps: -DMPICH_IGNORE_CXX_SEEK Base Optimization Flags ----------------------- C benchmarks: -O3 -xCORE-AVX512 -no-prec-div -ipo C++ benchmarks: 126.lammps: -O3 -xCORE-AVX512 -no-prec-div -ipo Fortran benchmarks: -O3 -xCORE-AVX512 -no-prec-div -ipo Benchmarks using both Fortran and C: -O3 -xCORE-AVX512 -no-prec-div -ipo The flags file that was used to format this result can be browsed at http://www.spec.org/mpi2007/flags/EM64T_Intel140_flags.20170822.html You can also download the XML flags source by saving the following link: http://www.spec.org/mpi2007/flags/EM64T_Intel140_flags.20170822.xml SPEC and SPEC MPI are registered trademarks of the Standard Performance Evaluation Corporation. All other brand and product names appearing in this result are trademarks or registered trademarks of their respective holders. ----------------------------------------------------------------------------- For questions about this result, please contact the tester. For other inquiries, please contact webmaster@spec.org. Copyright 2006-2010 Standard Performance Evaluation Corporation Tested with SPEC MPI2007 v2.0.1. Report generated on Tue Aug 22 18:38:20 2017 by MPI2007 ASCII formatter v1463. Originally published on 22 August 2017.