SPECsfs2008_nfs.v3 Result

EMC Corporation : EMC VNX VG8 Gateway/EMC VNX5700, 5 X-Blades (including 1 stdby)
SPECsfs2008_nfs.v3 = 497623 Ops/Sec (Overall Response Time = 0.96 msec)


49030 0.4
98167 0.4
147368 0.5
196664 0.7
245819 0.8
295373 0.9
344974 1.1
394453 1.3
444426 1.7
497623 3.3
Performance Graph

Product and Test Information

Tested By EMC Corporation
Product Name EMC VNX VG8 Gateway/EMC VNX5700, 5 X-Blades (including 1 stdby)
Hardware Available February 2011
Software Available February 2011
Date Tested February 2011
SFS License Number 47
Licensee Locations Hopkinton, MA

The EMC VNX VG8 Gateway allows you to reach new heights for performance, availability, scalability, and flexibility as it delivers NAS simplicity and efficiency to EMC Symmetrix, EMC VNX and EMC CLARiiON block arrays. The configuration tested here consists of an EMC VNX VG8 with four active X-Blades and one standby X-Blade connected to four EMC VNX5700 block arrays.

Configuration Bill of Materials

Item No Qty Type Vendor Model/Name Description
1 1 Enclosure EMC VG8-DME0 VG8 Gateway X-Blade Enclosure with Dual X-Blades included
2 2 Enclosure EMC VG8-DMEX VG8 Gateway Empty X-Blade Enclosure
3 3 X-Blade EMC VG8-DM VG8 Gateway Single X-Blade
4 10 SLIC EMC NS-MXG20-A 10GbE Optical Dual Port
5 5 SLIC EMC NS-M8GF-A 4 FC 8Gbit Ports
6 1 Software EMC UNIF-VG8 Unisphere for File on VG8
7 4 Software EMC VG8-BASE-L VG8 Base File License (include CIFS and FTP)
8 4 Software EMC VG8-ADV-L VG8 Adv File License (include NFS, MPFS, and pNFS)
9 1 Control Station EMC NS-CSB Gateway Control Station Series B
10 1 Control Station EMC NS-CSB2 Gateway Control Station Series B - Secondary (standby)
11 4 Enclosure EMC VNX5700SPE VNX5700 SPE 4x6Gbps SAS BE - EMC rack
12 4 Enclosure EMC VNXSPS1KW VNX57/75 1.2KW SPS 15/25 drv vault DAE - EMC rack
13 4 Enclosure EMC VNX6GSDAE15P VNX57/75 15 x 3.5" 6Gbps SAS Primary DAE - EMC rack
14 32 Enclosure EMC VNX6GSDAE15 VNX 15x3.5" 6Gbps SAS expansion DAE - EMC rack
15 4 Disk EMC V-VX-VS1530 3.5" 300GB SAS 15K vault pack for 6GSDAE (4 drives)
16 2 Disk EMC VX-VS15-300 3.5" 15K 300GB 520BPS 6GB SAS
17 436 Disk EMC VX-VS6F-200 3.5" 200GB 6Gbps SAS flash drive
18 4 SLIC EMC VSPM8GFFEA VNX 4 port 8G FC IO Module Pair
19 4 Software EMC UNIB-V57 Unisphere for Block for a EMC VNX5700
20 4 Software EMC VNXOE-57 VNX OE R31 License Model for EMC VNX5700 Block
21 1 Switch EMC DS-5300B-8G DS-5300B 48/80 ports 8Gbps Fibre Channel base switch
22 2 Switch EMC DS5300B8G16PU DS-5300B 8Gbps Fibre Channel 16 port upgrade kit
23 1 Switch Cisco Nexus 5020 Cisco Nexus 10GbE 48-port IP Switch

Server Software

OS Name and Version VNX File OE 7.0
Other Software EMC VNX Control Station Linux 2.6.18-
Filesystem Software VNX UxFS File System

Server Tuning

Name Value Description
ufs updateAccTime 0 Disable access time updates
file asyncThresholdPercentage 80 Total cached dirty blocks for NFSv3 async writes
file dnlcNents 16777213 Directory name lookup cache size

Server Tuning Notes

Non-default params were used on the EMC VNX5700 block arrays:

Disks and Filesystems

Description Number of Disks Usable Size
3.5" 200GB 6Gbps SAS flash 436 74.7 TB
3.5" 300GB SAS 15K RPM drive 21 5.3 TB
Total 457 80.0 TB
Number of Filesystems 8
Total Exported Capacity 60243 GiB
Filesystem Type UxFS
Filesystem Creation Options "server_mount server_x -o noprefetch" - disable prefetching for the mounted file system
Filesystem Config Each file system is striped (256 KB element size), across 42 LUNs, fs0 through fs7. Two file systems were mounted and exported from each X-Blade. The two file systems mounted on a given X-Blade reside on a EMC VNX5700 storage array.
Fileset Size 57400.6 GB

420 of the flash drives are for data, 16 are for hot spares. The 436 flash data drives were divided equally among the 4 EMC VNX5700 block arrays, each EMC VNX5700 block array having 109 flash drives installed, 105 data drives and 4 hot spares. The flash data drives were configured as 5 drive RAID groups, for a total of 21 RAID groups per EMC VNX5700. Each RAID group hosted 4 4+1 RAID5 LUNs for a total of 84 LUNs per EMC VNX5700, 42 LUNs owned by SPA and 42 owned by SPB.

The file systems were created in a manner that used the entire usable capacity (after RAID implementation) of the flash data drives. After completion of the benchmark, the file system was 98% full. The benchmark required the number of flash drives be installed in order to attain successful completion.

The 21 SAS drives were divided among the 4 EMC VNX5700 block arrays. The primary EMC VNX5700 block array had 6 SAS drives installed, 5 configured 4+1 RAID5, 1 hot spare. This system hosted the VNX control volumes for all the X-Blades on the 4+1 RAID5 RAID group. The remaining 3 EMC VNX5700 block arrays were configured with 5 SAS drives, 4 drives for the VNX vault, 1 hot spare.

Each client mounted 2 file systems per X-Blade through each network interface of the VNX. Each client mounted a total of 8 file systems.

Network Configuration

Item No Network Type Number of Ports Used Notes
1 Jumbo 10GbE 8 There are 2 10GbE network interfaces in use per active X-Blade, 1 port per 10GbE SLIC.

Network Configuration Notes

All 10GbE network interfaces were connected to a Cisco Nexus 5020 switch.

Benchmark Network

An MTU size of 9000 was set for all connections to the switch. Each X-Blade was connected to the network via 2 ports. The LG1 class workload machines were connected with one port.

Processing Elements

Item No Qty Type Description Processing Function
1 4 CPU Single socket Intel six core Westmere (Xeon X5660) 2.8 GHz VG8 with QPI speed 6400 MHz for each X-Blade server. 1 chip active for the workload. (1 Processor in the standby X-Blade is not in the quantity.) NFS protocol, UxFS file system

Processing Element Notes

Each X-Blade has one physical processor. 4 active X-Blades for a total of 4 physical processors. The 2 control stations listed in the BOM contain processors that are not counted for in the list of processors here. The control stations are for management only. There is no function of the control station that is in the workload's data path. Also note that there are 2 control stations solely for high availability. Two control stations are not required to manage the EMC VNX VG8 Gateway.


Description Size in GB Number of Instances Total GB Nonvolatile
Each X-Blade main memory. (24 GB in the standby X-Blade not in the quantity) 24 4 96 V
EMC VNX5700 storage array battery backed memory. 18 GB per EMC VNX5700 SP. The EMC VNX5700 cache was configured with 10GB mirrored write cache. 512 MB read cache was configured per SP. The balance of the memory is reserved for the Operating Environment (OE R31) 18 8 144 NV
Grand Total Memory Gigabytes     240  

Memory Notes

The EMC VNX5700 write cache is backed with sufficient battery power to safely destage all the cached data to the EMC VNX5700 vault drives before shutting down the SP's in the event of a power failure in an orderly fashion.

Stable Storage

8 NFS file systems were used. Each file system was striped over 42 LUNs. Each VG8 X-Blade had 2 Fibre Channel connections to the DS-5300B FC switch. The remaining 2 Fibre Channel ports of the VG8 X-Blade were not connected to the FC switch. Each EMC VNX5700 storage array had 8 Fibre Channel connections, 4 per SP. In this configuration, NFS stable write and commit operations are not acknowledged until after the EMC VNX5700 storage array has acknowledged that the related data has been stored in stable storage (i.e. battery backed memory or disk).

System Under Test Configuration Notes

The system under test consisted of 4 active VNX VG8 Gateway X-Blades attached to 4 EMC VNX5700 Storage Arrays each with 8 FC links using a DS-5300B Fibre Channel switch. The X-Blades were running VNX File OE 7.0. 2 10GbE Ethernet ports per X-Blade were connected to the network.

Other System Notes

Failover is supported by an additional VNX X-Blade that operates in standby mode. In the event of any of the 4 active X-Blades failing, the standby unit takes over the function of the failed unit. The standby X-Blade does not contribute to the performance of the system and it is not included in the active components listed above.

Test Environment Bill of Materials

Item No Qty Vendor Model/Name Description
1 34 Dell Dell PowerEdge R610 Dell server with 12 GB RAM and the Linux operating system

Load Generators

LG Type Name LG1
BOM Item # 1
Processor Name Intel(R) Xeon(TM) E5530
Processor Speed 2.4 GHz
Number of Processors (chips) 1
Number of Cores/Chip 4
Memory Size 12 GB
Operating System CentOS 5.4 Linux 2.6.18-164.el5
Network Type 1 x Intel X520 10GbE

Load Generator (LG) Configuration

Benchmark Parameters

Network Attached Storage Type NFS V3
Number of Load Generators 34
Number of Processes per LG 64
Biod Max Read Setting 8
Biod Max Write Setting 8
Block Size AUTO

Testbed Configuration

LG No LG Type Network Target Filesystems Notes
1..34 LG1 1 /fs0, ..., /fs7 N/A

Load Generator Configuration Notes

All file systems were mounted on all clients, which were connected to the same physical and logical network.

Uniform Access Rule Compliance

Each client has all file systems mounted from each active X-Blade.

Other Notes

Failover is supported by an additional VNX X-Blade that operates in standby mode. In the event of the X-Blade failure, the standby unit takes over the function of any failed unit. The standby X-Blade does not contribute to the performance of the system and it is not included in the components listed above.

The EMC VNX5700's were configured with 18 GB of memory per SP. The memory is backed up with sufficient battery power to safely destage all cached data onto the vault drives before shutting down the SP's in an orderly manner in the event of a power failure.

Config Diagrams

Generated on Tue Feb 22 17:10:32 2011 by SPECsfs2008 HTML Formatter
Copyright © 1997-2008 Standard Performance Evaluation Corporation

First published at SPEC.org on 22-Feb-2011