SPECsfs2008_nfs.v3 Result ================================================================================ NSPLab(SM) Performed : SPECsfs2008 Reference Platform (NFSv3) Benchmarking SPECsfs2008_nfs.v3 = 1470 Ops/Sec (Overall Response Time = 5.40 msec) ================================================================================ Performance =========== Throughput Response (ops/sec) (msec) --------------------- ---------------- 148 1.7 300 2.3 450 2.9 601 3.2 750 3.8 904 5.3 1054 6.9 1207 9.2 1356 12.0 1470 18.1 ================================================================================ Product and Test Information ============================ Tested By NSPLab(SM) Performed Benchmarking Product Name SPECsfs2008 Reference Platform (NFSv3) Hardware Available January 2003 Software Available November 2007 Date Tested December 2007 SFS License Number 2851 Licensee Locations El Paso, TX USA The PowerEdge 1600SC Server was setup as a small workgroup file server using AX100 and AX150 arrays to provide RAID 5 storage. The storage was directly connected to the server. Configuration Bill of Materials =============================== Ite m No Qty Type Vendor Model/Name Description --- --- ---- ------ ---------- ----------- 1 1 Server Generic PowerEdge PowerEdge 1600SC server with single Xeon 1600SC 2.4GHz processor, running Red Hat Enterprise Linux 5.1 software. 2 2 HBA Generic LightPulse Generic LightPulse 4Gb/s LPe1105-n dual-channel HBA 3 1 Disk Cont Generic AX100 This array includes 12 146GB/7200 rpm SATA roller drives. 4 1 Disk Cont Generic AX150 This array includes 12 230GB/7200 rpm SATA roller drives. 5 1 Ethernet Generic PowerConne 16 port Gigabit Ethernet Switch Switch ct 2616 Server Software =============== OS Name and Version Linux RHEL 5.1 Other Software Linux RHEL 5.1 kernel 2.6.16 Filesystem Software EXT3 Server Tuning ============= Name Valu Description e ---- ---- ----------- net.core.rmem_max 1048 The maximum receive socket buffer size 576 net.core.wmem_max 1048 The maximum send socket buffer size 576 net.core. 1048 The default send socket buffer size wmem_default 576 net.core. 1048 The default receive socket buffer size rmem_default 576 vm. 4096 Force the VM to keep a minimum number of kilobytes free min_free_kbytes 0 vm. 1 The percentage of total memory at which the pdflush dirty_background_ background writeback daemon will start writing out dirty ratio data. vm. 1 Data that has been in memory for longer than this dirty_expire_cent interval will be written out next time a pdflush daemon isecs wakes up. Expressed in 100ths of a second. vm. 99 The interval between wakeups of the pdflush daemons. dirty_writeback_c Expressed in 100th of a second. entisecs vm.dirty_ratio 10 The percentage of total memory at which a process that is generating disk writes will start writing out dirty data. Server Tuning Notes ------------------- None Disks and Filesystems ===================== Numb er of D Usable Description isks Size ----------- ---- ------ AX100 array configuration includes 12 146GB disks configured as 12 1.4 TB (10+1)RAID-5 bound as a single LUN on which a 1.4TB EXT3 file system is built. One disk is used as a Hot Spare. AX150 array configuration includes 12 230GB disks configured as 12 2.2 TB (10+1)RAID-5 bound as a single LUN on which a 2.3TB EXT3 file system is built. One disk is used as a Hot Spare. This is an internal SCSI disk drive used for the Linux OS for system 1 146.0 use. GB Total 25 3.8 TB Number of Filesystems 2 Total Exported Capacity 3.3 TB Filesystem Type EXT3 Filesystem Creation Options EXT3 Filesystem Config A filesystem is created on a LUN on the AX100 and another on a LUN on the AX150. The filesystems are fs1 and fs2. Fileset Size 172.0 GB Each of the arrays, AX100 and AX150 contained 12 disks that were configured as (10+1) RAID-5 groups with 64KB RAID stripe size. The 12th disk was configured as hot spare. Network Configuration ===================== Number of Ports Item No Network Type Used Notes ------- ------------ ----------------- ----- 1 Gigabit Ethernet 1 Used Gigabit Ethernet port on motherboard Network Configuration Notes --------------------------- All network interfaces were connected to a the switch. Benchmark Network ================= The PowerEdge 1600SC server single on-board network interface was connected to the switch. Each of the LG single on-board network interface was connected to the switch. Processing Elements =================== Item No Qty Type Description Processing Function ---- --- ---- ----------- ------------------- 1 1 CPU Intel Xeon 2.4 Ghz, 1 chip, 1 core per chip, NFS protocol, EXT3 512KB of (I+D) L2 cache. filesystem 2 4 CPU 1.2 Ghz Intel Celeron, 2 per AX100/AX150 RAID processing disk array Processing Element Notes ------------------------ None Memory ====== Number of Description Size in GB Instances Total GB Nonvolatile ----------- ---------- --------------- -------- ----------- NAS server main memory 4 1 4 V Storage Processor main memory 1 2 2 NV Grand Total Memory Gigabytes 6 Memory Notes ------------ Each storage array has dual storage processor units that work as an active-active failover pair. The mirrored write cache is backed up with a battery unit capable of saving the write cache to disk in the event of a power failure. In the event of a storage array failure, the second storage processor unit is capable of saving all state that was managed by the first (and vise versa), even with a simultaneous power failure. When one of the storage processors or battery units are off-line, the system turns off the write cache and writes directly to disk before acknowledging any write operations. The battery unit can support the retention of data for 2 minutes which is sufficient to write all necessary data twice. Storage processor A could have written 99% of its memory to disk and then fail. In that case storage processor B has enough battery to store its copy of A's data as well as its own. Stable Storage ============== 2 NFS file systems were used. Each RAID group was bound in a single LUN on each array. Each file system was configured on one LUN: fs1 on AX150 and fs2 on AX100. Each array had 2 Fibre Channel connections to the server although only one connection was enabled. In this configuration, NFS stable write and commit operations are not acknowledged until after the storage array has acknowledged that the related data has been stored in stable storage (i.e. NVRAM or disk). System Under Test Configuration Notes ===================================== See above for the tuning params. Other System Notes ================== A minimal amount of Linux tuning was performed to get the benchmark to run successfully. The goal was to measure performance as close to out of the box (default tunings) as possible. Test Environment Bill of Materials ================================== Item No Qty Vendor Model/Name Description ------ --- ------ ---------- ----------- 1 4 Generic 430SC Workstation with single Dual Core Xeon 2.8 GHz with 1 GB RAM Load Generators =============== LG Type Name LG1-4 BOM Item # 1 Processor Name Xeon Dual Core Processor Speed 2.8 GHz Number of Processors (chips) 1 Number of Cores/Chip 2 Memory Size 1 GB Operating System Linux RHES 4.2 Network Type BCM95751 Load Generator (LG) Configuration ================================= Benchmark Parameters -------------------- Network Attached Storage Type NFS V3 Number of Load Generators 4 Number of Processes per LG 8 Biod Max Read Setting 2 Biod Max Write Setting 2 Block Size AUTO Testbed Configuration --------------------- LG No LG Type Network Target Filesystems Notes ----- ------- ------- ------------------ ----- 1..4 LG1 1 /fs1,/fs2 N/A Load Generator Configuration Notes ---------------------------------- Both filesystems were mounted on all clients, which were connected to the same physical and logical network. Uniform Access Rule Compliance ============================== Each client has the both file systems mounted from each of the 2 storage arrays: AX150 and AX100. Other Notes =========== (No other report notes) ================================================================================ Generated on Tue Mar 18 23:24:56 2008 by SPECsfs2008 ASCII Formatter Copyright (C) 1997-2008 Standard Performance Evaluation Corporation