|NetApp, Inc.||:||FAS6080 (FCAL Disks)|
|SPECsfs2008_nfs.v3||=||120011 Ops/Sec (Overall Response Time = 1.95 msec)|
|Tested By||NetApp, Inc.|
|Product Name||FAS6080 (FCAL Disks)|
|Hardware Available||December 2007|
|Software Available||September 2009|
|Date Tested||August 2009|
|SFS License Number||33|
The NetApp(R) FAS6000 represents the top of the line of the FAS family of storage systems with NetApp's unified storage architecture. It features two models: the FAS6040 and FAS6080. FAS6000 performance is driven by a 64-bit architecture that uses high throughput, low latency links and PCI Express for all internal and external data transfers. With our FAS6000 series and Data ONTAP 8.0 you can efficiently consolidate SAN, NAS, primary, and secondary storage on a single platform. You get the ultimate in scalability, versatility, and availability when you use our flagship FAS6000 series. Data ONTAP 8.0 is designed to provide customers with the next generation of features and functionality to ensure they are able to meet the demands of their increasingly virtualized workloads. The FAS6080 system used in this test was configured using the new larger aggregate capability. We've designed our systems to make them easy for you to install, configure, manage, and upgrade so you can quickly adapt your storage infrastructure to meet your changing business needs. To help you maximize staff productivity, all of our FAS systems share a unified storage architecture based on the Data ONTAP(R) operating system and use our integrated suite of application-aware manageability software. You can minimize the use of data center resources including power, cooling, and floor space-by taking advantage of our comprehensive set of storage-saving software features in Data ONTAP like Deduplication and Thin Provisioning (FlexVols). Additionally, our FlexShare software is a powerful quality-of-service tool for Data ONTAP(R) storage systems. It lets you assign individual priorities to multiple application workloads consolidated on a single system - you can tune how your system should prioritize its resources during conditions of high system utilization. When you add massive scalability to other characteristics of the FAS family, the result is our FAS6000 series, the ideal platform for the largest applications and storage consolidations.
|1||2||Storage Controller||NetApp||FAS6080A-IB-BS2-R5||FAS6080A ACT-ACT System|
|2||24||Disk Drives w/Shelf||NetApp||X94015A-ESH4-R5-C||DS14MK4 SHLF,ACPS,14x300GB,15K,HDD,ESH4,-C,R5|
|3||2||10 Gigabit Ethernet Adapter||NetApp||X1008A-R6||NIC,2-Port,10GbE,Fiber,PCIe,R6|
|4||2||Software License||NetApp||SW-T7C-NFS||NFS Software,T7C|
|OS Name and Version||Data ONTAP 8.0|
|Filesystem Software||Data ONTAP 8.0|
|vol options 'volume' no_atime_update||on||Disable atime updates (applied to all volumes)|
|Description||Number of Disks||Usable Size|
|300GB FCAL 15K RPM Disk Drives||324||65.4 TB|
|Number of Filesystems||2|
|Total Exported Capacity||64.64 TB|
|Filesystem Creation Options||64-bit aggregate option was selected during creation of the aggregates that housed the SFS filesystems on each controller.|
|Filesystem Config||Each filesystem was striped across 162 disks|
|Fileset Size||14002.2 GB|
The storage configuration consisted of 24 shelves, each wth 14 disks. Groups of four shelves were daisy-chained such that the outputs of each shelf was attached to the inputs of the next shelf in the group. The first shelf in each group had two 4Gbit/s FC-AL loop connections, each one connected to one of six FC-AL ports (integrated on the mainboard) on a different storage controller. Each storage controller was the primary owner of 12 shelves, with 162 disks in those shelves placed into a single 64-bit aggregate. Each aggregate was composed of 9 RAID-DP groups, each RAID-DP group was composed of 16 data disks and 2 parity disks. Within each aggregate, a flexible volume (utilizing Data ONTAP FlexVol (TM) technology) was created to hold the SFS filesystem for that controller. Each volume was striped across all disks in the aggregate where it resided. Each controller was the owner of a single volume, but the disks in each aggregate were dual-attached so that, in the event of a fault, they could be managed by the other controller via an alternate loop. A separate flexible volume residing in a three-disk root aggregate on each controller was created to hold the Data ONTAP operating system and system files. The remaining three disks owned by each controller were reserved for spares.
|Item No||Network Type||Number of Ports Used||Notes|
|1||Jumbo Frame 10 Gigabit Ethernet||2||Dual-port 10 gigabit ethernet PCIe adapter|
There was a single, dual-port 10 gigabit ethernet network adapter configured on each storage controller. Only one interface on each controller was enabled during the test. The active interfaces were configured to use jumbo frames (MTU size of 9000 bytes). All network interfaces were connected to a Cisco 6509 switch, which provided connectivity to the clients.
An MTU size of 9000 was set for all connections to the switch. Each load generator was connected to the network via a single 1 GigE port, which was configured with 2 separate IP addresses on separate subnets.
|Item No||Qty||Type||Description||Processing Function|
|1||8||CPU||2.6GHz AMD Opteron(tm) Processor 885, 1MB L2 cache||Networking, NFS protocol, WAFL filesystem, RAID/Storage drivers|
Each storage controller has four physical processors, each with two cores.
|Description||Size in GB||Number of Instances||Total GB||Nonvolatile|
|Storage controller mainboard memory)||32||2||64||V|
|Non-volatile memory on NVRAM PCIe adapter||2||2||4||NV|
|Grand Total Memory Gigabytes||68|
Each storage controller has main memory that is used for the operating system and for caching filesystem data. A separate, battery-backed RAM module on a separate PCIe adapter is used to provide stable storage for writes that have not yet been written to disk.
The WAFL filesystem logs writes and other filesystem modifying transactions to the NVRAM adapter. In an active-active configuration, as in the system under test, such transacations are also logged to the NVRAM on the partner storage controller so that, in the event of a storage controller failure, any transactions on the failed controller can be completed by the partner controller. Filesystem modifying NFS operations are not acknowledged until after the storage system has confirmed that the related data are stored in NVRAM adapters of both storage controllers (when both controllers are active). The battery backing the NVRAM ensures that any uncommitted transactions are preserved for at least 72 hours.
The system under test consisted of two FAS6080 storage controllers and 24 storage shelves, each with 14 300GB FC-AL disk drives. The two controllers were executing Data ONTAP 8.0 software operating in 7G mode. They were configured in an active-active cluster failover configuration, using the high-availability cluster software option in conjunction with an InfiniBand cluster interconnect provided on the NVRAM adapter. One dual-port 10 gigabit ethernet host bus adapter was present in a PCI-e expansion slot on each storage controller. The storage shelves were configured in groups of four shelves, which were connected to each other via two 4Gbit/s FCAL connections. The first shelf in each group had two 4Gbit/s FC-AL input connections, one to each storage controller. The system under test was connected to a 10 gigabit ethernet switch via 2 network ports (one active port per dual-port network adapter on each storage controller, with the second port unused).
All standard data protection features, including background RAID and media error scrubbing, software validated RAID checksumming, and double disk failure protection via double parity RAID (RAID-DP) were enabled during the test.
|1||25||IBM||IBM eServer LS21 Blade||Bladecenter 14 blade chasis with 1GB RAM and Linux operating system|
|2||1||Cisco||6509||Cisco Catalyst 6509 Ethernet Switch|
|LG Type Name||LG1|
|BOM Item #||1|
|Processor Name||Dual-Core AMD Opteron Processor 2216 HE|
|Processor Speed||2.4 GHz|
|Number of Processors (chips)||1|
|Number of Cores/Chip||2|
|Memory Size||1 GB|
|Operating System||RHEL4u5 kernel 2.6.9-55.0.2.ELsmp|
|Network Type||1 x Broadcom NetXtreme II Gigabit Ethernet BCM5706S(Rev 02)|
|Network Attached Storage Type||NFS V3|
|Number of Load Generators||25|
|Number of Processes per LG||26|
|Biod Max Read Setting||8|
|Biod Max Write Setting||8|
|LG No||LG Type||Network||Target Filesystems||Notes|
All filesystems were mounted on all clients, which were connected to the same physical and logical network.
Each load-generating client hosted 26 processes. The assignment of processes to filesystems and network interfaces was done such that they were evenly divided across all filesystems and network paths to the storage controllers. The filesystem data was striped evenly across all disks and FC-AL loops on the storage backend.
Other test notes: None.
NetApp is a registered trademark and "Data ONTAP", "FlexVol", and "WAFL" are trademarks of NetApp, Inc. in the United States and other countries. All other trademarks belong to their respective owners and should be treated as such.
Generated on Tue Aug 25 15:46:57 2009 by SPECsfs2008 HTML Formatter
Copyright © 1997-2008 Standard Performance Evaluation Corporation
First published at SPEC.org on 25-Aug-2009