SPECsfs2008_cifs Result ================================================================================ NetApp, Inc. : FAS3140 (FCAL Disks) SPECsfs2008_cifs = 55476 Ops/Sec (Overall Response Time = 1.84 msec) ================================================================================ Performance =========== Throughput Response (ops/sec) (msec) --------------------- ---------------- 5506 1.0 11023 1.2 16545 1.4 22072 1.5 27595 1.7 33156 1.9 38754 2.3 44242 2.5 50066 2.9 55476 3.9 ================================================================================ Product and Test Information ============================ Tested By NetApp, Inc. Product Name FAS3140 (FCAL Disks) Hardware Available Sept 2009 Software Available March 2010 Date Tested March 2010 SFS License Number 33 Licensee Locations Sunnyvale, CA USA The NetApp(R) FAS3100 series is the midrange of the FAS family of storage systems with our Unified Storage Architecture, which is flexible enough to handle primary and/or secondary storage needs for NAS and/or SAN implementations. The FAS3100 series features three models: the FAS3140, FAS3160, and FAS3170. Performance is driven by a 64-bit processing design that uses high throughput, low latency links and PCI Express for all internal and external data transfers. All FAS3100 systems scale to as many as 36 Ethernet ports or 40 Fibre Channel ports, including support for both 8Gb Fibre Channel and 10Gb Ethernet. These systems are housed in an efficient, HA-ready 6U form factor which allows for a shared chassis and backplane for dual controllers by sharing of common system power resources. The FAS3140, which was used in these benchmark tests, scales to a maximum of 420 disk drives and 840TB of capacity. Configuration Bill of Materials =============================== Ite m No Qty Type Vendor Model/Name Description --- --- ---- ------ ---------- ----------- 1 2 Storage Controller NetApp FAS3140A-IB- FAS3140A,IB,ACT-ACT,OS,R5 BASE-R5 2 1 Controller Chassis NetApp FAS3140A- FAS3140,ACT-ACT,Chassis,AC CHASSIS-R5-C PS,-C,R5 3 16 Disk Drives w/ NetApp X94015A-ESH4-R5- DS14MK4 SHLF,ACPS,14x300GB, Shelf C 15K,HDD,ESH4,-C,R5 4 2 Gigabit Ethernet NetApp X1039A-R6 NIC,2-port,GbE,Copper,PCIe, Adapter R6 5 2 Software License NetApp SW-T3C-CIFS CIFS Software,T3C Server Software =============== OS Name and Version Data ONTAP 7.3.3 Other Software None Filesystem Software Data ONTAP 7.3.3 Server Tuning ============= Name Value Description ---- ----- ----------- vol options 'volume' on Disable atime updates (applied to all no_atime_update volumes) Server Tuning Notes ------------------- N/A Disks and Filesystems ===================== Description Number of Disks Usable Size ----------- --------------- ----------- 300GB FCAL 15K RPM Disk Drives 224 43.6 TB Total 224 43.6 TB Number of Filesystems 4 Total Exported Capacity 42.66 TB Filesystem Type WAFL Filesystem Creation Options Default Filesystem Config Each filesystem was striped across 56 disks Fileset Size 6440.9 GB The storage configuration consisted of 16 shelves, each with 14 disks. Groups of four shelves were daisy-chained such that the outputs of each shelf were attached to the inputs of the next shelf in the group. The first shelf in each group had two 4Gbit/s FC-AL loop connections, each one connected to one of four FC-AL ports (integrated on the mainboard) on a different storage controller. Each storage controller was the primary owner of 8 shelves, with the disks in those shelves divided into two disk pools or "aggregates". Each aggregate was composed of 4 RAID-DP groups, each RAID-DP group was composed of 12 data disks and 2 parity disks. Within each aggregate, a flexible volume (utilizing Data ONTAP FlexVol (TM) technology) was created to hold the SFS filesystems for that controller. Each volume was striped across all disks in the aggregate where it resided. Each controller was the owner of two volumes/filesystems, but the disks in each aggregate were dual-attached so that, in the event of a fault, they could be managed by the other controller via an alternate loop. A separate flexible volume residing in one of the aggregates owned by each controller held the Data ONTAP operating system and system files. Network Configuration ===================== Item Number of No Network Type Ports Used Notes ------ ------------ -------------- ----- 1 Jumbo Frame Gigabit 4 Dual-port gigabit ethernet Ethernet PCI-e adapter Network Configuration Notes --------------------------- There were two gigabit ethernet network interfaces on each storage controller. The interfaces were configured to use jumbo frames (MTU size of 9000 bytes). All network interfaces were connected to a Cisco 6509 switch, which provided connectivity to the clients. Benchmark Network ================= An MTU size of 9000 was set for all connections to the switch. Each load generator was connected to the network via a single 1 GigE port, which was configured with 4 separate IP addresses on separate subnets. Processing Elements =================== Item No Qty Type Description Processing Function ---- --- ---- ----------- ------------------- 1 2 CPU 2.4GHz Dual-Core AMD Networking, CIFS protocol, WAFL Opteron(tm) Processor 2216, filesystem, RAID/Storage drivers 2MB L2 cache Processing Element Notes ------------------------ Each storage controller has one physical processor, each with two processing cores. Memory ====== Size in Number of Total Nonvola Description GB Instances GB tile ----------- -------- ----------- ------- ------- Storage controller mainboard memory 4 2 8 V Storage controller integrated NVRAM module 0.5 2 1 NV Grand Total Memory Gigabytes 9 Memory Notes ------------ Each storage controller has main memory that is used for the operating system and for caching filesystem data. A separate, integrated battery-backed RAM module is used to provide stable storage for writes that have not yet been written to disk. Stable Storage ============== The WAFL filesystem logs writes and other filesystem modifying transactions to the integrated NVRAM module. In an active-active configuration, as in the system under test, such transactions are also logged to the NVRAM on the partner storage controller so that, in the event of a storage controller failure, any transactions on the failed controller can be completed by the partner controller. Filesystem modifying CIFS operations are not acknowledged until after the storage system has confirmed that the related data are stored in NVRAM modules of both storage controllers (when both controllers are active). The battery backing the NVRAM ensures that any uncommitted transactions are preserved for at least 72 hours. System Under Test Configuration Notes ===================================== The system under test consisted of two FAS3140 storage controllers housed in a single 6U chassis and 16 storage shelves, each with 14 300GB FC-AL disk drives. The two controllers were configured in an active-active cluster configuration using the high-availability cluster software option in conjunction with an InfiniBand cluster interconnect on the backplane of the shared chassis. A dual-port gigabit ethernet host bus adapter was present in a PCI-e expansion slot on each storage controller. The storage shelves were configured in groups of four and were connected to each other via two 4Gbit/s FCAL connections. The first shelf in each group had one 4Gbit/s FC-AL connection to each storage controller. The system under test was connected to a gigabit ethernet switch via 4 network ports (two per storage controller). Other System Notes ================== All standard data protection features, including background RAID and media error scrubbing, software validated RAID checksumming, and double disk failure protection via double parity RAID (RAID-DP) were enabled during the test. Test Environment Bill of Materials ================================== Item No Qty Vendor Model/Name Description ----- --- ------ ---------- ----------- 1 10 SuperMicro SuperServer Workstation with 2GB RAM and Linux 6014H-i2 operating system 2 1 Cisco 6509 Cisco Catalyst 6509 Ethernet Switch Load Generators =============== LG Type Name LG1 BOM Item # 1 Processor Name Intel Xeon Processor Speed 3.40 GHz Number of Processors (chips) 2 Number of Cores/Chip 2 Memory Size 2 GB Operating System RHEL4 kernel 2.6.9-55.0.6.ELsmp Network Type 1 x Intel 84546GB PCI-X GigE Load Generator (LG) Configuration ================================= Benchmark Parameters -------------------- Network Attached Storage Type CIFS Number of Load Generators 10 Number of Processes per LG 24 Testbed Configuration --------------------- LG No LG Type Network Target Filesystems Notes ----- ------- ------- ------------------ ----- 1..10 LG1 1 /vol/vol1 /vol/vol2 /vol/vol3 /vol/vol4 N/A Load Generator Configuration Notes ---------------------------------- All filesystems were mounted on all clients, which were connected to the same physical and logical network. Uniform Access Rule Compliance ============================== Each load-generating client hosted 24 processes. The assignment of processes to filesystems and network interfaces was done such that they were evenly divided across all filesystems and network paths to the storage controllers. The filesystem data was striped evenly across all disks and FC-AL loops on the storage backend. Other Notes =========== Other test notes: None. NetApp is a registered trademark and "Data ONTAP", "FlexVol", and "WAFL" are trademarks of NetApp, Inc. in the United States and other countries. All other trademarks belong to their respective owners and should be treated as such. ================================================================================ Generated on Tue May 04 17:58:59 2010 by SPECsfs2008 ASCII Formatter Copyright (C) 1997-2008 Standard Performance Evaluation Corporation