Avere Systems, Inc. | : | FXT 2500 (6 Node Cluster) |
SPECsfs2008_nfs.v3 | = | 131591 Ops/Sec (Overall Response Time = 1.38 msec) |
|
Tested By | Avere Systems, Inc. |
---|---|
Product Name | FXT 2500 (6 Node Cluster) |
Hardware Available | October 2009 |
Software Available | October 2009 |
Date Tested | September 2009 |
SFS License Number | 9020 |
Licensee Locations |
Pittsburgh, PA USA |
The Avere Systems FXT 2500 appliance provides tiered NAS storage that allows performance to scale independently of capacity. The FXT 2500 is built on a 64-bit architecture that is managed by Avere OS software. FXT clusters scale to as many as 25 nodes and support millions of IO operations per second and deliver multiple GB/s of bandwidth. The Avere OS software dynamically organizes data into tiers. Active data is placed on the FXT appliances while inactive data is placed on slower mass storage servers. The system accelerates performance of read, write and metadata operations. The FXT 2500 supports both NFSv3 and CIFS protocols. The tested configuration consisted of (6) FXT 2500 cluster nodes. The system also included a Linux server that acted as the mass storage system.
Item No | Qty | Type | Vendor | Model/Name | Description |
---|---|---|---|---|---|
1 | 6 | Storage Appliance | Avere Systems, Inc. | FXT 2500 | Avere Systems Tiered NAS Appliance running Avere OS 1.0 software. Includes (8) 450 GB SAS Disks. |
2 | 1 | Server Motherboard | Supermicro | X7DWN+ | Motherboard for Linux mass storage NFS Server. |
3 | 1 | Server Chassis | Supermicro | CSE-846E1-R900B | 4U 24-Drive Chassis for Linux mass storage NFS Server. |
4 | 24 | Disk Drive | Seagate | ST31000340NS | 1TB Barracuda ES.2 SATA 7200 RPM Disks for Linux mass storage NFS Server. |
5 | 1 | System Disk | Seagate | ST3500320NS | 500GB Barracuda ES.2 SATA 7200 RPM Disk for Linux mass storage NFS Server. |
6 | 1 | RAID Controller | LSI | SAS8888ELP | SAS/SATA RAID Controller with 512 MB Cache and External Battery Backup Unit for Linux mass storage NFS Server. |
7 | 1 | NVRAM Adapter | Curtiss-Wright | MM-5453CN | Micro Memory PCIe 1GB battery-backed NVRAM adapter for Linux mass storage NFS Server. |
OS Name and Version | Avere OS 1.0 |
---|---|
Other Software | Mass storage server runs CentOS 5.3 with Linux 2.6.30 kernel |
Filesystem Software | Avere OS 1.0 |
Name | Value | Description |
---|---|---|
Slider Age | 8 hours | Files may be modified up to 8 hours before being written back to the mass storage system. |
net.core.rmem_max | 1048576 | Maximum network receive socket buffer size (in bytes) on Linux mass storage system. |
net.core.rmem_default | 1048576 | Default network receive socket buffer size (in bytes) on Linux mass storage system. |
net.core.wmem_max | 1048576 | Maximum network send socket buffer size (in bytes) on Linux mass storage system. |
net.core.wmem_default | 1048576 | Default network send socket buffer size (in bytes) on Linux mass storage system. |
vm.min_free_kbytes | 40960 | Force Linux VM system on to keep 40MB free on Linux mass storage system. |
vm.dirty_background_ratio | 9 | Percentage of dirty pages at which the Linux VM system starts writing out dirty pages on Linux mass storage system. |
vm.dirty_expire_centisecs | 99 | Age at which (in centiseconds) a buffer will be written out on the Linux mass storage server. |
vm.dirty_writeback_centisecs | 100 | Interval (in centiseconds) that Linux flush daemons wake and write out dirty data on the Linux mass storage server. |
vm.dirty_ratio | 10 | Point (expressed in percentage of total memory) at which a process writing data will start writing out dirty data on Linux mass storage server. |
fs.xfs.xfssyncd_centisecs | 200 | XFS tunable. The interval at which the xfssyncd thread flushes metadata out to disk. This thread will flush log activity out, and do some processing on unlinked inodes. This setting is for the Linux mass storage system. |
fs.xfs.xfsbufd_centisecs | 100 | XFS tunable. The interval at which xfsbufd scans the dirty metadata buffers list. This setting is for the Linux mass storage system. |
fs.xfs.age_buffer_centisecs | 200 | XFS tunable. The age at which xfsbufd flushes dirty metadata buffers to disk. This setting is for the Linux mass storage system. |
/sys/block/sdc/queue/scheduler | noop | Turn off elevator scheduler for RAID array on Linux mass storage server. Hardware RAID controller provides sufficient scheduling. |
/sys/block/sdc/queue/read_ahead_kb | 0 | Turn off block read-ahead for RAID array on Linux mass storage server. Hardware RAID controller provides sufficient read-ahead. |
23+1 RAID5 array was created on the Linux mass storage server using the following LSI MegaCli command line 'MegaCli64 -CfgLdAdd -r5 '[10:0,10:1,10:2,10:3,10:4,10:5,10:6,10:7,10:8,10:9,10:10,10:11,10:12,10:13,10:14,10:15,10:16,10:17,10:18,10:19,10:20,10:21,10:22,10:23]' WB ADRA Cached NoCachedBadBBU -strpsz32 -a0' and disk caches were disabled as follows 'MegaCli64 -LDSetProp -DisDskCache -L0 -a0'. This MegaCli command line creates a RAID5 23+1 array with write-back caching enabled in the controller (WB), adaptive read-ahead (ADRA), controller caching (Cached), caching disabled if the battery-backup unit fails (NoCachedBadBBU), a 32KB stripe unit size on controller zero (-a0). Complete documentation for the MegaCli command line tool is available at http://www.lsi.com/DistributionSystem/AssetDocument/80-00156-01_RevH_SAS_SW_UG.pdf.
Description | Number of Disks | Usable Size |
---|---|---|
Each FXT 2500 node contains (8) 450 GB 15K RPM SAS disks. All FXT data resides on these disks. | 48 | 18.5 TB |
Each FXT 2500 node contains (1) 250 GB SATA disk. System disk. | 6 | 1.5 TB |
The mass storage system contains (24) 1 TB SATA disks. The XFS file system is used to manage these disks and the FXT nodes access them via NFSv3. | 24 | 20.9 TB |
The mass storage system contains (1) 500 GB SATA disk. System disk. | 1 | 500.0 GB |
Total | 79 | 41.4 TB |
Number of Filesystems | 1 |
---|---|
Total Exported Capacity | 21399 GB (Linux mass storage system capacity) |
Filesystem Type | TFS (Tiered File System) |
Filesystem Creation Options | Default on FXT nodes. Linux mass storage server XFS filesystem created with 'mkfs.xfs -d agsize=44876890112,sunit=64,swidth=1472,unwritten=0 -l logdev=/dev/umema,size=128m,version=2,sunit=8,lazy-count=1 -i align=1 -f /dev/sdc'. Mounted using 'mount -t xfs -o logbufs=8,logbsize=262144,allocsize=4096,nobarrier,logdev=/dev/umema /dev/sdc /export' |
Filesystem Config | 23+1 RAID5 configuration on Linux mass storage server. |
Fileset Size | 15320.6 GB |
Item No | Network Type | Number of Ports Used | Notes |
---|---|---|---|
1 | 10 Gigabit Ethernet | 6 | One 10 Gigabit Ethernet port used for each FXT 2500 appliance. |
2 | Gigabit Ethernet | 1 | The mass storage system is connected via a 1 Gigabit Ethernet. |
Each FXT 2500 was attached via a single 10 GbE port to an Arista DCS-7124S 24 port 10 GbE switch. The load generating clients were attached to (2) HP ProCurve 2900-48G switch via a single 1 GbE link. Each HP switch had a 10 GbE uplink to the Arista switch. The mass storage server was attached to the HP switch via a single 1 GbE link. A 1500 byte MTU was used throughout the network.
An MTU size of 1500 was set for all connections to the switch. Each load generator was connected to the network via a single 1 GbE port. The SUT was configured with 12 separate IP addresses on one subnet. Each cluster node was connected via a 10 GbE NIC and was sponsoring 2 IP addresses.
Item No | Qty | Type | Description | Processing Function |
---|---|---|---|---|
1 | 12 | CPU | Intel Xeon E5420 2.50 GHz Quad-Core Processor | FXT 2500 Avere OS, Network, NFS/CIFS, Filesystem, Device Drivers |
2 | 2 | CPU | Intel Xeon E5420 2.50 GHz Quad-Core Processor | Linux mass storage system |
Each file server has two physical processors.
Description | Size in GB | Number of Instances | Total GB | Nonvolatile |
---|---|---|---|---|
FXT System Memory | 64 | 6 | 384 | V |
Mass storage system memory | 32 | 1 | 32 | V |
FXT NVRAM module | 1 | 6 | 6 | NV |
Mass storage system RAID Controller NVRAM Module | .5 | 1 | 0.5 | NV |
Mass storage system NVRAM Module for XFS log | 1 | 1 | 1 | NV |
Grand Total Memory Gigabytes | 423.5 |
Each FXT node has main memory that is used for the operating system and for caching filesystem data. A separate, battery-backed NVRAM module is used to provide stable storage for writes that have not yet been written to disk.
The Avere filesystem logs writes and metadata updates to the NVRAM module. Filesystem modifying NFS operations are not acknowledged until the data has been safely stored in NVRAM. The battery backing the NVRAM ensures that any uncommitted transactions persist for at least 72 hours.
The system under test consisted of (6) Avere FXT 2500 nodes. Each node was attached to the network via 10 Gigabit Ethernet. Each FXT 2500 node contains (8) 450 GB SAS disks. The Linux mass storage system was attached to the network via a single 1 Gigabit Ethernet link. The mass storage server was a 4U Supermicro server configured with an (24) 1TB disk SATA RAID5 array.
N/A
Item No | Qty | Vendor | Model/Name | Description |
---|---|---|---|---|
1 | 13 | Supermicro | X7DWU | Supermicro Server with 8GB of RAM running CentOS 5.3 (Linux 2.6.18-128.el5) |
2 | 5 | Supermicro | X7DWU | Supermicro Server with 8GB of RAM running CentOS 5.3 (Linux 2.6.18-92.el5) |
3 | 1 | Arista Networks | DCS-7124S | Arista 24 Port 10 GbE Switch |
4 | 2 | HP | 2900-48G | HP ProCurve 48 Port 1 GbE Switch |
LG Type Name | LG1 |
---|---|
BOM Item # | 1 |
Processor Name | Intel Xeon E5420 2.50 GHz Quad-Core Processor |
Processor Speed | 2.50 GHz |
Number of Processors (chips) | 2 |
Number of Cores/Chip | 4 |
Memory Size | 8 GB |
Operating System | CentOS 5.3 (Linux 2.6.18-128.el5) |
Network Type | 1 x Intel Gigabit Ethernet Controller 82575EB |
LG Type Name | LG2 |
---|---|
BOM Item # | 2 |
Processor Name | Intel Xeon E5420 2.50 GHz Quad-Core Processor |
Processor Speed | 2.50 GHz |
Number of Processors (chips) | 2 |
Number of Cores/Chip | 4 |
Memory Size | 8 GB |
Operating System | CentOS 5.3 (Linux 2.6.18-92.el5) |
Network Type | 1 x Intel Gigabit Ethernet Controller 82575EB |
Network Attached Storage Type | NFS V3 |
---|---|
Number of Load Generators | 18 |
Number of Processes per LG | 48 |
Biod Max Read Setting | 2 |
Biod Max Write Setting | 2 |
Block Size | 0 |
LG No | LG Type | Network | Target Filesystems | Notes |
---|---|---|---|---|
1..7 | LG1 | 1 | /export/3 | LG1 nodes attached to the first 1 GbE switch |
8..9 | LG2 | 1 | /export/3 | LG2 nodes attached to the first 1 GbE switch |
10..15 | LG1 | 1 | /export/3 | LG1 nodes attached to the second 1 GbE switch |
16..18 | LG2 | 1 | /export/3 | LG2 nodes attached to the second 1 GbE switch |
All clients were connected to a single filesystem through both FXT nodes
Each load-generating client hosted 48 processes. The assignment of processes to network interfaces was done such that they were evenly divided across all network paths to the FXT appliances. The filesystem data was evenly distributed across all disks and FXT appliances.
N/A
Generated on Fri Oct 09 08:48:59 2009 by SPECsfs2008 HTML Formatter
Copyright © 1997-2008 Standard Performance Evaluation Corporation
First published at SPEC.org on 09-Oct-2009