SPECsfs2008_nfs.v3 Result ================================================================================ Oracle Corporation : Sun ZFS Storage 7420 SPECsfs2008_nfs.v3 = 267928 Ops/Sec (Overall Response Time = 1.31 msec) ================================================================================ Performance =========== Throughput Response (ops/sec) (msec) --------------------- ---------------- 12601 1.0 25221 0.8 37843 0.6 50466 0.7 63110 0.7 75801 0.7 88384 0.9 101131 1.0 113768 1.1 126414 1.1 139195 1.2 151962 1.3 164821 1.4 177424 1.5 189957 1.6 202943 1.7 215327 1.9 228602 1.9 241096 2.2 255978 2.6 267928 3.1 ================================================================================ Product and Test Information ============================ Tested By Oracle Corporation Product Name Sun ZFS Storage 7420 Hardware Available March 2012 Software Available Dec 2011 Date Tested March 2012 SFS License Number 6 Licensee Locations Broomfield, CO USA The Sun ZFS Storage 7420 is a high-performance storage system that offers enterprise-class NAS capabilities with industry-leading Oracle integration and storage efficiency, in a cost-effective high-availability configuration. The Sun ZFS Storage 7420's simplified set up and management combine with industry-leading storage analytics and a performance-optimized platform to make it the ideal solution for enterprise customers storage requirements. The Sun ZFS Storage 7420 can scale to 2 TB Memory, 80 CPU cores, and 1.7 PB capacity, with up to 4 TB of Flash Cache in a high-availability configuration. Configuration Bill of Materials =============================== Ite m Vendo No Qty Type r Model/Name Description --- --- ---- ----- ---------- ----------- 1 2 Storage Oracl 7420 Sun ZFS Storage 7420 Controller e Controller 2 4 10 Gigabit Oracl Sun PCI-E Dual 10GbE Dual port 10Gb Ethernet Ethernet e Fiber adapter Adapter 3 8 Short Wave Oracl 10Gbps Short Wave Short Wave Pluggable Pluggable e Pluggable Transceiver Transceiver Transceiver (SFP+) 4 2 Shelf w/Disk Oracl Sun Disk Shelf SAS-2 SAS-2 Disk Shelf with Drives e with 20 Disk Drives 20x300GB 15K,HDD 5 10 Shelf w/Disk Oracl Sun Disk Shelf SAS-2 SAS-2 Disk Shelf with Drives e with 24 Disk Drives 24x300GB 15K,HDD 6 8 SSD Drive Oracl 512GB Solid State Drive SSD Read Flash e SATA-2 Accelerator 512GB 7 8 SSD Drive Oracl SAS-2 73GB 3.5-inch SSD SSD Write Flash e Write Flash Accelerator Accelerator 73GB 8 4 SAS-2 Host Bus Oracl SAS-2 8-Port 6Gbps HBA SAS-2 8-Port 6Gbps Host Adapter e Bus Adapter Server Software =============== OS Name and Version 2011.1.1 Other Software None Filesystem Software ZFS Server Tuning ============= Name Value Description ---- ----- ----------- atime off Disable atime updates (all file systems) Server Tuning Notes ------------------- None Disks and Filesystems ===================== Number Description of Disks Usable Size ----------- -------- ----------- 300GB SAS 15K RPM Disk Drives 280 37.1 TB 500GB SATA 7.2K RPM Disk Drives Each controller contains 2 4 899.0 GB of these disk drives and are mirrored and are not used for cache data or data storage. Total 284 37.9 TB Number of Filesystems 32 Total Exported Capacity 36.32 TB Filesystem Type ZFS Filesystem Creation Options default Filesystem Config 32 ZFS file systems Fileset Size 31008.5 GB The storage configuration consists of 12 shelves, 10 with 24 disk drives, 2 with 20 disk drives and 4 write flash devices. Each controller head has 4 read flash accelerators. Each controller is configured to use 140 disk drives, 4 write flash accelerator devices and 4 read flash accelerator devices. The controller is then set up with pools by dividing the disk drives, write flash accelerators and read flash accelerators into 4 pools. Each of the controller's pools is configured with 34 disk drives, 1 write flash accelerator, 1 read flash accelerator and 1 spare disk drive. The pools are then set up to mirror the data (RAID1) across all 34 drives. The write flash accelerator in each pool is used for the ZFS Intent Log (ZIL) and the read flash accelerator is used as a level 2 cache (L2ARC) for the pool. All pools are configured with 4 ZFS file systems each. Since each controller is configured with 4 pools and each pool contains 4 ZFS file system, in total each of 2 controllers has 16 ZFS file systems. The SUT has 32 ZFS file systems for the benchmark. There are 2 internal mirrored system disk drives per controller and are used only for the controllers core operating system. These drives are not used for data cache or storing user data. Network Configuration ===================== Number of Item Ports No Network Type Used Notes ---- ------------ --------- ----- 1 10 Gigabit 4 Each controller has 2 Dual port 10 Gigabit Ethernet Ethernet cards using only a single port on each and using Jumbo Frames Network Configuration Notes --------------------------- There are 8 ports total but 4 are active at a time for high availibility. The MTU size is set to 9000 on each of the 10 Gigabit ports. Benchmark Network ================= Each controller was configured to use a single port of two dual port 10GbE network adapters for the benchmark network. All of the 10GbE network ports on all the load generators systems were connected to the Arista 7124SX switch that provided connectivity. Processing Elements =================== Item No Qty Type Description Processing Function ----- --- ---- ----------- ------------------- 1 8 CPU 2.4GHz 10-core Intel Xeon(tm) NFS, ZFS, TCP/IP, RAID/Storage Processor E7-4870 Drivers Processing Element Notes ------------------------ Each Sun ZFS Storage 7420 controller contains 4 physical processors, each with 10 processing cores. Memory ====== Size in Number of Nonvolati Description GB Instances Total GB le ----------- --------- --------- -------- --------- 7420 controller memory 1024 2 2048 V SSD 73GB SAS-2 Write Flash Accelerator 73 8 584 NV SSD 512GB SATA-2 Read Flash Accelerator 512 8 4096 NV Grand Total Memory Gigabytes 6728 Memory Notes ------------ The Sun ZFS Storage 7420 controllers' main memory is used for the Adaptive Replacement Cache (ARC) the data cache and operating system memory. A separate device, a write flash accelerator is used for the ZFS Intent Log (ZIL). The read flash accelerator is dedicated storage used also to hold cached data and is called the Level 2 Adaptive Replacement Cache (L2ARC). Stable Storage ============== The Stable Storage requirement is guaranteed by the ZFS Intent Log (ZIL) which logs writes and other file system changing transactions to either a write flash accelerator or a disk drive. Writes and other file system changing transactions are not acknowledged until the data is written to stable storage. Since this is an active-active cluster high availability system, in the event of a controller failing or power loss, the other active controller can take over for the failed controller. Since the write flash accelerators or disk drives are located in the disk shelves and can be accessed via the 4 backend SAS channels from both controllers, the remaining active controller can complete any outstanding transactions using the ZIL. In the event of power loss to both controllers, the ZIL is used after power is restored to reinstate any writes and other file system changes. System Under Test Configuration Notes ===================================== The system under test is a Sun ZFS Storage 7420 cluster set up in an active-active configuration. Other System Notes ================== None Test Environment Bill of Materials ================================== Item No Qty Vendor Model/Name Description ----- --- ------ ---------- ----------- 1 4 Oracle Sun Fire X4270 Sun Fire X4270 M2 with 144GB RAM and Oracle M2 Solaris 10 8/11 2 1 Arista 7124SX Arista 7124SX 24-port 10Gb Ethernet switch Load Generators =============== LG Type Name LG1 BOM Item # 1 Processor Name Intel Xeon(tm) X5680 Processor Speed 3.3 GHz Number of Processors (chips) 2 Number of Cores/Chip 6 Memory Size 144 GB Operating System Oracle Solaris 10 8/11 Network Type 10Gb Ethernet Load Generator (LG) Configuration ================================= Benchmark Parameters -------------------- Network Attached Storage Type NFS V3 Number of Load Generators 4 Number of Processes per LG 256 Biod Max Read Setting 2 Biod Max Write Setting 2 Block Size AUTO Testbed Configuration --------------------- LG No LG Type Network Target Filesystems Notes ----- ------- ------- ------------------ ----- 1..4 LG1 1 /export/sfs-1../export/sfs-32 None Load Generator Configuration Notes ---------------------------------- File systems were mounted on all clients and all were connected to the same physical and logical network. Uniform Access Rule Compliance ============================== Every client used 256 processes. All 32 file systems are mounted and accessed by each client and evenly divided amongst all network paths to the 7420 controllers. The file systems data was evenly distributed on the backend. Other Notes =========== Other Test Notes: None ================================================================================ Generated on Wed Apr 18 10:40:36 2012 by SPECsfs2008 ASCII Formatter Copyright (C) 1997-2008 Standard Performance Evaluation Corporation