Post on 17-Oct-2020
transcript
Presentation SE Technical Version
AVERE SYSTEMS, INC 5000 McKnight Road, Suite 404 Pittsburgh, PA 15237 (412) 635-7170
Mark Renault
averesystems.com
Company Overview
2 Proprietary & Confidential
• Mission – Provide Demand-Driven Storage™ solutions that dynamically organize data
into the most optimum storage tier yielding higher performance, global name space, LAN & WAN virtualization & lower cost
• Profile – Headquartered in Pittsburgh, PA – Menlo Ventures & Norwest Venture Partners
• Management Team – Ron Bianchini, CEO: NetApp, Spinnaker, FORE, Scalable Networks – Mike Kazar, CTO: NetApp, Spinnaker Networks, IBM, Transarc – John Dean, CFO: Vivisimo, NetApp, Spinnaker, P&G – Tom Hicks, VP Eng: NetApp, Spinnaker, FORE – Rebecca Thompson, Mkg: Vivisimo, FreeMarkets, FORE, Cisco – Brian Gladden, Sales: Gluster, Gear6, NetApp – Dan Nydick, Dir. Eng: NetApp, Spinnaker, FORE, Scalable Networks
Customer Challenge
• Hard disk drives (HDDs) are getting bigger not faster
• Many, costly 15k RPM drives required to achieve performance
• Challenging due to budget, power, cooling, floor space constraints
• Want SSD but solutions are expensive, complex, incomplete, vendor-specific
3 Proprietary & Confidential
Track
Sector
Platters
Linear Density (bpi)
Track Density (tpi)
Areal Density
Storage Media Comparison Summary
Small Large Random
Large Sequential
Archival SATA SATA SATA
Read RAM SSD SAS
Write RAM SAS SAS
Cap Price $/GB R Perf W Perf
SATA HDD 2,000GB $150 $0.08 130 130
SAS HDD 300GB $270 $0.90 400 360
SLC Flash 64GB $700 $11.00 24,500 1,000
DRAM 32GB $1280 $40.00 325,000 325,000
True Dynamic Tiering
• What? Finest level of granularity
• When? Data is tiered on-the-fly
• How? Automatic movement between tiers
– Automatic by frequency, access pattern and size
5
LUN Volume File Block
Weeks Days Hours On-The-Fly
Manual Disruptive
Manual Non-disruptive
Policy-based Automatic
How It Works
6 Proprietary & Confidential
• Tiered File System (TFS) dynamically places data on optimal media
• Active data owned by high-performance Avere FXT cluster • In-active data owned by Mass Storage System (MASS) • Offers a global view of all MASS filesystems locally & remotely • Avere algorithms monitor access patterns & manage data location • Policy mgmt keeps FXT cluster in sync with MASS for backup, etc.
Customer Benefits
• Performance acceleration: Active data moved to RAM, SSD, SAS • Cost savings: 5:1 reduction in disks, power, space • Simplicity: Seamless fit with existing clients, NAS servers • Scaling: Efficient, non-disruptive growth through clustering • Ease of management: Global Name Space & WAN capable
7 Proprietary & Confidential
Performance & capacity scale independently, more efficiently
SPECsfs2008: Industry Benchmark Posting Op Rate Latency #FileSys # Disks
Apple Xserve 8,053 1.37 6 49 Apple Xserve 18,511 2.63 16 65 Apple MacPro Leopard 9,189 2.18 32 65 Apple MacPro Snow Leopard 18,784 2.67 32 65 Avere 22,025 1.30 1 14 Avere 2-node cluster 43,796 1.33 1 26 Avere 6-node cluster 131,591 1.38 1 79 BlueArc Mercury 50 40,137 3.38 1 74 BlueArc Mercury 50 cluster 80,279 3.42 2 148 BlueArc Mercury 100 72,921 3.39 1 146 BlueArc Mercury 100 cluster 146,076 3.34 2 292 Exanet 2-node 29,921 1.96 1 148 Exanet 8-node 119,550 2.07 1 592 HP BL860c 4-node 134,689 2.53 48 584 Huawei Symantec 176,728 1.67 6 960 Isilon 10-node 46,635 1.91 1 120 NTAP 3140 FC 40,109 2.59 2 224 NTAP 3140 FC PAM 40,107 1.68 2 112 NTAP 3140 SATA PAM 40,011 2.75 4 112 NTAP 3160 60,409 2.18 4 224 NTAP 3160 FC PAM2 60,507 1.58 2 56 NTAP 3160 SATA PAM2 60,389 2.18 8 96 NTAP 6080 120,011 1.95 2 324 Onstor Cougar 6720 42,111 1.74 32 224 Onstor Cougar 3510 27,078 1.99 16 112 SGI 10,305 3.86 1 242 8
100% Linear Scaling
9 Proprietary & Confidential
22,025
43,796
131,591
y = 21931x
0
20,000
40,000
60,000
80,000
100,000
120,000
140,000
1 2 6
Avere SPECsfs2008 Throughput (ops/sec)
SPECsfs2008 Performance*
10 Proprietary & Confidential Only solution with 1 file system
*For more details: http://www.spec.org/sfs2008/results/sfs2008.html
7x reduction in disks, power & space on average *Includes disks used for storing data, not system/OS disks
Comparing 100k ops/sec Solutions*
*For more details: http://www.spec.org/sfs2008/results/sfs2008.html
$3.4/op
$11.3/op $76.3/op
Avere NetApp EMC Product FXT 2500, 6-cluster FAS6080, FCAL Disks Celerra NS-G8, Symmetrix V-Max
Performance (ORT)
131,591 ops/sec (1.38 msec)
120,011 ops/sec (1.95 msec)
110,621 ops/sec (2.32 ms ORT)
Usable Capacity 15.3TB SATA 14.0TB FC 12.9TB SSD
List Price $445,000 $1,351,000 $8,435,000
Rack Units 16 84 95
WAN Deployment
Satellite office – Minor installation w/o local MASS
– Multiple satellite offices supported – Write-around mode – Selectable cache timeout period
12 Proprietary & Confidential
Core office – Major datacenter w/ local MASS(es)
– Multiple core offices supported – Data retention & management at core office
• Accelerate data access at Satellite offices
• Coherent access to all data from all offices
• Hide WAN latency at Satellite office
• Centralize data management & retention at Core office
WAN*
*Typical WAN connection is private, reliable network
Avere FXT Cluster!
Avere FXT Cluster!(optional)!
– Optional FXT cluster (in WT mode)
Global Namespace
• Join exports from multiple MASSes into GNS
• Support heterogeneous MASS vendors & models
• Clients access all exports/MASSes from a single mountpoint, single IP
• NFS & CIFS support, simpler than automounter & DFS, no extra server
• Newly added exports are visible to clients without client reboot
• GNS “logical view” is admin-defined on Avere UI, client’s view of namespace
• Nesting of exports/junctions not supported
13 Proprietary & Confidential
Global Namespace
14 Proprietary & Confidential
GNS Logical View /
/eng /sales /finance /support
/hw_eng /sw_eng /pipeline /staffing /fy2009 /cust_data
NetApp1:
EMC:
NetApp2:
Sun_ZFS:
Data Center (MASS:/export) Remote Site (MASS:/export)
/mech_des
/src_code /pipeline
/cust_data
/fy2009
/staffing
Clients
FXT Cluster
/mech_des /src_code
WAN
Migration
• Non-disruptive migration between two MASSes (see below) • Export is the unit of migration • Enables…
– Moving exports to a newly installed MASS – Moving exports from an overloaded MASS – Moving exports from a soon-to-be-decommissioned MASS
• Checkpoints implemented, don’t need to restart if A or B fails • Resources consumed, peak performance not available
15 Proprietary & Confidential
NetApp1
NetApp2
Migration start
Migration complete
Migrating /src_code export from NetApp1 to NetApp2
Export moves from NetApp1 to NetApp2 Physical location changes, logical does not
Physical Logical
Before netapp1:/vol/vol0/src_code /eng/sw_eng/src_code
After netapp2:/vol/vol1/src_code /eng/sw_eng/src_code
User Interface
16 Proprietary & Confidential
Simple Administration – Install first FXT node in minutes – Additional nodes join cluster automatically – Email, web GUI alerts
Powerful GUI Monitoring – Historical monitoring of ops/sec, throughout, and latency – Per cluster, per vserver, and per node stats provided – Hot lists show most active files, client IPs, and CPUs – Support 3rd-party monitoring tools: XML API, RRD data format, SNMP
250k ops/sec Random IO, 50x Acceleration
17 Proprietary & Confidential
Configuration: 6 FXT 2700 nodes, NetApp MASS, 250k client ops/sec, 5k MASS ops/sec, 50x acceleration, seismic SRME application (Surface-Related Multiple Elimination)
2 GByte/sec Throughput, 50x Acceleration
18 Proprietary & Confidential
Configuration: 6 FXT 2700 nodes, NetApp MASS, 2 GB/sec client throughput, 40 MB/sec MASS throughput, 50x acceleration, seismic SRME application (Surface-Related Multiple Elimination)
50x Lower Latency with Avere
19 Proprietary & Confidential
Configuration: 6 FXT 2700 nodes, NetApp MASS, 0.4 msec (avg.) client latency, 20 msec (avg.) MASS latency, 50x acceleration, seismic SRME application (Surface-Related Multiple Elimination)
50x Lower Latency with Avere
20 Proprietary & Confidential
Configuration: 6 FXT 2700 nodes, NetApp MASS, 0.4 msec (avg.) client latency*, 20 msec (avg.) MASS latency, 50x acceleration, seismic SRME application (Surface-Related Multiple Elimination)
*See next slide for zoom-in on client-side latency
50x Lower Latency with Avere
21 Proprietary & Confidential
Configuration: 6 FXT 2700 nodes, NetApp MASS, 0.4 msec (avg.) client latency, 20 msec (avg.) MASS latency, 50x acceleration, seismic SRME application (Surface-Related Multiple Elimination)
Smoothing Out Latency Spikes of Slow MASS
22 Proprietary & Confidential
Configuration: 2 FXT 2300 nodes, Sun Thumper+Solaris+ZFS MASS, client latency < 2 msec, MASS latency > 15 msec
Operating Modes
Write-Through – Ultimate reliability – Writes commit to Avere
nodes & MASS – Writes limited by MASS – Read performance scales
23
Write-Around – Some users mount MASS
directly – Expected during initial
installation – Writes limited by MASS – Reads reduced by status
check – Selectable cache timeout
period
Write-Back – Expected configuration – Read & Write performance
scales – Performance scales
independently of MASS – Write-through scheduling to
sync with backup, etc.
Proprietary & Confidential
Avere FXT Series
24 Proprietary & Confidential
• Hardware – 2U Rack Mount System – 64GB DRAM, 1GB NVRAM – FXT 2700: 512GB SSD/Flash (SLC) – FXT 2500: 3.6TB HDD (15k SAS) – FXT 2300: 1.2TB HDD (15k SAS)
• Performance – Per node results below, performance scales linearly to 25 nodes per cluster
(1) FXT 2700 recommended for this workload, (2) FXT 2500 recommended for this workload
• Protocols – Client: NFSv3 (TCP/UDP), CIFS – MASS: NFSv3 (TCP)
• High Availability – N+1 failover – Persistent non-volatile memory – Redundant network ports & power
• Management – GUI, email alerts, SNMP, XML API, policy-based management
3-node FXT cluster shown
Perf. per FXT node
Random I/0 (ops/sec) Sequential I/O (MB/sec) SPEC (ops/sec) 300GB Working Set
256B read 4KB read 4KB write Read Write SFS'97 SFS'08 Rand. read Seq. read
FXT 2700 103k 96k 16k 1,600 330 49k (2) 28k ops/sec 870 MB/sec
FXT 2500 103k 94k 13k 1,560 330 49k 22k (1) (1)
FXT 2300 103k 94k 13k 1,560 330 49k (2) (1) (1)
Evidence
25 Proprietary & Confidential
• ESG Quote: “Conceptually, an architecture like this could quite literally change everything we thought we knew about storage and I/O. If the Avere architecture can perform as intended, it might just turn decades of thinking on its head,” said Steve Duplessie, Founder of ESG.
• Customer Quote: ''Before we added the Avere FXT Series to our storage network, we were seriously considering replacing some of our slower mass storage systems due to their inability to keep up with client demands,” said Bryan Nielsen, IT Architect at the Salk Institute. “The introduction of the FXT into our network took the load off of these devices, breathing new life into our current storage infrastructure investments. In addition, Avere’s FXT opens up new possibilities in price, performance and size considerations for future storage investments.”
Summary
26 Proprietary & Confidential
• Right Time – Storage industry at start of new era – Transition from HDD to SSD has begun – Heterogeneous Global Name Space
• Right Architecture – Leverage all media types – Tiering is granular, fast, and automatic – Support wide-range of application workloads – Simple to deploy and manage
• Right Team – Unique blend of clustered storage, file system and networking
expertise – Proven track record
Tiered NAS Strategy
SATA FC
Availability +, Access Density +
NVRAM RAM
SSD SAS 50:1
Dynamic Tiered NAS
Local De-Dupe
SATA
200:1
Private Remote
D.C.
Private Distributed
D.C.
Latency +, Reliability -
WAN Remote
D.C.
WAN Distributed
D.C.
500:1
Cloud
27
Type Company Limitation
NAS Server NetApp, EMC, Sun, Isilon, BlueArc
• Over provision & short stroke • Expensive due to disks, power & space • Forced to select expensive drive types
Caching Appliance NetApp FlexCache • Read only work loads (non-persistent) • One protocol (NFS) limitation typical • Limited scaling
SSD Adapter NetApp PAM, Fusion IO
• Inability to scale separately from server • Proprietary (NetApp) • Integration burden placed on end-user (Fusion IO)
SSD Array EMC, Texas Memory Systems
• High media cost • Wasteful, copy entire volume to SSD • Limited Tier-0 management
Switch F5/Acopia • Disruptive, non-transparent • Data migration between tiers is slow • Poor performance for small-file apps
28
Typical Vendor Approaches to Challenge
Thank you!
AVERE SYSTEMS, INC 5000 McKnight Road, Suite 404 Pittsburgh, PA 15237 (412) 635-7170
averesystems.com