Considerations and performance evaluations of shared storage area networks at NASA Goddard Space...
-
Upload
independent -
Category
Documents
-
view
2 -
download
0
Transcript of Considerations and performance evaluations of shared storage area networks at NASA Goddard Space...
1Shared Storage Networks at NASA GSFC
Architectural Considerations and Performance Evaluations
Of Shared Storage Area NetworksAt NASA Goddard Space Flight Center
Hoot Thompson, Curt Tilmes, Robert Cavey,Bill Fink, Paul Lang and Ben Kobler
[email protected] Technology Partners, LLC
www.ptpnow.com
The Twentieth IEEE/Eleventh NASA Goddard Conference on Mass Storage Systems and Technologies
2Shared Storage Networks at NASA GSFC
Introduction – Some Trends
• Storage consolidation continues to be a popular architectural construct – Storage Area Networks (SAN)
• Minimizing costs while expanding storage capacity and providing shared access is a driver
• Fibre Channel connected storage remains dominant at least in the high performance space
• Existing IP infrastructures versus ‘still’ expensive Fibre Channel represent a potentially attractive mechanism for connecting users to storage at the block level
3Shared Storage Networks at NASA GSFC
GSFC SAN Pilot Initiative
Shared Data Repository
Technology Research
Operational Testing
User Acceptance
High performanceScalableWide area accessSecure
6Shared Storage Networks at NASA GSFC
Some Definitions• Fibre Channel
– Industry standard high-speed SCSI transport technology– 1 or 2 Gbit/sec, 10 Gbit/sec coming
• Internet SCSI (iSCSI)– “represents a light switch approach to storage networking”
• Fibre Channel Over IP (FCIP)– “means of encapsulating Fibre Channel frames within TCP/IP
specifically for linking Fibre Channel SANs over wide areas”• Internet Fibre Channel Protocol (iFCP)
– “gateway-to-gateway protocol for providing Fibre Channel fabric services to Fibre Channel end devices over a TCP/IP network”
Notes:
Standards largely the work of Internet Engineering Task Force (IETF) – www.ietf.org
Definitions from “IP SANs – A Guide to iSCSI, iFCP, and FCIP Protocols for
Storage Area Networks” by Tom Clark
8Shared Storage Networks at NASA GSFC
GSFC Pilot SAN Configuration
RAIDStorage
SGI
Linux
Sun
Linux
Bldg 32
Bldg 33
Bldg 28
iSCSIRouter
LinuxiSCSIRouter
Universityof Maryland
Linux
LinuxBldg 2
Fibre ChannelSwitch
Fibre ChannelSwitch
TapeLibrary
Linux(OMI)
Fibre ChannelSwitch
Linux(OMI)
Fibre ChannelSwitch
GSFCScience and
Engineering Network(SEN)
IP
FC
9Shared Storage Networks at NASA GSFC
GSFC IP Testing
• iSCSI in comparison to native Fibre Channel• Primarily Linux based machines
– Sun platform also ‘looked at’– SGI currently does not support iSCSI– Windows® not the platform of choice
• TCP off-load engine (TOE) card– Evaluated on Windows machine
10Shared Storage Networks at NASA GSFC
Benchmarks
• Large file, large block sequential transfers– lmdd (http://www.bitmover.com/lmbench/lmdd.8.html)
• Small file, transaction oriented tests– bonnie++ (http://www.coker.com.au/bonnie++)– Postmark (http://www.netapp.com/tech_library/3022.html)
• Application testing– Composite generation from MODIS data by U of MD staff
• Different file systems– ext2fs (native Linux file system)– Global File System (GFS, Sistina)– CentraVision™ File System (CVFS) now StorNext File System
(ADIC)Notes:
Benchmark numbers for the most part are ‘out of the box’ results and should be
viewed as representative not definitive.
11Shared Storage Networks at NASA GSFC
Linux Fibre Channel Connected – Writes
48.00
50.00
52.00
54.00
56.00
58.00
60.00
1 2 4 8 16 32
Block Size (MB)
Tran
sfer
Rat
e (M
B/se
c)
ext2 gfs cvfs
12Shared Storage Networks at NASA GSFC
Linux FC Connected – Reads
45.0050.0055.0060.0065.0070.0075.0080.00
1 2 4 8 16 32
Block Size (MB)
Tran
sfer
Rat
e (M
B/se
c)
ext2 gfs cvfs
13Shared Storage Networks at NASA GSFC
Linux iSCSI Connected – Writes
25.00
30.00
35.00
40.00
45.00
1 2 4 8 16 32
Block Size (MB)
Tran
sfer
Rat
e (M
B/se
c)
ext2 gfs cvfs
14Shared Storage Networks at NASA GSFC
Linux iSCSI Connected – Reads
10.00
15.00
20.00
25.00
30.00
1 2 4 8 16 32
Block Size (MB)
Tran
sfer
Rat
e (M
B/s
ec)
ext2 gfs cvfs
15Shared Storage Networks at NASA GSFC
U of MD Linux iSCSI Connected – Writes
12.0014.0016.0018.0020.0022.0024.0026.00
1 2 4 8 16 32
Block Size (MB)
Tran
sfer
Rat
e (M
B/se
c)
ext2fs gfs cvfs
16Shared Storage Networks at NASA GSFC
U of MD Linux iSCSI Connected – Reads
10.00
11.00
12.00
13.00
14.00
15.00
16.00
1 2 4 8 16 32
Block Size (MB)
Tran
sfer
Rat
e (M
B/se
c)
ext2fs gfs cvfs
17Shared Storage Networks at NASA GSFC
iSCSI TOE Performance – Writes
0.005.00
10.0015.0020.0025.0030.0035.0040.0045.0050.00
1 2 4 8 16 32
Block Size (MB)
Tran
sfer
Rat
e (M
B/se
c)
0.00
5.00
10.00
15.00
20.00
25.00
30.00
CPU
NIC (MB/sec)TOE (MB/sec)NIC (%)TOE (%)
18Shared Storage Networks at NASA GSFC
iSCSI TOE Performance – Reads
0.005.00
10.0015.0020.0025.0030.0035.0040.0045.00
1 2 4 8 16 32
Block Size (MB)
Tran
sfer
Rat
e
0.00
5.00
10.00
15.00
20.00
25.00
30.00
CPU
NIC (MB/sec)
TOE (MB/sec)
NIC (% )
TOE (% )
19Shared Storage Networks at NASA GSFC
Bonnie++ and Postmark Results
• Compared various combinations of clients and file systems for small file, random operations
• Benchmarks highlighted file system design differences more so than iSCSI characteristics– ext2fs inodes on disk– GFS central lock manager with locally cached inodes– CVFS centralized metadata function
• Ext2fs performed the best then GFS then CVFS• Performance best for FC client, U of MD worst• Complete set of numbers available in paper
20Shared Storage Networks at NASA GSFC
Application Testing – U of MD
• Purpose:– Demonstrate operational value of shared data repository– Generate MODIS composite data at U of MD using
storage located at GSFC ~ 6 miles away• Test results:
– Local ext2fs dataset • > 1hr 45 min• Excluding ftp overhead time of 45 min
– GSFC resident dataset• gfs > 2 hr 8 min• cvfs > 3hr 15 min
21Shared Storage Networks at NASA GSFC
Ozone Monitoring Instrument System
• Purpose:– Study datasets from the Total Ozone Mapping
Spectrometer (TOMS) instruments and Solar Backscatter Ultraviolet (SBUV) instruments
– Prepare for Ozone Monitoring Instrument (OMI)• Architecture:
– Multi-building FC and NFS connected clients– CVFS SAN file system– DataDirect storage– Brocade switches
• Opinions to date:– Shared SAN storage performance comparable to local– NFS performance marginal but expected to improve
22Shared Storage Networks at NASA GSFC
OMI Architecture
clientworksation
clientworkstation
Bldg 32
Bldg 33
omifs3
omi916
RAIDStorage
Fibre ChannelSwitch
Fibre ChannelSwitch
Fibre ChannelSwitch
SEN
IP
FC
23Shared Storage Networks at NASA GSFC
Alaska-to-GSFC SAN File System
• Objective:– Leverage IP technology for moving data between
Alaska and GSFC • Trades:
– Standard FTP – SAN file system involving Alaska and GSFC
• Results to date – inconclusive, a work in progress:– First set-up used product designed for E-port expansion
– Not well suited for extending block device over the required distance
– Bandwidth was unacceptable– Tests used delay simulator and loopback– Additional tests planned using different equipment
25Shared Storage Networks at NASA GSFC
Conclusions
• IP is more than just a viable SAN technology– Painless to implement and test– Poised to have a dramatic impact– Market has yet to play out the options - iSCSI, FCIP and/or iFCP– Vendor commitment still forming– Standards:
• iSCSI passed• FCIP and iFCP are in work
• Gain in flexibility offsets bandwidth loss for potentially a large category of users– Easy to envision a SAN constructed completely of iSCSI connected
clients
27Shared Storage Networks at NASA GSFC
Plans For Additional Testing
• iSCSI– Jumbo frames– Distance limit testing
• 2Gbit Fibre Channel and Fibre Channel trunking• Geographically distributed SAN file systems• FCIP and iFCP equipment • SAN management software• Security products• Etc.
Build a happy operational base!