Considerations and performance evaluations of shared storage area networks at NASA Goddard Space...

28
1 Shared Storage Networks at NASA GSFC Architectural Considerations and Performance Evaluations Of Shared Storage Area Networks At NASA Goddard Space Flight Center Hoot Thompson, Curt Tilmes, Robert Cavey, Bill Fink, Paul Lang and Ben Kobler [email protected] Patuxent Technology Partners, LLC www.ptpnow.com The Twentieth IEEE/Eleventh NASA Goddard Conference on Mass Storage Systems and Technologies

Transcript of Considerations and performance evaluations of shared storage area networks at NASA Goddard Space...

1Shared Storage Networks at NASA GSFC

Architectural Considerations and Performance Evaluations

Of Shared Storage Area NetworksAt NASA Goddard Space Flight Center

Hoot Thompson, Curt Tilmes, Robert Cavey,Bill Fink, Paul Lang and Ben Kobler

[email protected] Technology Partners, LLC

www.ptpnow.com

The Twentieth IEEE/Eleventh NASA Goddard Conference on Mass Storage Systems and Technologies

2Shared Storage Networks at NASA GSFC

Introduction – Some Trends

• Storage consolidation continues to be a popular architectural construct – Storage Area Networks (SAN)

• Minimizing costs while expanding storage capacity and providing shared access is a driver

• Fibre Channel connected storage remains dominant at least in the high performance space

• Existing IP infrastructures versus ‘still’ expensive Fibre Channel represent a potentially attractive mechanism for connecting users to storage at the block level

3Shared Storage Networks at NASA GSFC

GSFC SAN Pilot Initiative

Shared Data Repository

Technology Research

Operational Testing

User Acceptance

High performanceScalableWide area accessSecure

4Shared Storage Networks at NASA GSFC

By Way of Review –Traditional Infrastructure

5Shared Storage Networks at NASA GSFC

Storage Area Network Infrastructure

6Shared Storage Networks at NASA GSFC

Some Definitions• Fibre Channel

– Industry standard high-speed SCSI transport technology– 1 or 2 Gbit/sec, 10 Gbit/sec coming

• Internet SCSI (iSCSI)– “represents a light switch approach to storage networking”

• Fibre Channel Over IP (FCIP)– “means of encapsulating Fibre Channel frames within TCP/IP

specifically for linking Fibre Channel SANs over wide areas”• Internet Fibre Channel Protocol (iFCP)

– “gateway-to-gateway protocol for providing Fibre Channel fabric services to Fibre Channel end devices over a TCP/IP network”

Notes:

Standards largely the work of Internet Engineering Task Force (IETF) – www.ietf.org

Definitions from “IP SANs – A Guide to iSCSI, iFCP, and FCIP Protocols for

Storage Area Networks” by Tom Clark

7Shared Storage Networks at NASA GSFC

IP-Based SAN Connections

8Shared Storage Networks at NASA GSFC

GSFC Pilot SAN Configuration

RAIDStorage

SGI

Linux

Sun

Linux

Bldg 32

Bldg 33

Bldg 28

iSCSIRouter

LinuxiSCSIRouter

Universityof Maryland

Linux

LinuxBldg 2

Fibre ChannelSwitch

Fibre ChannelSwitch

TapeLibrary

Linux(OMI)

Fibre ChannelSwitch

Linux(OMI)

Fibre ChannelSwitch

GSFCScience and

Engineering Network(SEN)

IP

FC

9Shared Storage Networks at NASA GSFC

GSFC IP Testing

• iSCSI in comparison to native Fibre Channel• Primarily Linux based machines

– Sun platform also ‘looked at’– SGI currently does not support iSCSI– Windows® not the platform of choice

• TCP off-load engine (TOE) card– Evaluated on Windows machine

10Shared Storage Networks at NASA GSFC

Benchmarks

• Large file, large block sequential transfers– lmdd (http://www.bitmover.com/lmbench/lmdd.8.html)

• Small file, transaction oriented tests– bonnie++ (http://www.coker.com.au/bonnie++)– Postmark (http://www.netapp.com/tech_library/3022.html)

• Application testing– Composite generation from MODIS data by U of MD staff

• Different file systems– ext2fs (native Linux file system)– Global File System (GFS, Sistina)– CentraVision™ File System (CVFS) now StorNext File System

(ADIC)Notes:

Benchmark numbers for the most part are ‘out of the box’ results and should be

viewed as representative not definitive.

11Shared Storage Networks at NASA GSFC

Linux Fibre Channel Connected – Writes

48.00

50.00

52.00

54.00

56.00

58.00

60.00

1 2 4 8 16 32

Block Size (MB)

Tran

sfer

Rat

e (M

B/se

c)

ext2 gfs cvfs

12Shared Storage Networks at NASA GSFC

Linux FC Connected – Reads

45.0050.0055.0060.0065.0070.0075.0080.00

1 2 4 8 16 32

Block Size (MB)

Tran

sfer

Rat

e (M

B/se

c)

ext2 gfs cvfs

13Shared Storage Networks at NASA GSFC

Linux iSCSI Connected – Writes

25.00

30.00

35.00

40.00

45.00

1 2 4 8 16 32

Block Size (MB)

Tran

sfer

Rat

e (M

B/se

c)

ext2 gfs cvfs

14Shared Storage Networks at NASA GSFC

Linux iSCSI Connected – Reads

10.00

15.00

20.00

25.00

30.00

1 2 4 8 16 32

Block Size (MB)

Tran

sfer

Rat

e (M

B/s

ec)

ext2 gfs cvfs

15Shared Storage Networks at NASA GSFC

U of MD Linux iSCSI Connected – Writes

12.0014.0016.0018.0020.0022.0024.0026.00

1 2 4 8 16 32

Block Size (MB)

Tran

sfer

Rat

e (M

B/se

c)

ext2fs gfs cvfs

16Shared Storage Networks at NASA GSFC

U of MD Linux iSCSI Connected – Reads

10.00

11.00

12.00

13.00

14.00

15.00

16.00

1 2 4 8 16 32

Block Size (MB)

Tran

sfer

Rat

e (M

B/se

c)

ext2fs gfs cvfs

17Shared Storage Networks at NASA GSFC

iSCSI TOE Performance – Writes

0.005.00

10.0015.0020.0025.0030.0035.0040.0045.0050.00

1 2 4 8 16 32

Block Size (MB)

Tran

sfer

Rat

e (M

B/se

c)

0.00

5.00

10.00

15.00

20.00

25.00

30.00

CPU

NIC (MB/sec)TOE (MB/sec)NIC (%)TOE (%)

18Shared Storage Networks at NASA GSFC

iSCSI TOE Performance – Reads

0.005.00

10.0015.0020.0025.0030.0035.0040.0045.00

1 2 4 8 16 32

Block Size (MB)

Tran

sfer

Rat

e

0.00

5.00

10.00

15.00

20.00

25.00

30.00

CPU

NIC (MB/sec)

TOE (MB/sec)

NIC (% )

TOE (% )

19Shared Storage Networks at NASA GSFC

Bonnie++ and Postmark Results

• Compared various combinations of clients and file systems for small file, random operations

• Benchmarks highlighted file system design differences more so than iSCSI characteristics– ext2fs inodes on disk– GFS central lock manager with locally cached inodes– CVFS centralized metadata function

• Ext2fs performed the best then GFS then CVFS• Performance best for FC client, U of MD worst• Complete set of numbers available in paper

20Shared Storage Networks at NASA GSFC

Application Testing – U of MD

• Purpose:– Demonstrate operational value of shared data repository– Generate MODIS composite data at U of MD using

storage located at GSFC ~ 6 miles away• Test results:

– Local ext2fs dataset • > 1hr 45 min• Excluding ftp overhead time of 45 min

– GSFC resident dataset• gfs > 2 hr 8 min• cvfs > 3hr 15 min

21Shared Storage Networks at NASA GSFC

Ozone Monitoring Instrument System

• Purpose:– Study datasets from the Total Ozone Mapping

Spectrometer (TOMS) instruments and Solar Backscatter Ultraviolet (SBUV) instruments

– Prepare for Ozone Monitoring Instrument (OMI)• Architecture:

– Multi-building FC and NFS connected clients– CVFS SAN file system– DataDirect storage– Brocade switches

• Opinions to date:– Shared SAN storage performance comparable to local– NFS performance marginal but expected to improve

22Shared Storage Networks at NASA GSFC

OMI Architecture

clientworksation

clientworkstation

Bldg 32

Bldg 33

omifs3

omi916

RAIDStorage

Fibre ChannelSwitch

Fibre ChannelSwitch

Fibre ChannelSwitch

SEN

IP

FC

23Shared Storage Networks at NASA GSFC

Alaska-to-GSFC SAN File System

• Objective:– Leverage IP technology for moving data between

Alaska and GSFC • Trades:

– Standard FTP – SAN file system involving Alaska and GSFC

• Results to date – inconclusive, a work in progress:– First set-up used product designed for E-port expansion

– Not well suited for extending block device over the required distance

– Bandwidth was unacceptable– Tests used delay simulator and loopback– Additional tests planned using different equipment

24Shared Storage Networks at NASA GSFC

Alaska SAN Architecture

25Shared Storage Networks at NASA GSFC

Conclusions

• IP is more than just a viable SAN technology– Painless to implement and test– Poised to have a dramatic impact– Market has yet to play out the options - iSCSI, FCIP and/or iFCP– Vendor commitment still forming– Standards:

• iSCSI passed• FCIP and iFCP are in work

• Gain in flexibility offsets bandwidth loss for potentially a large category of users– Easy to envision a SAN constructed completely of iSCSI connected

clients

26Shared Storage Networks at NASA GSFC

Simple iSCSI SAN

27Shared Storage Networks at NASA GSFC

Plans For Additional Testing

• iSCSI– Jumbo frames– Distance limit testing

• 2Gbit Fibre Channel and Fibre Channel trunking• Geographically distributed SAN file systems• FCIP and iFCP equipment • SAN management software• Security products• Etc.

Build a happy operational base!

28Shared Storage Networks at NASA GSFC

Acknowledgements

• J Patrick Gary, NASA GSFC• George Uhl, NASA GSFC• Tino Sciuto, NASA GSFC• Charlene DiMiceli, University of Maryland• Fritz McCall, University of Maryland• Mike Smorul, University of Maryland