Sunteți pe pagina 1din 11

White Paper

Backing Up and Restoring Multi-Terabyte Data Sets

SGI, LEGATO, and Others Break the 10TB-per-Hour Barrier


1.0 Taking Backup and Restore Performance To a New Level . . . . . . . . . . . . . . . 2
2.0 Measured Backup and Restore Performance . . . . . . . . . . . . . . . . . . . . . . . . . 3
2.1 A Real-World Data Set . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
2.2 File-By-File Backup and Restore (10TB) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
2.3 File-By-File Backup and Restore (1TB) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
2.4 Backup and Restore of a Single Large Filesystem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
2.5 Image Backup with Snapshots . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
2.6 Backup Using xfsdump (10TB) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6

3.0 A Hardware Architecture Optimized for Parallel I/O . . . . . . . . . . . . . . . . . . . 7


4.0 Advanced Software for Optimal Backup and Restore . . . . . . . . . . . . . . . . . . . 8
5.0 Beyond Benchmarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9

1
Executive Summary readily available from SGI or the individual
Backup and restore performance have become vendors. SGI® Professional Services can assist
increasingly important as data set size in high- in designing backup solutions that can meet or
performance computing (HPC) environments exceed the results reported in this paper and
has grown. Until now, published benchmarks that are tailored to unique requirements. Com-
of backup and restore speeds have focused bining these components with the SGI® CXFS™
almost exclusively on tests concerning a filesystem, SGI can create a high-performance
relational database, ignoring the large and data-sharing environment in which all systems
growing pool of file data. For this reason, SGI, share access to data at SAN speeds while
LEGATO, StorageTek, Brocade, and LSI Logic achieving exceptional backup and restore
Storage Systems undertook an effort to bench- performance.
mark backup and restore performance using a
storage area network (SAN) under conditions 1.0 Taking Backup and Restore
typical for HPC. A variety of scenarios using Performance to a New Level
both image and file-by-file backup were tested Organizations that depend on high-performance
to help customers better understand the computing for critical research and develop-
strengths and limitations of each approach. ment are struggling to cope with data sets that
are growing at astronomical rates. The data
Using LEGATO NetWorker®, this benchmark management problem is not just one of provid-
achieved results that far exceeded previously ing adequate storage, but also providing
published numbers in every category tested: adequate protection of critical data through
• Sustained throughput in excess of 10TB per backup and restore. In many cases, the scope
hour was demonstrated for file-by-file of important simulations must be limited
backup because of an inability to effectively manage
• Over 10TB of data was backed up within 60 the huge volumes of data that can result from
minutes of initiating backup detailed models.
• A 1TB data set was backed up in only seven
minutes from start to finish SGI has long recognized the unique problems
• Backup of a single 10TB SGI® XFS® filesys- associated with the backup and restore of multi-
tem demonstrated sustained throughput in terabyte data sets. In 1997, SGI—in working
excess of 6TB per hour with LEGATO, IBM, and Computer Associ-
• Restore performance for all file-by-file tests ates—was the first to demonstrate the ability
ranged from 4 to 4.5TB per hour, up to two to back up a 1TB database in less than one
times the performance of the previous best hour. This record stood for over four and a half
result years until VERITAS demonstrated a 2TB-per-
• Image backups—using snapshots to ensure hour result in May 2002. Later in 2002, a group
consistency—achieved 7.2TB per hour sus- led by Computer Associates demonstrated a
tained throughput for backup and a record database backup and restore solution with sus-
7.9TB per hour sustained throughput for tained throughput of 2.6TB per hour for
restore backup and 2.2TB per hour for restore. Early
in 2003, Hewlett-Packard further upped the
These results were achieved using a 32- ante with a database benchmark that achieved
processor SGI® Origin® 3000 server, SGI® throughput of 3.62TB per hour for backup and
TP9500 RAID storage arrays (developed by LSI a more modest 1.29TB per hour for restore.
Logic Storage Systems, Inc., and marketed as While impressive by most standards, these
SGI TP9500), 48 StorageTek® T9940B 2Gb historical results have been targeted for data-
Fibre Channel tape drives installed in a base backup and have ignored the large and
StorageTek® PowderHorn® 9310 tape library, growing pool of file data. Even performance at
Brocade® Fibre Channel switches, and these levels may be inadequate for the needs of
LEGATO NetWorker® 7. All components are today’s large technical data stores.

2
For this reason, SGI and its partners— achieved for each and examines the hardware
LEGATO, Brocade, LSI Logic Storage Systems, and software components that were used to
and StorageTek—set out to demonstrate a achieve these results. The information
data-protection solution for today’s multi- included covers the highlights of all backup
terabyte environments. The objective was to and restore results. A technical white paper is
improve upon the current industry-best currently in preparation which will describe
backup and restore benchmarks by a factor of each of the tests and their respective configu-
three. As part of this benchmark effort, a vari- rations in greater detail.
ety of test scenarios were carried out to help
customers choose the best backup strategy for 2.0 Measured Backup and Restore
their particular needs. These results break all Performance
previously established performance records The backup-and-restore approach best suited
for both file-by-file and image backup and to a particular environment depends on a vari-
restore. Results of the various tests performed ety of factors, such as the total amount of data
are summarized in figure 1, along with the being stored and the frequency with which
results from the previously mentioned tests stored data changes. For this reason SGI and
for comparison. Each test is described in more its partners tested a variety of scenarios that
detail in the following section. were considered to be of interest for various
HPC environments. Each scenario measures
The key to achieving this level of performance backup and restore performance for a 10TB
is parallelism in the hardware and software data set, not a relational database housing
components that make up the solution. Each 10TB of data. Online database backup perfor-
component of the solution can sustain many mance depends on the database management
parallel data streams without interruption. system and the ongoing workload. Offline
This paper examines the various tests that database backup is equivalent to an ordinary
were carried out and the performance image or file backup of the devices or files

12
10.1
10 8.9
9

8 7.2
Terabytes/hour

6.3 7.9
6

4 4.5
4.1 4.4 3.6
2.6
2
3 2
1.5
0 2.2
1.3
10TB

1TB

10TB Single FS

Backup
10TB Image

xfsdump

HP

Restore
CA

VERITAS

Original 1TB/hr

Current Results Previous Results


Fig. 1. Sustained throughput achieved by various tests described in this paper, plus SGI’s
previously demonstrated 1TB-per-hour database backup (1997)

3
where the database resides. These results sim- mechanism to locate the blocks that corre-
ulate what a customer might experience with a spond to a particular file in the tape image.
data protection solution designed for an envi-
ronment with a large amount of file data to Three key metrics were measured for both
back up. The specifics of each test and the backup and restore performance for each test
rationale for choosing that particular scenario scenario:
are described in the following sections. Two
categories of backup and restore were tested: 1. Average throughput—the total amount of
data transferred divided by the total amount
1. File-by-file backup and restore using Net- of time needed for the operation to complete
Worker® to protect and catalog each file in 2.Sustained peak throughput—the throughput
each test filesystem: The advantages of this observed when the backup is running at a
approach are that it supports both full and sustained rate for a minimum period of 60
incremental backups, and individual files minutes
can be easily restored. A possible disadvan- 3.Data transferred in first 60 minutes—the
tage is that restore speeds are typically amount of data transferred in the first 60
slower than backups. minutes after the test is initiated
2.Image backup and restore using NetWorker:
The raw data from the storage volume is To understand the rationale for these metrics,
streamed directly to tape, resulting in an consider the performance profile for the file-
exact image of the volume including filesys- level 10TB-per-hour backup result illustrated in
tem data structures. Image backups are figure 2.
restored by copying the entire image from
tape back to the original storage volume or As figure 2 illustrates, it takes several minutes
a similar-sized volume. Therefore, they gen- for the backup to ramp up to full throughput
erally offer the same or better restore while tapes are being loaded and the various
performance as backup performance. How- parallel operations are initiated. Likewise, an
ever, individual files are difficult or additional ramp-down period exists at the end
impossible to restore unless there is a of the process as individual jobs complete at

Throughput Rate for 10TB File-by-File Backup


12

10

8
Terabytes/hour

0
0 500 1000 1500 2000 2500 3000 3500 4000
Wallclock Time (Seconds)

Fig. 2. Performance profile for file-by-file 10TB-per-hour backup test

4
different times. As a result, the average sciences, manufacturing, and media and
throughput may be significantly less than the entertainment.
sustained peak throughput. The impact of
these ramp-up and ramp-down periods on the Files in the data set ranged in size from 2GB
total backup time obviously depends on the to over 42GB. Multiple filesystems were used
amount of data to be backed up and time in most of the tests. The results of all tests are
period for completion. For data sets larger summarized in table 1, and a description of
than the 10TB data set used in these tests, each test scenario and results are given in the
ramp up and ramp down would be correspond- following subsections.
ingly less significant, while they become more
significant for smaller backups. The three test 2.2 File-by-File Backup and Restore (10TB)
metrics should therefore provide a more com- In this test, a 10TB data set was divided
plete picture to help assess performance on between 48 filesystems. NetWorker, running
data sets of various sizes. on the SGI® Origin® server, was used to initi-
ate a file-by-file backup of each filesystem.
2.1 A Real-World Data Set This test achieved sustained peak throughput
The data set used for this benchmark was pro- in excess of 10TB per hour for backup and also
vided by a large HPC customer and contains a exceeded the overall goal of backing up more
variety of application data files and personal than 10TB in the first 60 minutes, including
productivity files. The customer currently the ramp-up time needed to initialize all tape
manages hundreds of terabytes of data, which drives. These numbers are more than three
is typical of HPC customers across many times the level of performance achieved for
industries, including government, energy, any previously reported backup benchmark of

Table 1. Summary of performance results for all tests


NetWorker NetWorker NetWorker NetWorker xfsdump
File Backup File Backup File Backup Image Backup
10TB 1TB 10TB w/Snapshot
Single
Filesystem
# of Filesystems 48 48 1 48 48
Sustained Peak 10.09TB/hr NA 6.26TB/hr 7.24TB/hr 8.95TB/hr
Throughput
(Backup)
Sustained Peak 4.52TB/hr NA 4.43TB/hr 7.90TB/hr 3.96TB/hr
Throughput
(Restore)
Average 9.00TB/hr 8.86TB/hr 6.10TB/hr 6.14TB/hr 7.83TB/hr
Throughput
(Backup)
Average 3.81TB/hr 4.07TB/hr 3.99TB/hr 6.67TB/hr 3.79TB/hr
Throughput
(Restore)
Data Transferred in 10.05TB NA 6.20TB 7.16TB 8.92TB
First 60 Minutes
(Backup)
Data Transferred in 4.50TB NA 4.36TB 7.88TB 3.92TB
First 60 Minutes
(Restore)

5
any kind. As is often the case for this type of While this is less than the performance
backup, restore performance was less than achieved with multiple, separate filesystems,
backup performance with a sustained rate of it’s still more than double the best previously
4.52TB per second. This number is still well in reported backup speed. Restore performance
excess of the industry’s previous best for is equivalent to that seen with the other file-
restore performance. by-file tests included in this benchmark that
were performed using multiple filesystems.
2.3 File-by-File Backup and Restore (1TB)
This test is identical to the previous test, 2.5 Image Backup with Snapshots
except that the data set size is reduced to 1TB. For this scenario, the 10TB data set was again
This test was intended to demonstrate how divided between 48 filesystems. SGI® XVM
quickly a smaller data set could be backed up Snapshot (see Section 4) was used to create a
to tape, including initialization time. The point-in-time snapshot of the volume contain-
entire 1TB backup was accomplished in just ing each filesystem. NetWorker was used to
over seven minutes and restore was accom- perform an image backup of the raw data from
plished in less than 16 minutes. These each snapshot to tape. The time necessary to
numbers are impressive, considering that just create the snapshots was not reflected in the
a few years ago the best reported 1TB backup image backup benchmark, though snapshots
took over an hour. Only average throughput typically take less than one second to generate.
was measured due to the short duration of the This approach generally delivers faster restore
tests. times in comparison with file-level backup
methods, making it of particular interest to
2.4 Backup and Restore of a Single Large customers that may require faster restores of a
Filesystem large filesystem or data set as part of a disas-
Many customers find that managing one or a ter recovery or data replication scheme.
few large filesystems is preferable to managing
a large number of smaller ones, although the The sustained peak throughput for this test
number of filesystems needed may also be dic- was 7.24TB per hour for backup and 7.90TB
tated by organizational or other constraints. per hour for restore. The increase in restore
Because of the high performance and scalabil- performance in comparison to other tests
ity of the SGI XFS filesystem, customers often illustrates the fact that restores of single, large
find they can achieve the necessary perfor- images are faster because they eliminate the
mance and simplify operations with a single overhead necessary with traditional file-level
large filesystem. Both XFS and the SGI CXFS backup.
shared filesystem scale to address up to
18 exabytes as a single filesystem and support 2.6 Backup Using xfsdump (10TB)
a single file size of nine exabytes. For this test, The SGI XFS filesystem ships with a simple
the entire 10TB data set was stored in a single backup utility called xfsdump. The companion
XFS filesystem and backed up using LEGATO utility, xfsrestore, provides restore capabilities.
NetWorker. The filesystem was divided into 48 While xfsdump lacks the advanced file cata-
directories, where each of the directories loging features of a full-featured backup
included 16 files for an aggregate size of application such as LEGATO NetWorker, it
214.5GB, and was assigned to an individual does have the ability to recognize and back up
tape drive for backup. The single filesystem extended attributes native to both XFS and
contained 10.3TB of data with a capacity of CXFS filesystems. Many backup applications
12.3TB. do not look for extended attributes, and there-
fore do not recognize or copy these attributes
Even when protecting a large single filesystem, in the backup operation. Both xfsdump and
the sustained peak throughput for backup per- xfsrestore are useful in conjunction with
formance was an impressive 6.26TB per hour. applications that utilize extended file attrib-

6
utes, such as the SGI® Data Migration Facility Brocade 2Gb Fibre Channel switches was used
(DMF) data management solution, a multitiered as an interconnect between the SGI Origin
data life-cycle management technology that 3000 server, the SGI TP9500 storage array,
administers migration and provisioning policies. and the 48 T9940B tape drives in the
StorageTek PowderHorn 9310 tape library. All
Because these utilities are in widespread use components were chosen for their proven abil-
in DMF environments and elsewhere, compar- ity to support a large number of I/O streams
ison tests were executed to determine the in parallel and configured to ensure that no
performance of both xfsdump and xfsrestore. bottlenecks occurred under test conditions.
These tests compare favorably with file-by-file
backups using LEGATO NetWorker under the SGI Origin 3000 server: An Origin 3000 server
same conditions. Sustained peak throughput with 32 processors, 8GB of system memory,
for both backup and restore are somewhat and 48 2Gb Fibre Channel host bus adapters
lower than with NetWorker (8.95TB per hour (HBAs) acted as the backup server for all tests.
for backup and 3.96TB per hour for restore), The NUMAflex™ architecture of the SGI Origin
but this level of performance is still well 3000 servers is well known for its tremendous
beyond that reported in any previous bench- I/O capabilities. A system can easily be tai-
marks. lored to meet almost any I/O requirement.
Origin systems capable of achieving sustained
3.0 A Hardware Architecture Optimized I/O bandwidth in excess of 7GB per second
for Parallel I/O (24.6TB per hour) to and from a single file-
The hardware solution that was used to achieve system have been demonstrated, ensuring that
the backup and restore results described in the Origin server has the I/O bandwidth to
this paper is illustrated in figure 3. A storage exceed the requirements for these tests.
area network (SAN) fabric composed of During testing, this configuration consistently

8
8
8
Brocade
Silkworm 3900 (3x)

SGI Origin 3000 12 12 12


12 12 12
12
12
8
StorageTek T9940B (48x)
4
12
12
12
SGI TP9500 Arrays
8

StorageTek
LEGATO PowderHorn 9310 12
NetWorker 7
4
8 Brocade Brocade
8 Silkworm 3800 Silkworm 3200
StorageTek ACSLSTM

Ethernet

Fig. 3. Solution architecture (numbers indicate the number of individual channels between each
pair of targets)

7
read data off the disk subsystem at a rate of Brocade® SilkWorm® Fibre Channel fabric
over 10.4TB per hour into memory, then piped switches: A SAN fabric composed of three
it out at over 10.4TB per hour through the fab- SilkWorm® 3900 switches, one SilkWorm®
ric to the tape drives. The concurrent read and 3800 switch, and one SilkWorm® 3200 switch
write operations performed in duplex demon- was used to interconnect the Origin server
strates an aggregate sustained throughput rate with the disk and tape storage systems for
of over 20.8TB per hour (5.8GB per second) optimal performance. All three switch models
through the Origin server’s I/O subsystem. offer full 2Gb performance in a nonblocking
architecture. Hardware-enforced zoning was
SGI® Total Performance 9500 (TP9500) RAID used to ensure that each tape drive achieved
storage array: The high-performance disk stor- optimal performance.
age subsystem used in this benchmark was
developed by LSI Logic Storage Systems, Inc. 4.0 Advanced Software for Optimal
The storage system is configured according to Backup and Restore
SGI’s specifications and sold by SGI and its The software used in this testing was chosen
partners as the SGI TP9500 storage array. SGI for its ability to support a high degree of par-
TP9500 is a full end-to-end 2Gb Fibre Channel allelism. Two main software products were
solution, which uses parallel high-performance used.
controllers and advanced cache management
to achieve the throughput needed for these LEGATO NetWorker: Protecting more than
tests. The configuration included 120 Fibre 30,000 enterprise sites worldwide, NetWorker
Channel disk drives with a raw capacity of is the premier solution for information protec-
17.5TB, capable of sustaining a read rate of tion in multi-terabyte, heterogeneous
over 10.4TB per hour through the storage area environments. NetWorker supports a broad
network. A total of six dual-controller RAID range of platforms, filesystems, tape libraries,
enclosures funneled data through 24 Fibre and disk-backup devices, and it can be config-
Channel connections to the Fibre Channel ured to carry out a large number of backup
mesh fabric. In addition to tremendous I/O and/or restore operations in parallel, to ensure
throughput, SGI TP9500 also offers advanced that any set of operations can be carried out
features for high availability and data in the shortest possible time. A single Net-
protection. Worker server is highly scalable and can
automate backup, restore, and archival activi-
StorageTek T9940B tape drives: All backup and ties of hundreds to thousands of network
restore operations were performed using Stor- clients in DAS, NAS, and SAN environments.
ageTek T9940B tape drives with 2Gb Fibre Data can be sent to the server itself or to
Channel interface installed in a StorageTek independent storage nodes to optimize net-
PowderHorn 9310 tape library. A single silo work bandwidth. storage nodes also enable
can be configured with up to 1200TB of native protection of large systems to locally or SAN-
capacity and 960 individual tape drives with attached tape or disk storage devices.
100TB-per-hour native throughput. For these
tests, a single library was configured with 48 The Origin 3000 server was configured as the
of the T9940B tape drives, where each drive NetWorker server and directed to perform
used a single 200GB cartridge for each backup backups of local filesystems. The image
and restore test. At the time of its release in backup and restore tests were done using the
September 2002, the T9940B Fibre Channel commercially available NetWorker version 7.0
tape drive offered the fastest data transfer rate Power Edition. For the traditional backup and
(30MB per second native and 70MB per sec- restore tests, NetWorker 7.0 with a direct I/O
ond with compression) and the highest- feature was used. This feature is available via
capacity cartridges (200GB native) of any tape patch for version 7.0 and will be native to ver-
drive on the market. sion 7.1, scheduled for release in fall 2003.

8
SGI XVM Snapshot: XVM Snapshot is an for optimum scalability and data consolidation,
optional extension for the SGI XVM volume creating a backup environment that is not only
manager that provides snapshot capabilities at highly scalable, but also simple—especially
the volume level. Volume-level snapshots are given the level of performance achieved—and
more convenient and efficient than snapshot easy to manage.
products that work at the LUN level, since a
single volume can span many LUNs. XVM The Origin backup server included both XFS
Snapshot can take a snapshot of an entire XFS and CXFS filesystems, but all tests were run
or CXFS filesystem, minimizing the number of using XFS. The SGI CXFS shared filesystem
snapshots needed, especially for environments has demonstrated performance levels similar
that use a single large filesystem. Among other to the XFS filesystem when implemented in a
purposes, snapshots are frequently used to SAN. While it was not explicitly tested during
provide a consistent image for backup of an this benchmark, SGI believes that results simi-
active filesystem. When a snapshot is in place, lar to the actual achievements in this
the original versions of changed blocks are benchmark would be achieved by using CXFS
automatically copied to a snapshot partition instead of XFS. Other distributed filesystems
(copy on write). When the snapshot is accessed don't provide the performance expected from
for backup, restore, or other purposes, the a native filesystem. CXFS provides data shar-
original blocks are read in lieu of the changed ing capabilities across a SAN with the
blocks, preserving a consistent image of the performance of a native filesystem. CXFS
volume from a single point in time. creates the opportunity to design a heteroge-
neous shared data environment in which all
For this testing, XVM Snapshot was used in computer systems have high-speed access to
combination with LEGATO NetWorker for shared data (multiple gigabytes per second),
image backups. XVM Snapshot can also be plus the backup and restore performance
used in combination with LEGATO NetWorker required for extremely large data sets. An Ori-
or xfsdump during file-by-file backups. Net- gin server in such a configuration acts as a
Worker has built-in capabilities to recognize backup server, accessing all data at native-
when individual files are changing to help filesystem speed via CXFS. The server can also
ensure consistent backups of live filesystems be used for other purposes when high-speed
in situations where snapshots are not used. backup and restore are not in progress.

5.0 Beyond Benchmarks Achieving optimal backup and restore perfor-


All hardware and software used in this bench- mance tailored to real requirements in the real
mark is readily available. Therefore, unlike world is not a trivial task. Most sites already
many laboratory benchmarks, customers will have a substantial investment in servers and
be able to duplicate this configuration with disk and tape storage systems that must be
the same or similar components and can preserved. SGI Professional Services can pro-
expect to achieve similar levels of performance. vide the storage expertise to help customers
choose the appropriate hardware and software
The key to achieving the levels of performance components for their needs and to integrate
demonstrated by these benchmarks is provid- those components in their existing environ-
ing the parallelism to achieve the desired level ment to maximize performance.
of throughput. Because of its unique
NUMAflex architecture, a single SGI Origin
server can scale to these levels and beyond.
The other components chosen for these
benchmarks offer similar levels of parallelism

9
Corporate Office North America 1(800) 800-7441
1600 Amphitheatre Pkwy. Latin America (52) 5267-1387
Mountain View, CA 94043 Europe (44) 118.925.75.00
(650) 960-1980 Japan (81) 3.5488.1811
www.sgi.com Asia Pacific (65) 6771.0290

© 2003 Silicon Graphics, Inc. All rights reserved. Silicon Graphics, SGI, XFS, Origin, and the SGI logo are registered trademarks and CXFS
and NUMAflex are trademarks of Silicon Graphics, Inc., in the United States and/or other countries worldwide. All other trademarks
mentioned herein are the trademarks of their respective owners.

3523 07/23/2003] J14309

10

S-ar putea să vă placă și