Documente Academic
Documente Profesional
Documente Cultură
Marc Farley President of Building Storage, Inc Author, Building Storage Networks, Inc.
Agenda
Lesson 1: Basics of SANs Lesson 2: The I/O path Lesson 3: Storage subsystems Lesson 4: RAID, volume management and virtualization Lesson 5: SAN network technology Lesson 6: File systems
Lesson #1
Connecting
Concentrator Router
Dish
VPN
Connecting
Networking or bus technology Cables + connectors System adapters + network device drivers Network devices such as hubs, switches, routers Virtual networking Flow control Network security
Storing
Host Software
Mirroring Software
Storage Protocol
Storage Devices
Tape Drives
Storing
Device (target) command and control
Drives, subsystems, device emulation
Filing
C:\directory\file
Database Object
User/Application View
User/Application View
Filing
Namespace presents data to end users and applications as files and directories (folders) Manages use of storage address spaces Metadata for identifying data
file name owner dates
Filing
Wiring Connecting
HBA or NIC
Cable Cable
Storing
Network Switch/hub
Disk Drive
NAS is filing over a network SAN is storing over a network NAS and SAN are independent technologies They can be implemented independently They can co-exist in the same environment They can both operate and provide services to the same users/applications
NAS
SAN Network
Storing Storing
Wiring Connecting
Storing Storing
Lesson #2
Memory
Processor
Memory Bus
MultiPathing
Device Driver
Switches, hubs, routers, bridges, gatways Port buffers, processors Backplane, bus, crossbar, mesh, memory
Fabric Services
Routing
Flow Control
Virtual Networking
Network Ports
Cache
Resource Manager
Disk drives
Tape drives
Tape Media
App
Cabling
Network Systems
Fabric Services
Routing
Flow Control
Virtual Networking
Cache
Resource Manager
Disk drives
Tape drives
Lesson #3
Storage subsystems
Cache Memory
Power
Multiple hot swappable power supplies Hot swappable cooling fans Data redundancy via RAID Multi-path support
Network ports to storage resources
Exported storage
Exported storage
Exported storage
Exported storage
Exported storage
Exported storage
unique ID (UUID) - sometimes referred to as a serial number An LU can be represented by multiple logical unit numbers (LUNs) Provisioning associates LUNs with LUs & subsystem ports
Provisioning storage
LUN 0 LUN 1
SCSI LU UUID A
Port S1
LUN 1 LUN 2 LUN 2
SCSI LU UUID B
Port S2
Port S3
LUN 3 LUN 3
SCSI LU UUID C
Port S4
LUN 0
SCSI LU UUID D
Controller functions
Multipathing
LUN X
Path 1
SCSI LU UUID A
MP SW
LUN X
Path 2
Caching
Exported Volume Exported Volume
Write Caches 1. Write Through (to disk) 2. Write Back (from cache)
Tape subsystems
Tape Drive Tape Drive Tape Drive Tape Drive
Tape Slots
Robot
Subsystem management
Ethernet/TCP/IP
Out-of-band management port
In-band management
Storage Subsystem
Data redundancy
Duplication
2n
Parity
n+1
Difference
-1
d(x) = f(x) f(x-1) f(x-1)
I/O Path
Mirroring Operator Terminate I/O & regenerate new I/Os Error recovery/notification
Point-in-time snapshot
Subsystem Snapshot
Host
Lesson #4
Duplication
2n
Parity
n+1
Difference
-1
d(x) = f(x) f(x-1) f(x-1)
Original goals were to reduce the cost and increase the capacity of large disk storage
Benefits of RAID
Capacity scaling
Disk mirroring (striping to 2 disks) Parity RAID (striping to more than 2 disks)
Capacity scaling
Combined extents 1 - 12 Storage extent 1 Storage extent 2 Storage extent 3 Storage extent 4
Storage extent 5
Storage extent 6
Storage extent 7
Storage extent 8
Storage extent11
Storage extent12
Performance
RAID controller (microsecond performance)
Disk drive
Disk drive
Disk drive
Disk drive
Disk drive
Disk drive
Disk drives (Millisecond performance) from rotational latency and seek time
Parity redundancy
RAID arrays use XOR for calculating parity
Operand 1 False False True True Operand 2 False True False True XOR Result False True True False
XOR {M1&M2&M3&P}
Parity rebuild
The process of recreating data on a replacement member is called a parity rebuild Parity rebuilds are often scheduled for non-production hours because performance disruptions can be so severe
XOR {M1&M2&M3&P}
RAID 0+1, 10
RAID Controller Hybrid RAID: 0+1
OS kernel
File system
Volume management
RAID & partition management Device driver layer between the kernel and storage I/O drivers
Volume Manager
Server system
Volume managers can use all available connections and resources and can span multiple SANs as well as SCSI and SAN resources
Virtual Storage
SAN HBA
In-band virtualization
Exported virtual storage
Disk subsystems
Out-of-band virtualization
Virtualization agents
Disk subsystems
Lesson #5
SAN networks
Fibre channel
The first major SAN networking technology Very low latency High reliability Fiber optic cables Copper cables Extended distance 1, 2 or 4 Gb transmission speeds Strongly typed
Fibre channel
A Fibre Channel fabric presents a consistent interface and set of services across all switches in a network Host and subsystems all 'see' the same resources
Leveraging the install base of Ethernet and TCP/IP networks iSCSI native SAN over IP FC/IP FC SAN extensions over IP
iSCSI
Native storage I/O over TCP/IP
New industry standard Locally over Gigabit Ethernet Remotely over ATM, SONET, 10Gb Ethernet
iSCSI equipment
Cables
Copper and fiber
Network systems
Switches/routers Firewalls
FC/IP
Extending FC SANs over TCP/IP networks FCIP gateways operate as virtual E-port connections FCIP creates a single fabric where all resources appear to be local
E-port
E-port
High-end SAN switches have latencies of 1 - 3 sec Transaction processing requires lowest latency
Most other applications do not
Directors
64+ ports HA redundancy Dual system supervisor Live SW upgrades
Lesson #6
File systems
Filing Filing
Storing Storing
Think of the storage address space as a sequence of storage locations (a flat address space)
1 2 10 11 19 20 28 29 37 38 46 . 55 . 64 . 73 . 82 83
3
12 21 . . . . . . 84
4
13 . . . . . . . 85
6 7
16 . . . . . . . 88
8
17 26 35 44 53 62 71 80 89
9
18 27 36 45 54 63 72 81 90
14 15 . . . . . . . . . . . . . . 86 87
Superblocks are known addresses used to find Superblocks file system roots (and mount the file system)
SB
SB
File systems must have a known and Filing and Scaling dependable address space
The fine print in scalability - How does the filing function know about the new storing address space?
1 7 13 19 25 31 37
2 8 14 20 26 32 38
3 9 15 21 27 33 39
4 10 16 22 28 34 40
5 11 17 23 29 35 41
6 12 18 24 30 36 42
Filing Filing
1 6 11 16 21
2 7 12 17 22
3 8 13 18 23
4 9 14 19 24
5 10 15 20 25