Beruflich Dokumente
Kultur Dokumente
Copy right IBM Corporation 2009 Course materials may not be reproduced in whole or in part without the prior written permission of IBM.
5.1
Unit objectives
After completing this unit, you should be able to: Discuss the hardware and architecture of the DS8000 Use virtualization terminology describing configuration of the DS8000 subsystem Describe the physical hardware components and resources Describe the models and features provided by each model Describe the types of disk arrays that can be configured for a DS8000 subsystem Explain the cabling between adapters and drive sets
Copyright IBM Corporation 2009
DS8000 highlights
Copy right IBM Corporation 2009 Course materials may not be reproduced in whole or in part without the prior written permission of IBM.
5.1
DS8000 highlights
New processor family - POWER5+ RISC (DS8000 Turbo)
DS8100 model 931 DS8300 model 932/ 9B2 for LPAR
I/O adapters
Fibre Channel/FICON host adapter (4 ports, 4 Gb/s) ESCON host adapter (2 ports, 18 MB/s) FC-AL device adapter (4 ports, 2 Gb/s)
FC-AL disks
73 GB, 146 GB, 300 GB or 450 GB at 15K rpm FATA disk drives of 500 GB or 1 TB / 7200 rpm 73 GB or 146 GB solid state disks
Copyright IBM Corporation 2009
Graphical user interface (GUI) or command line interface (CLI) allows: For high availability, hardware components are redundant
Copyright IBM Corporation 2009
DS8000 R2 highlights
R2: Announcing new features for ALL models: IBM POWER5+ processor: New DS8000 Turbo (93x/9Bx) Processor memory for POWER 5+ processor 4 Gb FCP/FICON adapter (available on all models 92x/9Ax and 93x/9Bx) 500 GB 7200 rpm FATA drives (available on all models 92x/9Ax and 93x/9Bx) Three-site Metro/Global Mirror Earthquake resistance kit Ethernet adapter pair (for TPC RM support) Performance Accelerator (models 932 and 92E only) 300 GB 15,000 rpm fibre channel drives HyperPAV (System z)
DS8000 R3 highlights
R3: Announcing new features for ALL models: Dynamic increase in size of existing volumes
No need to move data or delete/redefine May be used to prevent out of space conditions or to migrate from one volume size to another
Storage pool striping Adaptive multi-stream pre-fetching FlashCopy space efficient (SE) New Secure Socket Layer (SSL) option for call home New console, the System Storage Productivity Center (SSPC) to manage the full data center from a single point.
DS8000 R4 highlights
R4: Announcing new features for ALL models 450 GB/15,000 RPM DDMs 1 TB / 7.2 RPM FATA DDMs 73 GB and 146 GB SSDs RAID 6 DSCLI and Storage Manager changes GUI response time improvement on panel loads Disk encryption Variable LPAR Extended address volumes IPv6 Secure data overwrite service offering Laptop HMC
DS8000 R5 highlights
R5: Announcing a complete new model! new IO towers new DA cards new CECs new CPUs (Power6) increased performance increased stability easier upgrade path
DS8000
General support
http://www.ibm.com/servers/storage/support/disk/ds8300/
DS8000 CLI and Storage Manager client software levels and download
http://www.ibm.com/support/docview.wss?uid=ssg1S4000420
Intel servers:
Hewlett-Packard: Sun:
With DS6000 and DS8000, SDD is supported on the following operating systems:
Windows NetWare AIX HP-UX Sun Solaris Linux
SDD can coexist with RDAC (DS4000 multipath driver) on most operating systems as long they manage separate HBAs. SDD cannot be used with most other multipath drivers (in other words, Veritas, PV-Links, Powerpath)
Copyright IBM Corporation 2009
Both
Interfaces to DS8000 (2 of 2)
DS open application programming interface (API)
Non-proprietary storage management client application supporting:
Routine LUN management activities (creation, mapping, masking) Creation or deletion of RAID volume spaces Copy Services functions: FlashCopy, PPRC
Helps to integrate configuration management support into existing storage resource management (SRM) applications Enables automation of configuration management through customerwritten applications Complements the use of Web-based DS-SM and script-based DSCLI Implemented through IBM System Storage Common Information Model (CIM) agent Uses CIM technology to manage proprietary devices as open system devices through storage management applications Allows these applications to communicate with a storage unit Used by TPC for disk
Copyright IBM Corporation 2009
Copy right IBM Corporation 2009 Course materials may not be reproduced in whole or in part without the prior written permission of IBM.
5.1
External management console (optional, feature code 1100) Internal management console (feat code 1110)
For redundancy with high availability
Remote service
Call home and call back
Network Interface Server for logical configuration and invocation of advanced Copy Services functions Connection to storage facility (DS8000) through redundant private Ethernet networks only
AIX
Partition 2
POWER5 Hypervisor
Unassigned Resources
Ethernet
HMC
Non-Volatile RAM
Service Processor
Perm Temp
P5 HMC features:
Logical partition configuration Dynamic logical partitioning Capacity and resource management System status HMC management Service functions (microcode update, ) Remote HMC interface
One PCI modem for asynchronous Call home support. Corresponding private Ethernet ports of the external HMC (FC1110) would be plugged into port 2 of the switches as shown in next foil. To interconnect two DS8000 base frames, FC1190 would provide a pair of 31m Ethernet cables to connect from port 16 of each switch in the second base frame into port 15 of the first frame.
If the second HMC is installed in the second DS8000, it would remain plugged into port 1 of its Ethernet switches.
Copyright IBM Corporation 2009
PCI Modem
Call home to IBM Services is possible through dial-up (PCI Modem in the HMC) or Internet connection VPNs Dial-up or Internet connection VPNs are also available for IBM service to provide remote service and support Recommended configuration is to connect HMC to customers public network for support Network connectivity and remote support is managed by the HMC
Support will use WebSM GUI for all service actions Downloading of problem determination data favors the use of a high-speed network
This view shows the laptop HMC display opened and ready to use. Note that the USB cables plugged into the side of the laptop HMC and the power cord at the rear. You can also see the power cord transformer strapped to the rear of the tray with Velcro. To open the DVD tray, the laptop HMC can rotated in clockwise direction.
Here the turntable and laptop HMC are fully rotated 90 degrees and the DVD drives DVD tray is extended.
DS8000 Subsystems
DMZ VPN
Redundant HMC Ethernet Fabric eth eth modem eth integrated Firewall Proxy
Internet Internet
VPN DMZ
Hardware Management Console Demilitarized Zone Virtual Private Network Ethernet Port
The VPN Server security complies with IBM corporate security standards ITCS104
This is an IBM internal security measure for all IBM secure data.
VPN Tunnel
DS-6000 SMC Firewall
RS3
IB M S u p p o rt
Customer Site
Copyright IBM Corporation 2009
IBM Site
Hardware components
Copy right IBM Corporation 2009 Course materials may not be reproduced in whole or in part without the prior written permission of IBM.
5.1
Redundant Power
BBU:
Battery Backup Units
Host Adapter
4 FCP/FICON Ports or 2 ESCON Ports
Front
Device Adapter
4 FC-AL Ports
Rear
DS8000 terminology
Storage complex Storage unit
A group of DS8000s managed by a single Storage-Hardware Management Console A single DS8000 including expansion frames. One P5-570 p Series server Two processor complexes form a redundant pair Divided into one LPAR (models 931 or 932) or two LPARs (model 9B2) The software that uses an LPAR: Has access to a percentage of resources available on the processor complex for the LPAR At GA, this percentage is 50% (model 9B2) or 100 % (models 931 or 932) Union of 2 LPARs, one from each processor complex Each LPAR hosts one storage server.
Copyright IBM Corporation 2009
Processor complex
Storage server
Under normal operation, both DS8000 servers are actively processing I/O requests.
Copyright IBM Corporation 2009
Processor complex
Processor complex
Processor complex
Processor complex
CEC enclosures in the Model 921/931 each have one processor card (2-way) CEC enclosures in the Model 922/932 and 9A2/9B2 each have two processor cards (4-way)
CEC: Computer Electronic Complex CEC enclosures contain components such as the processor cards, cache memory, and CEC hard drives.
The p5 hardware and Hypervisor manage the real to virtual memory mapping to provide robust isolation between LPARs. IBM has been doing LPARs for 20 years in mainframes and 8 years in System p. A 4-way has two processors to one LPAR and two processors to the other LPAR. LPARs only possible in the 4-way P5s (RIO-G cannot be shared in 2-way).
Post GA, 25-75 possible.
Cache memory ranges from 16 GB to 256 GB Persistent memory ranges from 1 GB to 8 GB: dependent on cache size Copyright IBM Corporation 2009
In an LPAR:
An operating system instance runs with dedicated (AIX 5.2) or shared (AIX 5.3) resources: processors, memory, and I/O slots These resources are assigned to the logical partition The total amount of assignable resources is limited by the physically installed resources in the system Isolation between LPARs to prevent unauthorized access between partition boundaries Fault isolation such that one LPARs operation does not interfere with the operation of other LPARs Support for multiple independent workloads, different operating systems, operating system levels, applications, and so on
LPARs provide:
Processor compl ex 1
LPAR11
DS8300
Comprised of two eServers P5 570 (Processor complex) Each processor complex supports one or more LPARs Currently, each processor complex divided into two LPARs
LPAR01
LPAR02
LPAR12
LPARxy x=Pr ocessorcom plex num ber y=Stor age facility num ber
(Persistent memory)
In case of power failure, if the 2107 has pinned data in cache, it is written to an extra set of two disk drives located in each of the CEC enclosures Two disk drives total in each CEC:
For LIC (LVM Mirrored AIX 5.3 + DS8000 code) For pinned data and other CEC functions
During the recovery process, the pinned data can be restored from the extra set of CEC disk drives just as it would have been from the NVS Copyright IBM Corporation cards on2009 the ESS 800
Processor complex
Processor complex
Accesses cache via RIO-G internal bus Each adapter has its own PowerPC processor Owned by processors in LPAR Uses system power control network (SPCN)
Controls and monitors the status of the power and cooling within the I/O enclosure Cabled as a loop between the different I/O enclosures
Copyright IBM Corporation 2009
Processor complex
Processor complex
The personality of the port is changeable via the DS storage management tools (GUI or CLI).
Ports cannot operate as FCP and FICON simultaneously FCP port can be longwave or shortwave
Shortwave ports support a distance of 300m (non-repeated) Longwave ports support a distance of 10Km (non-repeated)
PPC 750GX
Processor 1 GHz
Buf f er
Protocol Chipset
QDR
Data Mover
New high function/high performance ASIC Metadata creation/checking Configured at port level Fibre or FICON
SW or LW
4 Gb / 2 GB HA performance comparison
Processor complex
Processor complex
Two simultaneous I/O ops per FC-AL connection possible Switched FC-AL or switched bunch of disks (SBOF) used for back-end access
Device adapters are attached to a FC switch with the enclosure Four paths to each drive: 2 FC-AL loops X dual port access
(Detailed later with storage enclosures cabling)
Copyright IBM Corporation 2009
PPC 750FX
NVRAM
Bridge
Fibre Channel Protocol Engine RAID: Data Protection Data Mov er ASIC Buf f er
Protocol Chipset
Data Mover
Four 2 Gbps fibre channel ports New high function/high performance ASIC Metadata checking
Device adapter
Device Adapter
cards counted left to right, starting with 0 ports counted top to bottom, starting with 0 naming scheme in DSCLI (lsioport): I0XYZ X=enclosure number (0-7) Y=card number (0-5) Z=port number (0-3) e.g. I0312: IO enclosure 3 (1st rack, bottom,
right)
Slot 2 Slot 4 Enclosure 1 I0210 I0211 I0212 I0213 I0230 I0231 I0232 I0233
Slot 2 Slot 4 Enclosure 2 I0310 I0311 I0312 I0313 I0330 I0331 I0332 I0333
Slot 2 Slot 4 Enclosure 5 I0610 I0611 I0612 I0613 I0630 I0631 I0632 I0633
Slot 2 Slot 4 Enclosure 6 I0710 I0711 I0712 I0713 I0730 I0731 I0732 I0733
Bottom Half (8 DDMs) Interface card (FCIC) Disk Drive Module (DDM)
DS8000: Disk enclosures installed in pairs: one in front and one in back
Copyright IBM Corporation 2009
Switched connections to each DDM Each DDM connected to 2 switches Each switch connected to 2 device adapters Each device adapter belongs to a different CEC 4 connections to each DDM high redundancy
62
Switched FC-AL
Drives attached in point-to-point connection Faster arbitration message processing 200 MB/sec external transfer rate Improved RAS Switch detects individual failures Intermittent/permanent
The DS8000 architecture employs dual redundant switched FC-AL access to each of the disk enclosures. The key benefits of doing this are:
Two independent switched networks to access the disk enclosures Four access paths to each DDM in DS8000 architecture (dual switches) Each device adapter port operates independently Double the bandwidth over traditional FC-AL loop implementations
When a connection is made between the device adapter and a disk, the connection is a switched connection that uses arbitrated loop protocol.
Copyright IBM Corporation 2009
This means that a mini-loop is created between the device adapter and the disk Results in four simultaneous and independent connections, one from each device adapter port
Copy right IBM Corporation 2009 Course materials may not be reproduced in whole or in part without the prior written permission of IBM.
5.1
DS8000 frames
Base frame:
The base frame contains two processor complexes: eServer p5 570 servers
Each of them contains the processor and memory that drive all functions within the DS8000.
The base frame can contain up to eight disk enclosures; each can contain up to 16 disk drives. The base frame contains four I/O enclosures.
In a maximum configuration, the base frame can hold 128 disk drives. I/O enclosures provide connectivity between the adapters and the processors. The adapters contained in the I/O enclosures can be either device or host adapters (DAs or HAs).
The communication path used for adapter to processor complex communication is the RIO-G loop.
Expansion frames:
Each expansion frame can hold up to 16 disk enclosures which contain the disk drives. Expansion frames can contain four I/O enclosures and adapters if they are the first expansion frame that is attached to either a model 932 or a model 9B2.
In a maximum configuration, an expansion frame can hold 256 disk drives.
Up to 128 disks
Batteries
I/O drawers
Up to 640 Disks
(maximum configuration)
2 2 0 0 C0 C1
3 3 1 1 2 2 0 0
b
Balanced configuration at 256 DDMs: In other words, 64 DDMs per DA pair DA (card) plugging order: 2 / 0 / 3 / 1
Up to eight DA pairs
DA pairs 0 to 7 Automatically configured from DDMs
2 2 0 0 C0 C1
6 6 4 4 7 7 5 5
b
3 3 1 1 2 2 0 0
b
3 3 1 1 2 2 0 0
b
6 6 4 4 7 7 5 5
b
3 3 1 1 0 0 1 1
b
4/6 6/4
Copy right IBM Corporation 2009 Course materials may not be reproduced in whole or in part without the prior written permission of IBM.
5.1
SARC uses:
Demand paging for all standard disk I/O Sequential pre-fetch for sequential I/O patterns
1 0.8 0.6 0.4 0.2 0 0 64 128 192 256 Cache Size (GB) z/OS Open
Benefits
Nimrod Megiddo and Dharmendra S. Modha, "Outperf orming LRU with an Adaptiv e Replacement Cache Algorithm," IEEE Computer, pp. 4-11, April
Copyright 2004. IBM Corporation 2009
AMP optimizes cache efficiency by incorporating an autonomic, workloadresponsive, self-optimizing prefetching technology .
The algorithm dynamically decides what to prefetch and when to prefetch Delivers up to a two-fold increase in the sequential read capacity of RAID 5 arrays The bandwidth for a fully configured DS8000 remains unchanged May improve sequential read performance for smaller configurations and single arrays Reduces the potential for array hot spots due to extreme sequential workload demands May significantly reduce elapsed time for sequential read applications constrained by array bandwidth such as BI and critical batch processing workloads
Copy right IBM Corporation 2009 Course materials may not be reproduced in whole or in part without the prior written permission of IBM.
5.1
Server RAS (1 of 2)
The DS8000 provides data integrity when performing write operations and server failover.
Metadata check: The metadata is checked by various internal components to validate the integrity of the data as it moves throughout the disk system or sent back to the host. Server failover and failback:
LSS and server affinity:
LSS with even number have an affinity with server 0 LSS with odd number have an affinity with server 1
When a host operating system issues a write to a logical volume, the DS8000 host adapter directs that write to the server that owns the LSS to which that logical volume is a member.
Server RAS (2 of 2)
Under normal operation, both DS8000 servers are actively processing I/O requests
Each write is placed into the cache memory of the server owning the volume and also into the NVS memory of the alternate server.
Failover: In case of one server failure, the remaining server is able to take over all of its functions
RAID arrays which are connected to both servers can be accessed from the device adapters of the remaining server. Since the DS8000 has only one copy of data in cache of remaining server, it will now take the following steps:
It de-stages the contents of its NVS to the disk subsystem. The NVS and cache of remaining server are divided in two, half for the odd LSSs and half for the even LSSs. Remaining server now begins processing the writes (and reads) for all the LSSs.
Failback: When the failed server has been repaired, failback process is activated
Copyright IBM Corporation 2009
It starts in less than 8 seconds, will finish in less than 15 minutes, and is invisible to the attached hosts.
LPAR Hypervisor
takes par tof takes par tof takes par tof
Server failover
Normal flow of data for a write:
1. Data is written to cache memory in the owning server. 2. Data is written to NVS memory of the alternate server. 3. The write is reported to the attached host as having been completed. 4. The write is destaged from the cache memory to disk. 5. The write is then discarded from the NVS memory of the alternate server.
NV S f orodd LS Ss NV S f oreven LS S s
After a failover, remaining server processes all I/Os with cache and NVS divided by two, one for odd LSSs and one for even LSSs.
NV S f orodd LS S s
V S N V SN o r f o r f o dd even LS S s LS S s
Ser ver 1
Failover
Copyright IBM Corporation 2009
Scenario at power-on
Processor complexes power-on and perform power-on self-test Each server boots up During boot-up, each server detects NVS data on its disks and destages it to FC-AL disks When battery units reach a certain level of charge, the servers come online
NVS contents preserved indefinitely Note: The servers will not come online until the batteries are fully
Copyright IBM Corporation 2009
Disk subsystem (1 of 2)
RAID 5, RAID 6, and RAID 10
RAID 5 (7+P or 6+P+S) or RAID 10 (2x4 or 2x3 + 2S)
RAID 6 adds extra parity drive, the Q drive, now 6+P+Q versus 7+P or 5+P+Q+S versus 6+P+S
A minimum of one spare is created for each array site defined until the following conditions are met:
A minimum of four spares per DA pair A minimum of four spares of the largest capacity array site on the DA pair A minimum of two spares of capacity and RPM greater than or equal to the fastest array site of any given capacity on the DA pair
Floating spare
The DS8000 microcode may choose to migrate new spare disks to a more optimal position to better balance the spares across the DA pairs, the loops, and the enclosures.
Useful after a drive replacement that became a spare drive
Disk subsystem (2 of 2)
Each DDM attached to two FC switches
Each disk has two separate connections on the backplane.
Each DA connected to the two switches DDMs hot-pluggable Incorporates predictive Failure Analysis (PFA)
Anticipates failures
Microcode update
Concurrent code update Installation process:
HMC can hold six different versions of code Each server can hold three different versions of code Internal HMC code update New DS8000 LIC downloaded on the internal HMC LIC uploaded from HMC to each DS8000 server internal storage New firmware can be loaded from HMC directly into each device
May require server reboot with failover of its logical subsystems to the other server
Redundant HMCs
Each DS8000 can be connected via the redundant Ethernet switches to both HMCs.
Unit summary
Having completed this unit, you should be able to: Discuss the hardware and architecture of the DS8000 Use virtualization terminology describing configuration of the DS8000 subsystem Describe the physical hardware components and resources Describe the models and features provided by each model Describe the types of disk arrays that can be configured for a DS8000 subsystem
2107-932
Base rack, 4 way
2107-9B2
Base rack, 4 way, LPAR model
242x-9yy
New models which include the warranty in the model number x: 1, 2, 3, 4 (years of warranty) For all rack models
242x-92E
Expansion rack to a 931 or 932
242x-9BE
Expansion rack to a 9B2
93 DS8000 | Concepts and Architecture | K. Jehnen 2008 IBM Corporation