Beruflich Dokumente
Kultur Dokumente
TotalStorage
Productivity Center V3.3
Update Guide
Implement TPC V3.3 on supported
platforms
Mary Lovelace
Mathias Defiebre
Harsha Gunatilaka
Curtis Neal
Yuan Xu
ibm.com/redbooks
International Technical Support Organization
April 2008
SG24-7490-00
Note: Before using this information and the product it supports, read the information in “Notices” on
page vii.
This edition applies to Version 3, Release 3 of TotalStorage Productivity Center (product number 5608-VC0).
Notices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vii
Trademarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . viii
Preface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ix
The team that wrote this book . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ix
Become a published author . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xi
Comments welcome. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xi
Contents v
Appendix C. Performance metrics in TPC Performance Reports. . . . . . . . . . . . . . . . 315
Performance metric collection . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 316
By Storage Subsystem report . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 317
By Controller report . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 323
By I/O Group report . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 326
By Node report . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 331
By Array report . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 336
By Managed Disk Group report . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 339
By Volume report . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 341
By Managed Disk report. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 344
By Port report - Storage . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 345
By Port report - Fabric . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 347
Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 351
This information was developed for products and services offered in the U.S.A.
IBM may not offer the products, services, or features discussed in this document in other countries. Consult
your local IBM representative for information on the products and services currently available in your area. Any
reference to an IBM product, program, or service is not intended to state or imply that only that IBM product,
program, or service may be used. Any functionally equivalent product, program, or service that does not
infringe any IBM intellectual property right may be used instead. However, it is the user's responsibility to
evaluate and verify the operation of any non-IBM product, program, or service.
IBM may have patents or pending patent applications covering subject matter described in this document. The
furnishing of this document does not give you any license to these patents. You can send license inquiries, in
writing, to:
IBM Director of Licensing, IBM Corporation, North Castle Drive, Armonk, NY 10504-1785 U.S.A.
The following paragraph does not apply to the United Kingdom or any other country where such
provisions are inconsistent with local law: INTERNATIONAL BUSINESS MACHINES CORPORATION
PROVIDES THIS PUBLICATION "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESS OR
IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF NON-INFRINGEMENT,
MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Some states do not allow disclaimer of
express or implied warranties in certain transactions, therefore, this statement may not apply to you.
This information could include technical inaccuracies or typographical errors. Changes are periodically made
to the information herein; these changes will be incorporated in new editions of the publication. IBM may make
improvements and/or changes in the product(s) and/or the program(s) described in this publication at any time
without notice.
Any references in this information to non-IBM Web sites are provided for convenience only and do not in any
manner serve as an endorsement of those Web sites. The materials at those Web sites are not part of the
materials for this IBM product and use of those Web sites is at your own risk.
IBM may use or distribute any of the information you supply in any way it believes appropriate without incurring
any obligation to you.
Information concerning non-IBM products was obtained from the suppliers of those products, their published
announcements or other publicly available sources. IBM has not tested those products and cannot confirm the
accuracy of performance, compatibility or any other claims related to non-IBM products. Questions on the
capabilities of non-IBM products should be addressed to the suppliers of those products.
This information contains examples of data and reports used in daily business operations. To illustrate them
as completely as possible, the examples include the names of individuals, companies, brands, and products.
All of these names are fictitious and any similarity to the names and addresses used by an actual business
enterprise is entirely coincidental.
COPYRIGHT LICENSE:
This information contains sample application programs in source language, which illustrate programming
techniques on various operating platforms. You may copy, modify, and distribute these sample programs in
any form without payment to IBM, for the purposes of developing, using, marketing or distributing application
programs conforming to the application programming interface for the operating platform for which the sample
programs are written. These examples have not been thoroughly tested under all conditions. IBM, therefore,
cannot guarantee or imply reliability, serviceability, or function of these programs.
Oracle, JD Edwards, PeopleSoft, Siebel, and TopLink are registered trademarks of Oracle Corporation and/or
its affiliates.
Snapshot, Network Appliance, FilerView, Data ONTAP, NetApp, and the Network Appliance logo are
trademarks or registered trademarks of Network Appliance, Inc. in the U.S. and other countries.
Java, JVM, J2SE, Solaris, and all Java-based trademarks are trademarks of Sun Microsystems, Inc. in the
United States, other countries, or both.
Active Directory, Microsoft, SQL Server, Windows, and the Windows logo are trademarks of Microsoft
Corporation in the United States, other countries, or both.
UNIX is a registered trademark of The Open Group in the United States and other countries.
Linux is a trademark of Linus Torvalds in the United States, other countries, or both.
Other company, product, or service names may be trademarks or service marks of others.
IBM TotalStorage Productivity Center V3.3 continues to build on the function provided in prior
releases. This IBM® Redbooks® publication will take you through what is new in TotalStorage
Productivity Center and explain how to implement and use the new function.
Mathias Defiebre is an IT Specialist for Proof of Concepts in the ATS Customer Solutions
team in Mainz, Germany. His areas of expertise include setup and demonstration of IBM
System Storage™ and TotalStorage solutions in Open Systems environments. He has
worked at IBM for six years after graduating from the University of Cooperative Education
Mannheim in 2001 with a Bachelor of Science and a German Diploma in Information
Technology Management. Mathias is an IBM Certified Specialist - TotalStorage Networking
and Virtualization Architecture.
Curtis Neal is a Senior IT Specialist working for the System Storage Group in San Jose,
California. He has over 25 years of experience in various technical capacities, including
mainframe and open system test, design, and implementation. For the past seven years, he
has led the Open Storage Competency Center, which helps customers and IBM Business
Partners with the planning, demonstration, and integration of IBM System Storage Solutions.
Yuan Xu is a Senior IT Specialist with Advanced Technical Support in China. He has 20 years
of IT experience in software development and open system solution design, implementation,
and support. He joined IBM China in 1999. His areas of expertise include post-sale support
and pre-sale support of IBM storage products, he has extensive experience on
Performance-Benchmark testing and Proof of Concepts demonstration with IBM Enterprise
storage and Mid-range storage, as well as business continuity solution. He holds a degree in
Information Management from the People's University of China.
Bob Haimowitz
Rich Conway
Sangam Racherla
Yvonne Lyon
International Technical Support Organization
Diana Duan
Paul Lee
Tivoli L2 Technical Support
Roland Cao
Harsha Gunatilaka
Thiha Than
Bill Tuminaro
Sarah Hovey
Tivoli Development
Doug Dunham
Tivoli Storage Management Consultant
Chris Katsura
Miki Walter
Tivoli Information Development
Eric Butler
Almaden Research
Silviano Gaona
Brian Steffler
Brocade
Your efforts will help increase product acceptance and customer satisfaction. As a bonus, you
will develop a network of contacts in IBM development labs, and increase your productivity
and marketability.
Find out more about the residency program, browse the residency index, and apply online at:
ibm.com/redbooks/residencies.html
Comments welcome
Your comments are important to us!
We want our books to be as helpful as possible. Send us your comments about this book or
other IBM Redbooks® in one of the following ways:
Use the online Contact us review Redbooks form found at:
ibm.com/redbooks
Send your comments in an e-mail to:
redbooks@us.ibm.com
Mail your comments to:
IBM Corporation, International Technical Support Organization
Dept. HYTD Mail Station P099
2455 South Road
Poughkeepsie, NY 12601-5400
Preface xi
xii TotalStorage Productivity Center V3.3 Update Guide
1
The overall goal of the Topology Viewer is to provide a central location to view a storage
environment, quickly monitor and troubleshoot problems, and gain access to additional tasks
and functions within the TotalStorage Productivity Center User Interface without users losing
their orientation to the environment. This kind of flexibility through the Topology Viewer User
Interface displays better cognitive mapping between the entities within the environment, and
provides data about entities and access to additional tasks and functionality associated with
the current environmental view and the user's role.
A possible scenario utilizing this feature is an application on a host that is running slow. The
system administrator wants to find out what the health of the associated I/O path for that
application is. Are all components along that path healthy? Are there any component level
performance problems that might be causing the slow application response?
The system administrator might also want to find out whether the I/O paths for two
applications (on separate host LUNs) are in conflict with each other. They could be sharing a
common component (such as a switch). After viewing the I/O paths for these two applications,
the administrator makes the required zoning or connectivity change to alleviate the problem.
With TotalStorage Productivity Center V3.3, the reports and alerts are integrated into the GUI
console. It provides a mechanism to navigate from an entity shown on the Topology Viewer to
other areas of the console to show reports or perform management of the selected entity.
This enhancement will enable users to directly jump to the appropriate portions of the console
from the Topology Viewer. Figure 1-3 shows how TotalStorage Productivity Center V3.3 can
provide reports integrated into the console.
In TPC V3.1, the pin list in the Topology Viewer is not persistent across sessions. If you close
the TPC GUI (or even just the Topology Viewer panel), you will lose the current pin list. With
TPC V3.3 the pin list is persistent across sessions and per user. This means that different
users (people using different usernames to log into TPC) will be able to have their own
persistent and private pin lists, which will not be lost when the TPC GUI is closed and which
will not be affected by pin list operations performed by other users.
When using TotalStorage Productivity Center V3.3 with multiple users log onto the same TPC
server simultaneously using the same username, they will share the same pin list. When one
user pins or unpins an entity, the other users will not see that update on their clients right
away. They will see the changes only after they refresh the Overview panel or other affected
views.
Figure 1-4 illustrates right-clicking on a pinned entity in the Topology Viewer Overview panel,
which brings up the context menu for that entity.
The configuration history is displayed through the Topology Viewer-like graphic viewer.
You can identify the changes in this viewer. The change overlay is a topology overlay that
becomes active in Change Rover mode. The purpose of this overlay is to show whether an
entity visible in the topology viewer has been changed between point TimeA and TimeB in
time. In essence there are only four possible states of any entity in this situation:
The entity has not been changed between TimeA and TimeB.
The entity has been created since TimeA.
The entity has been deleted since TimeA.
The entity has been modified since TimeA.
Today, with TotalStorage Productivity Center V3.3, you have the ability to probe Storage
Subsystems and report on their storage. This new capability allows you to create reporting
groups for storage subsystems, and then be able to use these storage subsystem reporting
groups to create TPC reports presenting aggregate views of storage subsystem storage
(based on storage subsystem groups).
It also supports the ability able to use the existing definitions of computer and filesystem
reporting groups, to create TPC reports presenting aggregate views of storage subsystem
storage, in the context of the filesystem and computer groups.
Finally, it provides the user the ability to use the definitions of storage subsystem reporting
groups, to create TPC reports presenting aggregate views of disk capacity in the context of
the storage subsystem groups. This is shown in Figure 1-5.
You can use each planner separately, but you must use all three planners together (Volume
Planner, Path Planner, and Zone Planner) to use storage provisioning.
The Volume Planner helps administrators plan for the provisioning of subsystem storage
based on capacity, storage controller type, number of volumes, volume size, performance
requirements, RAID level, performance utilization, and capacity utilization. This is true if
performance data for the underlying storage subsystem exists. If a prior performance monitor
has not been run for that storage subsystem, the Volume Planner will make its decisions
based on capacity only. The Volume Planner generates a plan that presents the storage
controllers and storage pools that can satisfy the request. If you explicitly specify the storage
pool and controller information, then the Volume Planner checks to see whether the input
performance and capacity requirements can be satisfied.
The Path Planner only has support for IBM SDD multipath drivers.
The goal of the planning framework is to allow users to select high level policies with respect
to performance, availability, and security, and the underlying planning system based on the
current resource utilization, and selected policies determines the best course of action and it
generates a plan.
You have the option of either accepting the advice and then triggering the associated actions
(such as creating a volume or performing zoning), or you can reject the advice and then
perform re-planning.
When you start a planning operation, you have to provide the following types of input:
Users have the option of choosing and specifying values for a set of policies.
Users have the option of selecting hosts, switches, and storage controllers and storage
pools that they want to consider as part of the planning process.
Users can perform storage planning, host planning or security planning in isolation or in a
integrated manner.
Plan Mode: In this mode, the user inputs which storage controller and storage pool to select,
and the wizard determines the number of required paths from the host to the storage
controller, and how the zoning and LUN masking/mapping should be performed.
Storage Planner: In this mode, the Storage Planner provides support for DS8000, DS6000,
and ESS 800 controllers; it takes:
Input with respect to capacity and storage controller preference
Input such as storage controller type, number of volumes, volume size, performance
requirements, use of unassigned volumes, RAID level.
Performance utilization and capacity utilization input from the TPC database; it generates
a plan that presents the storage controllers and storage pools that can satisfy this request.
Figure 1-6 illustrates this strategic feature as you build up a hierarchy of TotalStorage
Productivity Center V3.3 servers. A TotalStorage Productivity Center V3.3 instance is
configured as a master. One or more master TotalStorage Productivity Center V3.3 servers
are configured to collect roll-up information from the subordinates. The master server will
communicate with the subordinate servers using the Device Server API.
A subordinate TotalStorage Productivity Center V3.3 server can have one or more master
servers that are configured to collect roll-up information. The health of the subordinate
servers is based on the result of the check operation and the status from the probe actions.
All TotalStorage Productivity Center V3.3 instances can be configured for discovery and
probing for enterprise reports.
If the TotalStorage Productivity Center for Replication is installed and running on the same
host as the TotalStorage Productivity Center server, an external tool will automatically be
added to the TotalStorage Productivity Center during TotalStorage Productivity Center server
startup. Then you can launch the TotalStorage Productivity Center Replication GUI from the
TotalStorage Productivity Center using an external Web browser. This is similar to launching
an Element Manager, with the difference being that the external tool definitions are stored in
the TPC database and can therefore be shared between TPC users.
VMWare allows a single physical computer system to be divided into logical virtual machines
running various operating systems. To the applications running in the VM, it is a computer
system with a unique IP address and access to storage that is virtualized by the hosting
system (Hypervisor™).
TotalStorage Productivity Center V3.3 will support ESX servers at level 3.0.2 and above. It will
collect hosted VM information. The VMs must be running an operating system in the TPC for
Data agent support list and be running a TPC for Data agent. GSX servers, VMWare servers,
and VMWare workstations are not supported. No support is provided for Virtual VMotion.
TPC will collect Hypervisor information from the system or Virtual Center using the VMWare
Virtual Infrastructure API (proprietary) API of ESX 3 and Virtual Center 2.
SNMP discovery of the topology will be supported. However, McData recommends having the
switches on private networks and therefore SNMP connectivity is generally not possible.
Since the CIM/OM is generally on the public network, the SMI-S communication for fabric
discovery, monitoring, and zone control would be available and is the expected
communication channel.
You can change this path during installation setup. There are two types of installation: typical
and custom.
Typical installation
The Typical installation allows you to install all the components of the TotalStorage
Productivity Center on the local server by selecting the options Servers, Agents, and Clients.
Our recommendation is not to use the Typical installation, because the control of the
installation process is much better when you use the Custom installation method.
Custom installation
The Custom installation allows you to install each component of the TotalStorage Productivity
Center separately and deploy remote Fabric and or Data agents on different computers. This
is the installation method we recommend.
Note: When installing TPC server from Disk1, the installer will prompt you to insert Disk2
to copy files from it. Then it will copy the remote installation files for Data agent and Fabric
agent so that it can perform remote installation for those agents. If you are only installing
local agents, you can use Disk3.
Physical media
The physical media shipped with the TPC V3.3 product consists of a DVD and a CD. The DVD
contains the Disk1 and Disk2 content described in “Passport Advantage and Web media
content” on page 16. The physical media CD is the same as the Web Disk3 media.
http://www-304.ibm.com/jct01004c/systems/support/supportsite.wss/supportresources?
taskind=3&brandind=5000033&familyind=5329731
http://www-304.ibm.com/jct01004c/systems/support/supportsite.wss/supportresources?
taskind=3&brandind=5000033&familyind=5329731
If you install all the components in one step, if any part of the installation fails for any
reason (for example, space or passwords), the installation suspends and rolls back,
uninstalling all the previously installed components.
5. In the next panel, verify that Primary DNS suffix field displays a domain name. Click OK
(see Figure 2-3).
6. If you made any changes, you might need to restart your computer (see Figure 2-4).
Figure 2-4 You must restart the computer for changes to take effect
On the WINS tab, select Enable NetBIOS over TCP/IP and click OK (Figure 2-6).
Before beginning the installation of TotalStorage Productivity Center, you must do one of the
following actions:
Change the IIS port to something other than 80, for example, 8080.
Uninstall IIS.
Disable IIS.
2.4.4 Create Windows user ID to install Device server and Data server
In order to install Device server and Data server, you must have a Windows User ID with all
the proper required rights. We created a unique user ID, as described in Table 2-3 on
page 22.
Table 2-1 points you to the appropriate table that contains the user IDs and passwords used
during the installation of TotalStorage Productivity Center.
Table 2-1 Index to tables describing required user IDs and passwords
Item Table
Table 2-2 on page 22 through Table 2-13 on page 26 contain information about the user IDs
and passwords used during the installation of the TotalStorage Productivity Center
prerequisites and components.
To install a GUI or CLI, you do not need any particular authority or special user ID.
DB2 All Used to run the DB2 Specified when Used by the DB2
administration administration server on DB2 is installed GUI tools to
server user your system. Used by perform
the DB2 GUI tools to administration
perform administration tasks.
tasks. See rules below.
Note: Adding the user ID used to install TotalStorage Productivity Center to the DB2 Admin
group gives the user ID the necessary administrative rights.
Important: Do not change the Agent Registration password under any circumstances.
Changing this password will render the certificates unusable.
Common agent All This is the Specified when you Used during
registration password required install Agent common agent,
by the common Manager Data agent and
agent to register Fabric agent
with the Agent installation
Manager
changeMe changeMe
(is the default)
WebSphere ALL You can use Specified when Used when Device
Application tpcadmin, in order you install Device server has to
Server to avoid to create a server communicate with
WebSphere new one WebSphere
administrator
user ID and Group User ID Password ITSO’s user ID
password and password
Attention: If you update DB2 from an older version, for example, from DB2 7.2 to DB2 8.2,
the TotalStorage Productivity Center installer might not recognize the DB2 version.
4. The next panel is the license agreement, click I accept the terms in the license
agreement (Figure 2-11).
6. Accept the defaults and proceed with Install DB2 Enterprise Server Edition on this
computer (see Figure 2-13). Click Next to continue.
8. Set the user information for the DB2 Administration Server; choose the domain of this
user. If it is a local user, leave the field blank. Click Next to continue.
Type a user name and password of the DB2 user account that you want to create
(Figure 2-15). You can refer to Table 2-5 on page 23.
DB2 creates a user with the following administrative rights:
– Act as a part of an operating system.
– Create a token object.
– Increase quotas.
– Replace a process-level token.
– Log on as a service.
9. Accept the defaults in the panel shown in Figure 2-16, and click Next to continue.
11.In the Configure DB2 instances panel, accept the default and click Next to continue
(see Figure 2-18).
13.In the panel shown in Figure 2-20, click Defer the task until after installation is
complete and then click Next to continue.
18. Click Exit First Steps (Figure 2-26) to complete the installation.
Open a Command Prompt window and enter the db2level command to check the version
installed as shown in Figure 2-27.
Figure 2-28 shows the DB2 window services created at the end of the installation.
In the system tray, there is a green DB2 icon indicating DB2 is up and running (Figure 2-29).
When you install the Agent Manager, you will also be installing the Embedded version of IBM
WebSphere Application Server - Express, V6.0.2 (WebSphere Express).
2. The Installation Wizard starts; you can see a panel similar to the one in Figure 2-30.
7. The Database User Information panel is shown in Figure 2-35. Enter the database user
name and password, this is the DB2 administrator user ID that is in the DB2ADMNS group
and Administrator group.
If you want to use a different user ID for installation of Agent Manager, you can select the
Use a different user ID during the installation, and enter the user ID and password.
Note that if you do not select the check box, the following Database Administrator User ID
and Password will not be used.
We recommend that you use the DB2 ID and password from the DB2 installation if you
use the DB2 only for TPC.
You can refer to Table 2-5 on page 23 for the ID and password. In our installation, we use
the DB2 ID and password from the DB2 installation. Click Next to continue.
8. The WebSphere Application Server Connection Information panel is shown as Table 2-36
on page 45. Enter the following information, and click Next to continue.
– Host Name or Alias of Agent Manager:
Review the preinstallation task mentioned in 2.4.1, “Verify primary domain name
systems” on page 18. Use the fully qualified host name. For example, specify
tpcsrv.itsosj.sanjose.ibm.com. This value is used for the URLs for all Agent Manager
services. It is preferable to use the fully qualified host name rather than an IP address.
If you specify an IP address, you will see the warning panel shown in Figure 2-37 on
page 46.
– Registration Port:
Use the default port of 9511 for the server-side SSL.
– Secure Port:
Use the default port of 9512 for client authentication, two-way SSL.
– Public Port and Alternate Port for the Agent Recovery Service:
Use the public communication port default of 9513.
– Do not use port 80 for the agent recovery service:
Accept the default and do not check this box.
Note: If you want to check this box, make sure that port 80 is not being used by another
application. To check for other applications that are using port 80, run this command:
netstat -an
9. If you specify an IP address instead of a fully qualified host name for the Host Name or
Alias of Agent Manager, you see the panel shown in Figure 2-37. We recommend that you
click the Back button and specify a fully qualified host name.
10.In the WebSphere Application Server Connection Information panel shown in Figure 2-38,
accept the defaults and click Next to continue.
Figure 2-38 WebSphere Application Server Connection Information for Application Server Name
12.In the panel shown in Figure 2-40, specify the Security Certificate settings. To create
Certificates, you must specify a Certificate Authority Password. You must specify this
password to look at the certificate files after they are generated. Make sure that you record
this password in the worksheets in Appendix B, “Worksheets” on page 309.
After entering the passwords, click Next to continue.
At the time of writing this book, entering a password other than “password” will
cause the Data Server and Device Server instal to fail. Check the TPC flashes to
verify if this has been fixed before entering a password other than “password”.
Note: Agent Manager 1.2 installer provided a default password as changeMe. For
Agent Manager 1.3, there is no default password.
You specify a unique password and record it in the worksheets provided in Appendix B,
“Worksheets” on page 309. You must provide a password here, otherwise you cannot
continue the installation.
14.The User Input Summary panel is displayed (see Figure 2-42). If you want to change any
settings, click Back and return to the window where you set the value. If you do not need
to make any changes, click Next to continue.
17.The Summary of Installation and Configuration Results panel is displayed in Figure 2-48.
Verify that the Agent Manager has successfully installed all of its components. Review the
panel and click Next to continue.
18.The last panel (Figure 2-49) shows that the Agent Manager has been successfully
installed. Click Finish to complete the Agent Manager installation.
Verify that the ARS.version field shows the level you have installed (in our case it is 1.3.2.13)
and that at the end you see the message, Health Check passed, as shown in Figure 2-51.
Before starting the installation, verify that DB2 8.2 Enterprise Edition FixPak 14 has been
installed and has been started.
Important: Log on to your system as a local administrator with database authority, for
Windows.
1. For Windows, if Windows autorun is enabled, the installation program should start
automatically. If it does not, open Windows Explorer and go to the TotalStorage
Productivity Center CD–ROM drive or directory. Double-click setup.exe.
2. Chose your language and click OK (see Figure 2-53).
4. Figure 2-55 shows how to select typical or custom installation. You have the following
options:
– Typical installation:
This selection allows you to install all of the components on the same computer by
selecting Servers, Agents, and Clients.
– Custom installation:
This selection allows you to install each component separately.
– Installation licenses:
This selection installs the TotalStorage Productivity Center licenses. The TotalStorage
Productivity Center license is on the CD. You only need to run this option when you add
a license to a TotalStorage Productivity Center package that has already been installed
on your system.
For example, if you have installed TotalStorage Productivity Center for Data package,
the license will be installed automatically when you install the product. If you decide to
later enable TotalStorage Productivity Center for Fabric, run the installer and select
Installation licenses. This option will allow you to install the license key from the CD.
You do not have to install the IBM TotalStorage Productivity Center for Fabric product.
In this chapter, we document Custom Installation. Click Next to continue.
5. In the Custom installation, you can select all the components in the panel shown in
Figure 2-56. This is the recommended installation scenario. In our scenario, we show the
installation in stages. By default all components (except the Remote Data agent and
Remote Fabric agent) are checked. As the first step, we only select the option to Create
database schema. and click Next to proceed (see Figure 2-56).
7. Enter your DB2 user ID and password again (see Table 2-5 on page 23). Do not take the
default of Use Local Database. Click Create local database. By default, a database
named TPCDB is created. Click Schema creation details to continue (Figure 2-58).
.
8. You will see the TPC installation information you selected as shown in Figure 2-60;
click Install to continue.
9. Upon completion, the successfully installed panel is displayed. Click Finish to continue
(Figure 2-62).
Attention: Do not edit or modify anything in DB2 Control Center. This could cause serious
damage to your tablespace. Simply use DB2 Control Center to browse your configuration.
Log files
Check for errors and java exceptions in the log files at the following locations:
<InstallLocation>\TPC.log
<InstallLocation>\log\dbSchema\install
For Windows, the default InstallLocation is c:\Program Files\IBM\TPC
Check for the success message at the end of the log files for successful installation.
Preinstallation tasks
To install Data Server and Device Server components, you must log on to Windows 2003 with
a User ID that has the following rights:
Log on as a service.
Act as part of the operating system.
Adjust memory quotas for a process.
Create a token object.
Debug programs.
Replace a process-level token.
Custom installation
To perform a custom installation, follow these steps:
1. Start the TotalStorage Productivity Center installer.
2. Chose the language to be used for installation.
3. Accept the terms of the License Agreement.
4. Select the Custom Installation.
5. Select the components you want to install. In our scenario, we select the four server
components, as shown in Figure 2-64. Notice that the field, Create database schema,
is grayed out.
Tip: We recommend that you install Data agent and Device agent in a separate step.
If you install all the components at the same time, if one fails for any reason (for
example, space or passwords), the installation suspends and a rollback occurs,
uninstalling all the previously installed components.
6. Click Use local database. We will use the database TPCDB we just created in the
previous step. Click Next to continue (Figure 2-66).
8. In the panel shown in Figure 2-68, enter the Agent Manager information. You must specify
the following information:
– Hostname or IP address:
Fully qualified name or IP address of the agent manager server. For further details
about the fully qualified name, refer to 2.4.1, “Verify primary domain name systems” on
page 18.
– Port (Secured):
Port number of the Agent Manager server. If acceptable (not in use by any other
application), use the default port 9511.
– Port (Public):
The public communication port. If acceptable (not in use by any other application), use
the default of 9513.
– User ID:
This is the user ID used to register the Data Server or Device Server with the Agent
Manager. You have to use the built-in ID manager since it is not allowed to specify it
during the Agent Manager 1.3.2 installation (see Figure 2-41 on page 49).
– Password:
This is the password used to register the Data Server or Device Server with the Agent
Manager. You previously specified this user ID during the Agent Manager install (see
Figure 2-41 on page 49).
– Password - common agent registration password:
This is the password used by the common agent to register with the agent manager, it
was specified when you installed the Agent Manager (see Figure 2-41 on page 49).
Click Next to continue.
9. The Summary information panel is displayed. Review the information, then click Install to
continue (see Figure 2-69).
If you install from the electronic image, the installer will prompt you to change to the
directory of the second disk. Click Browse and choose the directory of Disk2, and click
OK to continue.
Tip: You extract the files from disk1 and disk2 into a directory of the names: disk1 and
disk2. The TotalStorage Productivity Center installation program will be able to find the
disk2 files; it will not prompt the Insert Next Disk panel. Put disk1 and disk2 in the same
directory. For example, on Windows (note that this is case-sensitive), you might do this:
C:\disk1
C:\disk2
or
C:\tpc33\disk1
C:\tpc33\disk2
10.After the GUI and CLI installation messages, you see the summary information panel
(Figure 2-73). Read and verify the information and click Finish to complete the
installation.
Check that the Administrators group contains the newly created TPC user ID. The user ID
TSRMsrv1 is created by default by the install program.
Regarding support and machine requirements, consult the TPC support Web site.
As an alternative, you can use the command line installer. To do this, run db2_install. This will
only install the DB2 data to your operating system. Configuration has to be done manually
then.
Using the graphical installer to install DB2, you have three different options to select the type
of installation you want to make. These options are Typical, Compact, and Custom. We
describe the Custom option to install DB2.
When the DB2 Setup wizard is done loading, you see the DB2 Setup wizard welcome
panel, similar to Figure 3-4. Click Next to continue the installation of IBM DB2 UDB.
7. The features selection panel, similar to Figure 3-8, is displayed. Select the features to be
installed. From the Server support tab, choose the options shown in Figure 3-8. You
might want to deselect DB2 Data Source Support, because it is only needed to connect
to remote databases residing on the System i™ or System z™ server running z/OS or
OS/390®.
8. Select the Client support tab. Make sure that the features shown in Figure 3-9 are
selected. This selection should reflect the default values.
10.Select the Getting started tab. Select the features shown in Figure 3-11. This selection
should reflect the default values. Click Next.
11.The Languages panel enables you to install additional languages. Select the languages
you want to install and click Next.
13.The DAS User panel is now presented. If you would like the installer to create a DB2
Administration Server (DAS) user ID, you must enter a unique username for the DAS user
in the User name field. You must also enter a password in both the Password and the
Confirm password fields.
If you leave the UID and GID fields blank and check the Use default boxes, the system will
assign a UID and GID for you. Alternatively, you can check the Existing user button and
select the name of an existing user ID, which will become the DAS user. When you have
completed this form, click Next as shown in Figure 3-14.
14.The Set up a DB2 instance window opens and appears similar to the one shown in
Figure 3-15. Select Create a DB2 Instance and click Next.
19.You advance to the instance properties panel, similar to Figure 3-20. Select authentication
type Server, check the Autostart the instance at system startup checkbox and click
Next.
20.The DB2 catalog preparation panel is now shown, similar to Figure 3-21. Choose Do not
prepare the DB2 tools catalog on this computer and click Next.
22.You proceed to the health monitoring notification setup, similar to Figure 3-24. If you
disabled the SMTP notification, you are presented a warning message like Figure 3-23.
Click OK to confirm that a setup without SMTP notification is intended.
23.The health monitoring notification setup is presented, similar to Figure 3-24. Select New
contact and specify a Name and an E-mail address for the administration contact for this
instance, or select Defer this task until after installation is complete, and click Next.
24.You get to the summary panel, similar to Figure 3-25. Verify that the summary information
fits your installation needs and click Finish to start the DB2 installation.
26.When the installation is finished, you are presented a completion message similar to
Figure 3-27. Check the Post-installation steps and click Status report to verify installation
success.
28.Congratulations, you have successfully completed the first step of the DB2 installation.
To install the Agent Manager, make sure you have a possibility to work with graphical
installers. Prepare the GUI according to our instructions:
1. Go to your Agent Manager install resource and start the installation by issuing the
following command:
[root@helium EmbeddedInstaller]# ./setupLinux.bin
2. The Agent Manager Installer opens up. The first selection you have to make is whether
you want to install the Agent Manager in combination with an already existing installation
of the WebSphere Application Server, or if you want to use the embedded WebSphere
Application Server, which is delivered with the Agent Manager. We do not have a
WebSphere Apllication Server installed already, so we use the embedded one.
3. Click Next when you are ready to continue the installation. See Figure 3-29.
5. This panel prompts you for the type and location of the database used for the Tivoli Agent
Manager registry. You can choose among the following six options:
– DB2 database on this computer (this is the default)
– DB2 database on another computer (without DB2 Administration Client)
– Local alias to DB2 database on another computer (using DB2 Administration Client)
– Oracle® database on this computer
– Oracle database on another computer (using Oracle database Client)
– Derby database on this computer
We choose DB2 database on this computer, which is the default. It is suitable for the
local installation of DB2 that we have.
20.Finally, the installation prompts you with a summary information that the installation of the
Agent Manager is complete and the Agent Manager has been started. Click Finish to end
the installation of the Agent Manager.
Before you continue with the installation, you have to enable the root user to be able to install
the following TPC components. Therefore the root user needs to be added to the linux group
of the db2 instance owner.
If your DB2 Instance owner is db2inst1, logon as db2inst1 and issue the command groups:
[db2inst1@helium db2inst1]$ groups
db2grp1 dasadm1
You are presented the groups which db2inst1 is member of. We are interested in the group
db2grp1. Now we need to add root to the db2grp1 group.
For Red Hat, in the example given, the following command would add root to the db2grp1.
Logged on as root, execute the following command.
usermod -G root,bin,daemon,sys,adm,disk,wheel,db2grp1 root
After that, log off and logon as root again to see the changes you made. Execute the
command groups as root again:
[root@helium root]# groups
root bin daemon sys adm disk wheel db2grp1
Now you can continue with the installation of the TPC components.
We consider it a best practice to split the installation of TPC into the following three parts:
1. Install the TPC DB2 database schema.
2. Install the Data Server and the Device Server, the GUI, the CLI, the Data agent, and the
Fabric agent.
3. Optional: Install Remote Data agent and / or Remote Fabric agent.
First, install the TPC DB2 database schema. Start the installation of TPC by executing:
[root@helium disk1]# ./setup.sh
The graphical installer will prompt you with a selection window to choose the installer’s
language. Choose English as shown in Figure 3-47 and click OK to continue.
4. The initialization of the installation wizard takes a brief moment. During this time you can
see the panel shown in Figure 3-48.
6. The panel in Figure 3-50 prompts you for the type of installation you want to perform.
Choose Custom Installation to be able to perform the four tier installation approach we
take. Choose an installation location for TPC with enough space left. We use the default
/opt/IBM/TPC. Click Next to continue.
8. This panel allows you to enter the database administrator information that is used to
connect to the database during installation and uninstallation. In our environment, the
database administrator username is db2inst1(see Figure 3-52). Click Next to continue.
10.The installer will present you a summary information. Check it carefully to ensure that it
represents what you want to do as shown in Figure 3-54. If everything is acceptable, click
Next to continue.
11.The installation now begins. The installer gives you progress updates as shown in
Figure 3-55.
12.After the installation is finished, you are presented an installation summary. Make sure that
the installation was successful and click Finish to end the installation as shown in
Figure 3-56.
3.2.4 Installing TPC V3.3 Data Server, Device Server, CLI, and GUI
The next step is to install the remaining TPC components.
1. To continue with the TPC installation, start the installer again by executing the following
command where helium is the name of the server on which we are installing the code.
[root@helium disk1]# ./setup.sh
The graphical installer will prompt you with a selection window to choose the installer’s
language. Choose English as shown in Figure 3-57 and click OK to continue.
2. The initialization of the installation wizard takes a brief moment. During this time, you can
see the following panel shown in Figure 3-58.
4. This panel prompts you for the type of installation you want to perform. Choose Custom
Installation to be able to continue the two-tier installation approach we take. The
installation location for TPC is detected automatically. We use the default /opt/IBM/TPC as
shown in Figure 3-60. Click Next to continue.
6. Enter the database administrator information to be used during the installation in the panel
shown in Figure 3-62 and click Next to continue.
8. The next dialog, as shown in Figure 3-64, lets you specify Data Server, Device Server and
Data agent information. Enter a fully qualified hostname for the Data Server and the
Device Server. This hostname should be resolvable by your DNS system from all
machines you plan to use in combination with TPC. Specify a Data Server and a Device
Server port. Specify the TPC superuser.
Specify a password for the Fabric agents to communicate with the Device Server, the host
communication password. You will not be able to specify a Data Server account password.
This is for Windows installations only. Specify a WebSphere Application Server Admin ID
and a password. Click Security roles... to advance to Figure 3-64, click NAS discovery
to advance to Figure 3-66 on page 123, or click Data agent options to advance to
Figure 3-67 on page 123. Click Next to continue with the installation.
You have the possibility to create advanced security roles mapping. Stay with the defaults
and click OK as shown in Figure 3-65 to get back.
Configure the default options for Data agents in this dialog. Enable the checkbox to have
an agent run a scan when first installed. Enable the checkbox to allow an agent to run a
script sent by the server. Click OK to continue.
10.The next panel, as shown in Figure 3-69, prompts you with the common agent selection.
Click Install a new common agent at the location listed below. The default location is
/opt/IBM/TPC/ca with an agent port of 9510. Click Next to continue.
12.After the installation is finished, you are presented with a summary panel as shown in
Figure 3-71. Click Finish to end the installation.
Congratulations, you have successfully installed TPC. You can now start the TPC GUI by
issuing the following command:
[root@helium root]# /usr/local/bin/TPC
In this chapter we guide you through a TotalStorage Productivity Center V3.3 upgrade and
describe the requirements necessary to accomplish this goal. We also discuss the
preparation steps required before you do an upgrade. We discuss the following topics:
IBM upgrade strategy
Upgrade technical considerations
Step-by-step upgrade process
TPC for Data and TPC for Fabric share a Common Agent to manage the fabric as well as
capacity utilization of file systems and databases. Figure 4-1 shows the TotalStorage
Productivity Center physical structure.
The Data server is the control point for product scheduling functions, configuration, event
information, reporting, and GUI support. It coordinates communication with agents and data
collection from agents that scan file systems and databases to gather storage demographics
and populate the database with results.
Automated actions can be defined to perform file system extension, data deletion, and Tivoli
Storage Manager backup or archiving or event reporting when defined thresholds are
encountered.
The single database instance serves as the repository for all TotalStorage Productivity Center
components. The Data agents and Fabric agents gather host, application, and SAN fabric
information and send this information to the Data server or Device server. The GUI allows you
to enter information or receive information for all TotalStorage Productivity Center
components.
You can upgrade previous releases of TotalStorage Productivity Center to V3.3. The
TotalStorage Productivity Center architecture flow is illustrated in Figure 4-1. You can upgrade
the following IBM TotalStorage Productivity Center releases to IBM TotalStorage Productivity
Center version 3.3:
IBM TotalStorage Productivity Center for Data version 2.3 to 3.3
IBM TotalStorage Productivity Center version 3.1.1, 3.1.2, or 3.1.3 to 3.3
IBM TotalStorage Productivity Center version 3.2 or 3.2.1 to 3.3
If you have TPC V2.1 or V2.2, you must first upgrade to V2.3, then upgrade to IBM
TotalStorage Productivity Center V3.3. For information about upgrading IBM TotalStorage
Productivity Center for Data from V2.1 or 2.2 to 2.3, see the IBM TotalStorage Productivity
Center for Data Installation and Configuration Guide for V2.3, GC32-1727, found at:
http://publib.boulder.ibm.com/infocenter/tivihelp/v4r1/index.jsp
Note: The Migration Utility allows you to migrate data from IBM TotalStorage Productivity
Center V2.3 to V3.3. You can migrate Disk, Performance, Fabric or TotalStorage Expert
data. You do not have to migrate data from IBM TotalStorage Productivity Center for Data
V2.3 because that data is migrated through the normal upgrade process to IBM
TotalStorage Productivity Center V3.3.
4. Double-click setup.exe from the temporary working directory you created in step 1 to start
the upgrade.
5. Click Run to proceed as shown in Figure 4-4.
6. You will see the first panel, as shown in Figure 4-5. Select Install Product to proceed.
The DB2 installation proceeds, and you see a progress panel similar to the one shown in
Figure 4-9.
10.The installer reminds you to reboot your system after you upgraded your DB2, as shown in
Figure 4-11. Click Yes to reboot your server.
2. Open a Command Prompt window and enter the db2level command to check the DB2
version installed as shown in Figure 4-13.
2. It is considered a good idea to install the most current IBM DB2 FixPak. For TPC V3.3 we
install the DB2 FixPak 11. Download it from the IBM Web page. Navigate through the path
Software → Information Management → DB2 Product Family → DB2 UDB Version 8
FixPaks and Clients Install
3. Prepare your DB2 for upgrade during FixPak download. Shut down DB2 by entering these
commands:
su - db2inst1
. $HOME/sqllib/db2profile
db2 force applications all
db2 terminate
db2stop
db2licd -end
$HOME/sqllib/bin/ipclean
exit
4. Continue with the shutdown of the Administration server by entering these commands as
shown in Figure 4-16:
su - db2tpc
.$HOME/das/dasprofile
db2admin stop
exit
/opt/IBM/db2/V8.1/bin/db2fmcu -d
/opt/IBM/db2/V8.1/bin/db2fm -i db2tpc -D
6. Enter the FixPak directory and start the installation of the Fixpak by entering this
command as shown in Figure 4-17.
installFixPak -y
7. After successfully installing the FixPak, update the DB2 database instances. As a
preparation to the instance update, you can list the database instances by entering this
command:
/opt/IBM/db2/V8.1/instance/db2ilist
9. List the DB2 database administration servers by entering this command as shown in
Figure 4-20 on page 139.
/opt/IBM/db2/V8.1/instance/daslist
11.Start up DB2 again by entering the following commands (see Figure 4-22).
su - db2inst1
. $HOME/sqllib/db2profile
db2start
exit
su - db2tpc
.$HOME/das/dasprofile
db2admin start
exit
Attention: A new version of Agent Manager (V1.3.2) is available for use with TPC V3.3.
It is mandatory that you upgrade to V1.3.2, because the embedded WebSphere shipped
with Agent Manager V1.2 is no longer supported.
8. A series of installing panels for the Agent Manager are displayed indicating the progress of
the installation (see Figure 4-28). Wait for the installation to complete.
10.A summary of the Installation and Configuration Results panel is displayed (see
Figure 4-30). This panel indicates if the Agent Manager has successfully validated and
installed all of its components. Review the information and click Next to continue.
11.The summary information panel is displayed. For a successful upgrade, you will see the
panel shown in Figure 4-31. Click Finish to continue.
The following panels take you through an upgrade from TPC V3.1 to TPC V3.3 on a Windows
2003 server.
1. Start the IBM TotalStorage Productivity Center installation program.
2. The Select a language panel is displayed. Select a language and click OK as shown in
Figure 4-33, “Installer language selection” on page 148.
3. The Software License Agreement panel is displayed as shown in Figure 4-34. Read the
terms of the license agreement. If you agree with the terms of the license agreement,
select I accept the terms of the license agreement. Click Next to continue.
4. The default selection is a Typical Installation illustrated in the next panel. It is important
that you select Custom installation (see Figure 4-35). Then click Next to continue.
6. The Database administrator information panel is displayed with information filled in by the
installation program (Figure 4-37). Click Next.
8. The Data server, Device server, Data agent, and Agent information panel is displayed.
Enter the Host authentication password, which is used when you install Fabric agent, to
communicate with the Device server. Also Enter the WebSphere Application Server (WAS)
administration ID and Passwood. All fields are case sensitive. Click Next to continue.
9. The Summary information panel is displayed (see Figure 4-40). The items listed will be the
components upgraded. Verify the list of components and click Next to continue.
11.If you are installing using the Web images, a panel will be displayed to insert Disk2. Enter
the directory path where Disk2 is extracted (see Figure 4-42 on page 153). When Disk 2 is
located, click Next to continue.
Unlike other Data agents, there is no agent code to install on NAS devices, TPC issues
SNMP queries to the NAS device to collect the summary information for the aspect of the
NAS device, and TPC also utilizes a proxy agent implementation to collect more detailed
information. A Data agent is designated as the “proxy agent” responsible for collecting
asset/quota information from assigned NAS devices via SNMP. TPC collects the mounted file
system information or shares via NFS or CIFS viewed by the system hosting the Data agent.
After collecting the NAS device information, the NAS devices are displayed in the Topology
Viewer as computers. You can check information about them just as you would for computers.
The collected information is also used for the following reports, which can be found in the
Navigation Tree under Reporting → Asset → By OS Type → Network Appliance.
Controllers
Disks
File System or Logical Volumes
Exports or Shares
Monitored Directories
Check the following Web site for the IBM N Series Models and NetApp filers supported by
TPC V3.3:
http://www-1.ibm.com/support/docview.wss?rs=597&uid=ssg1S1003019
The account used for scanning NAS for Windows must be a Domain account that can log into
both the Windows agent machine and the NAS device.
Note: The Windows server used as a proxy Data agent must be a member of the Windows
domain, and the NAS filer should also be added to the Windows domain.
Note: You must use a Domain User that has Domain Administrator privileges.
3. Expand the Administrative Services in the Navigation Tree and select Administrative
Services → Discovery.
Note: You must verify that the correct SNMP community name is defined in the
Windows Domain, NAS and SAN FS job. To verify, go to Administrative Services →
Discovery → Windows Domain, NAS and SAN FGS and select the Options panel.
Add the correct SNMP community name for the filer.
To run a NAS/NetApp discovery job, right-click Windows Domain, NAS and SAN FS and
select Run Now as shown in Figure 5-7.
Note: The discovery will discover all entities in the Windows Domain. To shorten the
time taken for discovery, you can put a check mark on Skip Workstations in the
Discovery properties panel (see Figure 5-5).
To check the running job status, right-click Windows Domain, NAS and SAN FS and
select Update Job Status. Wait until the discovery job finishes.
4. After the discovery job completes, but before TotalStorage Productivity Center can perform
operations such as Probe and Scan against NetWare/NAS/NetApp filers, first they must
be licensed. In a very large environment, you might not want to automatically license all
the discovered NAS devices, so you have a choice as to which servers to license.
Note: If you manually add a NAS device, you do not need to do this, the filer will be
licensed automatically.
Important: You must save the changes after you license the NAS filer before you leave this
panel, otherwise TPC will not save the change to its repository.
Note: Set Filer Login/Password is not required if it is a UNIX attached NAS filer.
Note: Keep in mind that you are assigning work to computers, and that these scans go
over the IP network, so it is better to pick a proxy agent on the same LAN. These scans will
be slower than a local scan, and it might take a very long time to run if the scan is also
doing a Proxy scan on a large device. For large NAS devices, use multiple proxy agents.
Normally the default scan of once per day is really not required, so look into weekly scans.
Now you can see that the NAS filer filesystems have been assigned the Scan/Probe agent.
Make sure you click the Save button in the toolbar to save the changes (see Figure 5-15).
Click the When to Run tab, and choose the radio button of Run Now, and save the Probe job
by clicking the Save button on the toolbar (see Figure 5-17). The Probe jobs starts. You can
right-click on the Probe job and select Update Job Status to check the running job status.
When the Probe job has been successfully completed, you can verify that TPC has the filer
data by viewing the TPC dashboard. On the Monitored Server Summary Panel, you should
see the total number of Network Appliance devices monitored and the total Filesystem/Disk
Capacities (see Figure 5-18).
Click the When to Run tab, and choose the radio button of Run Now. Save the Scan job by
clicking the Save button on the toolbar (see Figure 5-21). You will be prompted for a name for
the Scan job, enter a job name and click OK to start the Scan job. You can right-click the job
name and select Update Job Status to check the running job status.
Important: You must issue a save after you license the NAS filer, but before you leave this
panel. Otherwise TPC will not save the change to its repository.
You do not need to enter the Filer Login tab, because it is only required for a Windows
environment.
4. Run a discovery again.
We need to run a discovery job again to get further information on the NAS filer. Refer to
Figure 5-7 on page 165.
If all the information you provided above is correct, you will see the NAS filer added to the
panel shown in Figure 5-26.
Note: Keep in mind that you are assigning work to computers, these scans go over the IP
network, so it is better to pick a proxy agent on the same LAN. These scans will be slower
than a local scan, and it might take a very long time to run if the scan is also doing a Proxy
scan on a large device. For large NAS devices, use multiple proxy agents. Normally the
default scan of once per day is really not required, so look into weekly scans.
When a probe has been successfully completed, you can verify that TPC has the filer data by
viewing the TPC dashboard. On the Monitored Server Summary Panel, you should see the
total number of Network Appliance devices monitored and the total Filesystem/Disk
Capacities (see Figure 5-18).
Figure 5-30 TotalStorage Productivity Center Dashboard showing NAS Filer Information
In the Profiles tab, we select all of the default profiles and apply them to file systems and
directories by clicking the >> button as shown in Figure 5-20. Profiles allow us to specify what
statistical information is gathered and to fine-tune and control what files are scanned.
Click When to Run tab, and choose the radio button of Run Now, and save the Scan job by
click the Save button on the toolbar (see Figure 5-32). TPC will prompt for a name for the
Scan job. Enter a job name and click OK to start the Scan job. You can right-click the job
name and select Update Job Status to check the running job status.
In this section we will show you some examples of how to retrieve and display data for NAS
filers.
2. From the following panel, highlight the NAS filer and click the magnifier icon in front of it
(see Figure 5-37).
3. in the next panel, highlight all the mount points you are interested in under the Mount
Point column you are interested in, and right-click on them. Choose Chart space usage
for selected as shown in Figure 5-38.
The Filesystem free space chart is presented in the next panel. This chart shows the current
free space on each volume on the NAS filer. You can right-click the chart and select the
Customize this chart button to customize the chart. On the pop up panel, we choose 4 in the
Maximum number of charts or series per screen drop-down menu (see Figure 5-39).
For further information regarding NetApp Quotas, refer to the latest version of the
IBM TotalStorage Productivity Center Version 3.3 User’s Guide, GC32-1775.
The TotalStorage Productivity Center uses SLP as the method for CIM clients to locate
managed objects. When a CIM agent implementation is available for a supported device, the
device can be accessed and configured by management applications using industry-standard
XML-over-HTTP transactions.
In order to manage Brocade Mi10k fabric using SMI-Specifications V1.1.0, we require the
software component mentioned on this site:
http://www.snia.org/ctp/conformingproviders/mcdata
At the time of writing this book, this is the McDATA OPENconnectors SMI-S Interface.
The McDATA OPENconnectors SMI-S Interface is a software product that provides support
for the SMI-S for Brocade director and switch products. McDATA OPENconnectors SMI-S
Interface provides an SMI-S interface for management of Brocade products. It exposes a
WBEM (specifically, CIM XML) interface for management.
Prerequisites
The McDATA OPENconnectors SMI-S Interface requires the hardware and software
described here.
Figure 6-1 lists the switch hardware supported for this release of McDATA OPENconnectors
SMI-S Interface and the minimum levels of firmware required by the hardware.
After logging into the McData File Center, the McData OPEN connectors SMI-S Interface can
be found by selecting DOCUMENTS from the top toolbar. In the Find documents where box,
check the box Category is one or more of the following and select SMI-S from the scroll
down on the right. Select SMI-S 1.5.2 with the description Brocade OPENconnectors SMI-S
interface version 1.5.2 for Windows. This is the same as the McDATA OPENconnectors
SMI-S Interface Version 1.5.1.
The McData OPENconnectors SMI-S Interface User Guide can be downloaded from the
Resource Library.
http://www.mcdata.com/wwwapp/resourcelibrary/jsp/navpages/index.jsp?mcdata_categor
y=www_resource&resource=index
The installer will launch, displaying the License Agreement panel shown in Figure 6-3.
Select I Accept the terms of the lIcense Agreement and click Next.
The installer gives you the option of installing Service Location Profile (SLP) software. If SLP
is not installed, the post-install configuration will fail.
To install both components select SMI-S Interface and SLP as shown in Figure 6-5 and then
click Next.
The next panel is the Pre-installation Summary panel (Figure 6-7) that allows you to start the
installation. Click Install to continue.
After the McDATA OPENconnectors SMI-S Interface is installed, the installer launches the
SMI-S interface server configuration tool, as detailed in the next section.
When you click on the Test Connection, you are presented the progress panel shown in
Figure 6-10.
If you are not at the required EFCM 9.2.0 level, you receive the error message in Figure 6-11,
and the configuration will end.
If you are at the required EFCM 9.2.0 level, when the connection is successfully tested, you
see the message shown in Figure 6-12. Click OK to continue.
You are now ready to work with the McData SMI-S provider through TotalStorage Productivity
Center.
You are prompted for the following data to be provided (see Figure 6-14):
Host: Enter the IP address or the hostname of the SMI-S Interface Server.
Port: By default, you can use port 5989 used by HTTPS.
Username: The default userid you can use is Administrator.
Password: The password for Administrator is password.
Interoperability namespace: The namespace is /interop.
Protocol: Use either HTTP or HTTPS depending on your setting. We recommend HTTPS.
Truststore Location: Leave blank.
Truststore Passphrase: Leave blank.
Display Name: You can enter any name you want to identify the CIM Agent.
Description: You can enter a description for the CIM Agent.
Test CIMOM connectivity: We suggest that you leave the check mark in the check box
and let TPC perform the connection test.
Note: We recommend that you do not add an out-of-band Fabric agent for the Mi10k. This
will cause invalid information to be added to the TPC repository.
To view information regarding the switches, generate the SAN Assets (Switches) report under
IBM TotalStorage Productivity Center → My Reports → System Reports → Fabric
(see Figure 6-17).
Important: In order to see connection information for the Mi10k, an in-band Fabric agent is
required on the fabric.
Figure 6-18 Topology Viewer showing the Brocade Mi10k switch and connected entities
When the Switch Performance Monitor Job has completed, you are able to view the switch
performance data in TotalStorage Productivity Center. To view the port performance data,
go to Fabric Manager → Reporting → Switch Performance → By Port and generate the
report. The data is shown in tabular form (see Figure 6-20).
To view the switch performance data in graphical mode, you can click on the pie chart icon on
the upper left side and select the desired chart (see Figure 6-21).
If you have TPC for Replication already installed and running on the same host as
TotalStorage Productivity Center, an external tool will be automatically discovered and
available under IBM TotalStorage Productivity Center → External Tools (see Figure 7-1).
Figure 7-1 TPC for Replication External Tool Automatically Discovered by TPC
Figure 7-2 Add IBM TotalStorage Productivity Center for Replication GUI
You are prompted for the following data to be provided (see Figure 7-3):
URL/Command: Enter the IP address or the hostname of the TPC for Replication Server,
along with the port number and location.
Label: You can define a label to easily identify the external tool.
Description: You can enter a description for the external tool.
Figure 7-3 Add TotalStorage Productivity Center for Replication GUI External Tool to TPC
When you have the storage subsystem probed, you will see the FlashCopy Target Capacity
value on the TotalStorage Productivity Center dashboard populated with the total FlashCopy
Target Capacity of all storage subsystems probed (see Figure 7-5).
Note: The usable LUN capacity now excludes the capacity of volumes that are defined as
FlashCopy targets.
Figure 7-7 View FlashCopy Attributes for Individual Volume or Virtual Disk
The IBM TS3310 has an embedded SMI-S agent, so no external agent setup is required.
The TotalStorage Productivity Center uses SLP as the method for CIM clients to locate
managed objects. When a CIM agent implementation is available for a supported device, the
device can be accessed and configured by management applications using industry-standard
XML-over-HTTP transactions.
You are now ready to work with the TS3310 SMI-S provider through the TotalStorage
Productivity Center.
When the probe has been completed, we can generate reports and view relevant information
about the TS3310 Tape Library.
When the SMI-S Provider has been configured on the TPC Server, we can generate relevant
reports.
To view information regarding the tape libraries in your environment, view the Tape Libraries
panel under Tape Manager (see Figure 8-4).
Figure 8-6 Topology View showing the IBM TS3310 Tape Library
The overall goal of the Topology Viewer is to provide a central location to view a storage
environment, quickly monitor and troubleshoot problems, and gain access to additional tasks
and functions within the TotalStorage Productivity Center graphical user interface (GUI)
without users losing their orientation to the environment. This kind of flexibility through the
Topology Viewer UI displays better cognitive mapping between the entities within the
environment, and provides data about entities and access to additional tasks and functionality
associated with the current environmental view and the user's role.
The Topology Viewer uses the TotalStorage Productivity Center database as the central
repository for all data it displays. It actually reads the data in user-defined intervals from the
database and, if necessary, updates the displayed information automatically.
The Topology Viewer is an easy-to-use and powerful tool within TotalStorage Productivity
Center. It can make a storage manager’s life easier. But as is true for every tool, to get the
most out of it, you first have to understand the basics, the concept, and the “dos and don’ts.”
In TPC V3.1, the pin list in the Topology Viewer is not persistent across sessions. If you close
the TPC GUI (or even just the Topology Viewer panel), you will lose the current pin list. Now,
with TPC V3.3, the pin list is persistent across sessions and per user.
This enhancement makes pin lists for longer durations. The pins will remain visible until you
unpin them in the Topology Viewer. So it can be used to easily refer to a small number of
entities for your whole monitoring period of time.
To pin an entity in the graphical view, open the Background Context Menu by selecting an
entry and right-click, then select the Pin or Unpin menu option (see Figure 9-1).
Figure 9-2 shows the entities we pinned in the Topology Viewer Overview panel. Pinning
provides a direct path to look at an entity in detail by just double clicking on it. Note that in a
collapsed group, pinned entities are surfaced (see Figure 9-2), whereas in an expanded
group, pinning is shown on each pinned entity.
With TotalStorage Productivity Center V3.3, the reports and alerts are integrated into the GUI
console. It provides a mechanism to navigate from an entity shown on the Topology Viewer to
other areas of the console to show reports or perform management of the selected entity.
This enhancement enables users to directly jump to the appropriate portions of the console
Navigation Tree from the Topology Viewer.
Link to alerts
From the Topology Viewer, you might notice that there is a red exclamation point (!) shown
beside an entity or in the group title indicating there are alerts for this entity or some of the
entities in the same group. You can hover to the red exclamation point; it will show you how
many alerts are now open for the selected entities.
You would right-click on the entity and choose Show <entity type> Alerts (in our example,
we are looking at Switches Group, so it is Show Switch Alerts) as shown in Figure 9-4.
Link to reports
In the Topology View, from the L0-L2 graphic view of Computers and Storage classes, you
can link to reports that reflect the entities you choose from. This is very convenient when you
are looking at an entity or a certain group of entities, and you want to get a report showing
more information.
Here is an example of how to create a report from the Computer group graphic view.
Right-click on the title of the Computers group, then click Reports from the pop-up context
menu as shown in Figure 9-6.
Note: Do not right-click on the computer entity, if you do so, the report will be related only
to that computer.
The pop-up panel shows the possible reports that can be generated related to the selected
entity. In our example, we choose Data Manager → Capacity → Filesystem Used Space →
By Computer, and click Create Report (see Figure 9-7).
You will see the report shown in Figure 9-9 that contains filesystem used space information
for the selected computer group in Topology View.
To launch the Data Path Explorer, in Topology Viewer, select one or more hosts, subsystems,
disks, volumes, or Mdisks, and click the Data Path Explorer shortcut in the MiniMap tool
(Figure 9-10).
Note: In order to view the Data Path Explorer for a host, you must either have an in-band
Fabric agent installed or both a Data agent and an out-of-band Fabric agent on the SAN on
which the host is connected.
The Data Path Explorer view for hosts is shown in Figure 9-11. From this view, you can see
how the host volumes connect to the storage subsystem through the fabric.
Data Path Explorer view for subsystem is shown in Figure 9-12. if the subsystem is an SVC,
from this view, you can see how the SVC Mdisks connect to its back-end storage subsystem
through the fabric. If the subsystem is not an SVC, from this view, you can see how the
subsystem connects to its hosts.
The Data Path Explorer view for volume is shown in Figure 9-14. From this view, you can see
how the volume from the subsystem connects to SVC or hosts.
Change Monitoring is the detection of changes in the environment over time and the
identification of the changes to the user.
The configuration history is displayed through the Topology Viewer, like a graphic viewer.
You can identify the changes in this viewer. The change overlay is a topology overlay that
becomes active in Change Rover mode. The purpose of this overlay is to show whether an
entity visible in the Topology Viewer has been changed between point TimeA and TimeB in
time. In essence, there are only four possible states of any entity in this situation:
The entity has not been changed between TimeA and TimeB.
The entity has been created since TimeA.
The entity has been deleted since TimeA.
The entity has been modified since TimeA.
The left Time Range slider covers the range of time from the oldest snapshot in the system to
the current time. The right Snapshots in Range slider allows you to select any two snapshots
from the time interval specified by the Time Range slider. The value in parentheses beside
the Snapshots in Range slider indicates the total snapshots in the currently selected time
range. The Snapshots in Range slider has one check mark for each snapshot from the time
interval that you specified in the Time Range slider.
The Configuration History page (a variation of the Topology Viewer, see Figure 9-17) displays
the configuration’s entities. To distinguish them from tabs in the Topology Viewer page, tabs in
the Configuration History page (Overview, Computers, Fabrics, Storage, and Other) have a
light gray background and are outlined in orange. The minimap in the Configuration History
page uses the following colors to indicate the aggregated change status of groups:
Blue: One or more entities in the group have changed. Note that the addition or removal of
an entity is considered a change.
Gray: All of the entities in the group are unchanged.
Note: In the Configuration History view, the performance and alert overlays are disabled
and the minimap's shortcut to the Data Path Explorer is not available.
Table 9-1 describes the colors and icons of the change overlay.
Light gray entities without an Entities that did not exist at the time that the snapshot was
overlay icon taken or at the time that a later snapshot was taken.
Background: Light Gray
Entities that changed between the time that the snapshot was
Yellow pencil taken and the time that a later snapshot was taken.
Background: Blue /turquoise
Entities that did not change between the time that the
Light gray circle snapshot was taken and the time that a later snapshot was
Background: Dray Gray taken.
Entities that were created between the time that the snapshot
Green cross was taken and the time that a later snapshot was taken.
Background: Green
Entities that were deleted between the time that the snapshot
Red minus sign was taken and the time that a later snapshot was taken.
Background: Red
A policy is a rule that is applied to the storage environment, for example: “All the HBAs from
the same vendor and model type should have the same firmware level.” You are notified of
violations of these policies.
You can specify the configuration check for all fabrics, one fabric, or one zoneset. This
provides you some extend of flexibility when you just want to check certain scope of the fabric.
The configuration check job can be run as needed or scheduled just like other probe or scan
jobs.
You are notified of violations of these policies through Alerts and logfiles. The entities violating
these SAN Configuration policies will be flagged and displayed in the Topology Viewer.
Table 9-2 lists and explains the best practice policies in TPC V3.3.
2. Each HBA accesses storage Determines whether an HBA accesses both storage subsystem
subsystem ports or tape ports, and tape ports. Because HBA buffer management is configured
but not both. differently for storage subsystems and tape, it is not desirable to
use the same HBA for both disk and tape traffic. The Fabric and
Zone Set scopes are not supported by this policy because an
HBA can be connected to multiple fabrics.
3. Each volume is accessed Determines whether computers that run different operating
only by computers running the systems access the same storage volumes. Use of the same
same type and version of volumes by computers that run different operating systems can
operating system. corrupt the data that is stored on the volumes. This applies,
regardless of whether the computers are in the same zone.
4. Each zone contains only Determines whether HBAs from different vendor types are in their
HBAs from a single vendor. own zone. Receiving a registered state change notification
(RSCN) can cause an HBA to lose a zoned device, preventing the
HBA from seeing or communicating with other devices in the
zone. To avoid losing a zoned device, keep HBAs from different
vendor types in their own zone.
5. Each zone contains only a Determines whether different storage subsystems are in the
single model of storage same zone. While no technical problem is associated with
subsystem. storage subsystems from different vendors and of different
models being in the same zone, an administrator might find them
more difficult to organize. When similar storage systems are in
the same zone, an administrator can easily group them for
different applications.
6. Each zone is part of a zone Determines the presence of orphan zones. Orphan zones are not
set. associated with any zone set. They are not useful because their
definitions are not used and they take up switch resources.
7. Each host must be zoned so Determines whether the zones that were configured by the
that it can access all of its storage administrator allow each computer to access all of the
assigned volumes. storage volumes that are assigned to it. The administrator
specifies the storage subsystem ports through which the
computer port accesses volumes, but might forget to configure
zones that enable the ports to communicate during volume
assignment. This policy also determines whether zoning makes
assigned volumes inaccessible to the computer ports. The Fabric
scope is not supported by this policy.
8. Each computer has only Checks whether there is only one type of HBA in each computer.
HBAs of the same model and Using only one type of HBA minimizes configuration problems.
firmware version. The policy also checks whether firmware upgrades have been
done properly for all HBAs in a computer. The Zone Set scope is
not supported by this policy.
9. For each host type and Determines whether all firmware upgrades have been done for
operating system, every HBA of the HBAs in the operating system. The Zone Set scope is not
a given model must have the supported by this policy.
same firmware version.
10. Every SAN switch of a given Determines whether firmware upgrades have been done for all
model must have the same switches of the same type. For example, if you have four identical
firmware version. models of SAN switches from the same vendor and you perform
a firmware upgrade on one, it is best to perform the upgrade on
all of the others. The Zone set scope is not supported by this
policy.
11. Every storage subsystem of Determines whether firmware upgrades have been done for all
a given model must have the storage subsystems of the same type. For example, if you have
same firmware version. four identical storage subsystems from the same vendor and you
perform a firmware upgrade on one, it is best to perform the
upgrade on all of the others. The Zone Set scope is not supported
by this policy.
12. Each fabric can have a Checks whether the number of zone definitions in the fabric is
maximum of x zones. larger than the number that you entered. In large fabrics, too
large a number of zone definitions can become a problem. Fabric
zone definitions are controlled by one of the switches in that
fabric, and limiting their number ensures that the switch's zoning
tables do not run out of space. The Zone Set scope is not
supported by this policy. You can enter up to a maximum of 9999
zones.
13. Each zone can have a Checks whether the number of zone members in a zone is larger
maximum of x zone members. than the number that you entered. In large fabrics, too large a
number of zone members can become a problem. Fabric zone
members are controlled by one of the switches in that fabric, and
limiting their number ensures that the switch’s zoning tables do
not run out of space. You can enter up to a maximum of 9999
zones.
3. Scroll down to Scheduling to specify how often to run the analysis, in our case, we choose
Run Now, and click File Save, and give the analysis job a name in the pop-up box (see
Figure 9-20) to start the job.
The Job log file page displays the following information (see Figure 9-22):
– Scope of the run (all fabrics, one fabric, or one zone set)
– Policies that were checked
– Total violations
– Breakdown of violations per policy
– Whether the run completed successfully or the errors it encountered
Note: Tabular view Alert tab from Overview shows all violations, if you choose
Computers, Fabrics, Storage, or Other, the Alert tab only shows alerts related to that
specified device group.
As an example, we have Policy 6 violated (each zone is part of a zone set); check the affected
entities. We found in Fabric (100000051E35D514), zone (COLORADO_HBA_ITSOSVC02)
and zone (GALLIUM_HBA_ITSOSVC02) are defined but are not part of a zone set, and in
Fabric (100000051E34E895), zone (COLORADO_HBA_ITSOSVC02) is defined but is not
part of a zone set. To fix the violation, we can add the zones to the zoneset in each of the
fabrics, or if the zones are not needed, we can delete the zones from the fabrics, and run the
configuration analysis again.
We also have Policy 10 violated (every SAN switch of a given model must have the same
firmware version); check the switch configuration. We found that the two switches in our lab
are at different firmware level s (see Figure 9-25). To fix the violation, we can upgrade the
firmware to the same level, and run the configuration analysis again.
In this chapter, we will explain the SAN Planner functionality and go through a step-by-step
creation and implementation of a plan.
In TotalStorage Productivity Center V3.3, if multiple storage pools from different controllers
can potentially satisfy your provisioning request, the Volume Planner uses rated utilization of
the pools (the sum of the previous provisioning performance requirements, which might be
greater than the current utilization) to break the ties and select a candidate storage pool.
Note: Before using the Volume Planner to allocate storage based on performance
characteristics, a Performance Monitor job must be run on the target subsystem.
Note: IBM Subsystem Device Driver (SDD) must be installed on a host in order to invoke
the Path Planner.
The Path Planner is used for specifying multiple paths options between selected hosts and
storage subsystems. Path Planner assists the administrator in the multipath tuning process
through the selection of these policies:
The Multipath option specifies how the driver uses the paths between the host and the
storage subsystem. The options are as follows:
– Load Balancing sends Input/Output on all paths
– Round Robin sends Input/Output on one path until a time interval expires (set in an
SDD setting at the host) or stops to use another path.
– Fail-Over sends Input/Output on one path until a failure occurs and fails over (switches)
to another path.
The Specify number of paths option specifies the number of paths between each host and
the storage subsystem.
Workload profiles provide the Path Planner with estimates of required I/O traffic. These
estimates are used to determine the number of paths required from each host to the storage
subsystem and select the multipath driver path utilization. The number of paths and the driver
multipath mode can be adjusted by the performance requirements specified through the
selection of a workload profile.
The Path planner does not directly interact with the Zone planner. It provides the path
information which the Zone planner uses. Each path is represented by a host port WWPN,
target port WWPN, and a Volume ID for the volume on the target which is mapped to the host
port. These paths are created when the Zone planner is implemented.
The Zone Planner expects a list of host port and storage port pairs as input. If the Path
Planner has been invoked prior to the Zone Planner, its output is used as input to the Zone
Planner. If the subsystem/host are within the same fabric and Zone Planner is not checked,
then existing zones or zone sets are used. If Zone Planner is checked, this creates a new
zone or zone set.
Note: For the Zone Planner to create zones, the host and subsystem must be within the
same fabric.
For the case where the host and subsystem reside in more than one of the same fabrics, you
are given two options. The first option is to create identical zones in all of the fabrics. The
second option is to select specific fabrics to create identical zones in. The guidance policies
used for zone planning are as follows:
One zone per host bus adapter (HBA)
One zone per host
One zone per host port
Auto zone: largest zone that satisfies the validation policies
The following validation policies are used for zone planning (where N is specified by the user):
No two controllers of different types should be in the same zone.
Maximum number of zone members in a zone = N.
Maximum number of zones in a fabric = N.
Note: SAN Planner supports the ESS, DS6000, and DS8000 subsystems. Unsupported
subsystems cannot be selected by the user.
For planning, TotalStorage Productivity Center must be managing the host system,
subsystem and the fabric interconnecting them. If the host, subsystem, or fabric
information is not available in the repository, then the planner will not be able generate the
plan or execute portions of the plan and will issue an error message.
For volume creation, TotalStorage Productivity Center must be managing the subsystem.
You will need an active CIMOM and must have a completed a subsystem probe.
For volume assignment, TotalStorage Productivity Center must be managing the host,
fabric, and subsystem. The host and subsystem must be in the same fabric.
For zoning configuration, TotalStorage Productivity Center must be managing the fabric.
For Brocade fabrics, an out-of-band Fabric agent must be configured with the user ID and
password for the switch. For other fabrics, an in-band Fabric agent must be connected to
the fabric for zone control operations.
A subsystem performance monitor must be run in order to select the Workload Profile
options. If a performance monitor has not been run, the Space Only workload profile
option is allowed for planning on capacity only.
The IBM Subsystem Device Driver (SDD) is required on each host for multipath planning.
2. The Planner Selection Topology Viewer panel (see Figure 10-2) contains:
– Topology Viewer in the graphical and tabular view of the configuration, from which
configuration elements can be selected.
– Select Elements pane, into which the selected elements are moved. These elements
are used later by the planners during the planning and implementation stages.
10.2.1 Introduction
You can create a SAN Plan by right-clicking the IBM TotalStorage Productivity Center →
Analytics → SAN Planner node and selecting Create Plan (see Figure 10-4).
2. Through the Launch Planner selection from the TotalStorage Productivity Center Topology
Viewer. In the graphical view of the Topology Viewer, highlight one or more supported
elements, right-click, and select Launch Planner from the context menu (see
Figure 10-7). The selection is placed into the Planner Selection pane of the SAN Planner
configuration panel.
Both methods provide a Topology Viewer from which you can select a variety of elements.
The elements allowed for selection are fabrics, storage subsystems, storage pools, volumes,
and computers. You can select a storage subsystem, not select any storage subsystems, or
select a storage subsystem and one or more storage pools. You might also select one or
more fabrics for use in path and zone planning.
10.3.1 Introduction
You must run a fabric discovery and complete subsystem and fabric probes in order to use
the SAN Planner.
Note: The SAN Planner is available for ESS, DS6000, and DS8000 systems. Unsupported
subsystems will not be allowed to be selected.
A Data agent is required to obtain host information used in volume assignments. An in-band
Fabric agent is optional if the fabric information is available through another in-band agent in
the same fabric, an outbound agent, or through a CIMOM managed switch. Run the in-band
host agent probe to get the host operating system information. If the in-band host agent has
not been run and the storage controller also does not know about the host operating system,
an asterisk is displayed next to the host port. Ensure that the operating system-imposed limits
on the number of volumes that can be accessed are not exceeded.
2. In the Create Plan pane, enter a short description of the plan into the Description field.
3. The Planner Selection pane will be empty if no input has been made.
4. Using the Add button to select elements for the Planner Selection pane is described in the
following steps:
a. Click the Add button. This brings up the Planner Selection Topology Viewer panel.
Select the storage controllers, hosts, fabrics, and storage pools to be used for volume
provisioning considerations.
The topology viewer displays the actual configuration and the plan selections. The
Topology Viewer pane displays the configuration from which selections are made. The
Select Elements pane displays the selection of elements to be used in the plan.
b. Expand the Graphical Storage view by double-clicking on the box title for a subsystem.
This expands the storage view into the L0 level view of the available subsystems.
c. Select the elements by clicking on any of the supported elements. For multiple element
selections, press the Ctrl key while you click on each element icon (see Figure 10-9).
d. Click >> to move the selected elements into the Select Elements pane. If you decide to
remove the selected elements from the Select Elements pane, click <<.
The Select Elements pane lists the fabrics, computers, subsystems, pools, and
volumes if selected. All selections must be located within the same fabric for SAN
Planner to work. The Fabrics section lists the fabrics (by WWN) and the corresponding
selected subsystems and selected hosts within each fabric. A subsystem might appear
multiple times if it has been configured in different fabrics. The Subsystems section will
contain the selected storage subsystems. Next to each subsystem, the parenthesis
contains the number of fabrics the subsystem belongs to. Selected pools will appear
under their respective storage subsystem. Selected volumes will appear under their
respective pools. The Computers section lists the selected hosts. In parenthesis, the
number of fabrics the host belongs to will be displayed. The host(s) will be used in path
planning and zone planning if Path Planner and Zone Planner are selected in the
configuration panel.
e. Click OK when you are satisfied with all your selections. This will return to the SAN
Planner Configuration panel with the selections displayed in the Planner Selection
pane.
5. When you have returned to the configuration panel, the Planner Selection pane will
contain the selected elements (see Figure 10-10). If you decide to remove an element
from the Planner Selection pane, click on the item to highlight it and click the Remove
button. The element will disappear from the Planner Selection pane and will not be used
for the SAN plan.
You might decide to save your configuration inputs to review at a later time. This is done by
clicking the Save icon on the TotalStorage Productivity Center toolbar. If hardware
configurations are modified and the saved plan is reactivated, the Planner Selection pane
contains all the elements as they were at the time the plan was saved.
6. You are ready to start the volume planning. Click on + to expand the Volume Planner
options. If you do not want to do any volume planning, uncheck the Volume Planner box to
disable the fields within the pane.
Note: Based on the user inputs, the Volume Planner selects appropriate storage
controllers, storage pools, and storage volumes when using unassigned volumes.
a. Enter the total capacity of the volumes to be used for provisioning in the Total Capacity
field.
b. Click on Divide capacity between if you want to divide the Total Capacity among one
to several volumes. For example, enter 1 and 1 volumes if you want the total capacity to
be on 1 volume. Enter 1 and 5 volumes if you want the total capacity to be divided
among 1 to 5 volumes.
c. Click on Divide capacity among volumes of size if you want to keep the capacity of
each volume between x.x GB and y.y GB.
d. On Workload Profile, select the workload profile that represents how the new volumes
will be used.
Note: An unassigned volume cannot be selected if it does not meet the performance
requirements selected in Workload Profiles and RAID Level.
h. Click the Suggest Storage Pools button to obtain a selection of storage pools.
Note: When the user has specified storage controllers in the Planner Selection, the
planner lists the storage pools from those storage controllers which have the storage
capacity as requested by the user.
Suggest Storage Pools will select a set of storage pools to create the volumes from.
Note that Suggest Storage Pools cannot be used if volumes were selected during the
selection process and are visible in the Planner Selection pane. Storage pools are
used if they are not full or are not visible from all the hosts. They also have to be fixed
block to be used by the planners.
7. Click the Get Recommendation button if you want the SAN planner to select one or more
volumes depending on the inputs you provided.
If storage controllers and storage pools were selected, then clicking on Get
Recommendation will cause the planner to consider only those pools visible in the
Planner Selection pane.
8. You are ready to start the path planning. Click on + to expand the Path Planner options
(see Figure 10-12). If you do not want to do any path planning, uncheck the Path Planner
box to disable the inputs to the Path Planner pane.
a. Select the Multipath option to determine how I/O will be distributed across all paths.
The predefined I/O options are as follows:
• Load Balancing: Sends /IO on all paths.
• Round Robin: Sends I/O on one path until a time interval expires (set in an SDD
setting at the host) or stops to use another path.
• Fail-Over: Sends I/O on one path until a failure occurs and fails over (switches) to
another path.
b. Check the box for Specify number of paths to input the number of paths you wish to
configure. Enter the number of paths in the box on the right.
c. Check the box for Use Fully redundant paths to use the paths from host to storage
subsystem(s) through a minimum of 2 fabrics. Note that this requires 2 fabrics.
9. You are ready to start the zone planning. Click on + to expand the Zone Planner options
(see Figure 10-13). If you do not want to do any zone planning, uncheck the Zone Planner
box to disable the inputs to the Zone Planner pane.
Note: Based on the user inputs, the Zone Planner allows the zoning to be changed to
ensure that hosts can see the new storage.
a. Select the Automatically create zone to indicate where zoning will be done. The
predefined options are as follows:
• <auto-zone> This is the default. The plan will be generated using the maximum
number of zones without grouping the data paths based on host, host ports, or host
bus adapters (HBAs).
• ...for each host: Create a zone for each host.
• ...for each HBA: Create a zone for each HBA.
• ...for each host port: Create a zone for each host port.
b. Check the box for Specify maximum number zones. Enter the maximum number of
zones in the box to the right.
c. Check the box for Specify maximum zone members per zone. Enter maximum
number of zone members (per zone) in the box to the right.
d. Check the box for Use active zone set if you want any zone set that is available to be
selected.
e. Check the box for Append zone name prefix and enter a zone name prefix in the box
to the right if you want to set a prefix name for each zone.
10.When the settings are specified, click the Get Recommendation button. All the settings
will be validated and surface any errors. If there are no errors, a plan is generated and the
recommended configuration is displayed in a new panel (see Figure 10-14).
a. In the Create Plan pane, enter a description in the Description field.
b. If you are satisfied with your selections, go to the When to run pane and either click
Run Now to start the task immediately when it will be submitted, or click Run Once to
start the task at the specified date and time you select.
c. The TotalStorage Productivity Center server and the GUI can be located in different
time zones. The How to handle time zones pane will allow control on running the task
at the specified time zone. Click on Use the time zone that the server runs in if you
want to use the timezone where the TotalStorage Productivity Center server resides.
Click on Use this time zone and select the timezone from the drop-down list box. This
will start the task at the specified date and time entered in the When to run pane and
use the timezone selected here.
d. Click the Execute Plan button to save the plan and to execute it. This starts the job at
the date, time, and time zone specified in the When to run and the How to handle time
zones panes. The executed task will have a job status with the ability to view the job
log(s).
11.When you execute the plan, a pop-up window will appear to confirm (Figure 10-15).
12.After the job has been submitted, a job entry will appear under IBM TotalStorage
Productivity Center → Analytics → SAN Planner → [SAN Planner Name] (see
Figure 10-16).
Customers require a way to roll up network wide summary metrics from multiple TPC servers
and report on their storage environment from a network wide perspective. In this chapter we
describe:
How to define subordinate TPC servers to the master TPC server.
How to combine reports for multiple TPC servers into an enterprise-wide rollup report that
can give the customer a full view of their environment.
Rolled up data from TPC subordinate servers are stored into new tables in the master TPC
server. Otherwise, all of the existing TPC reports would need to be modified if rollup data is
put into the existing T_RES_*** tables. The existing reports would all have to be modified to
exclude the rollup data. Using TPC V3.3 with new table definitions allow you to created, store,
and retrieve the rolled up data.
TPC subordinate servers should have no more than 1200 unique data sources. This number
includes Data agents, Fabric agents (in-band and out-of-band), CIM agents, and VM agents
(VMWare). When this threshold has been met, a new TPC subordinate server should be
deployed and all new agents should be pointed to it. If the same storage entity is managed by
multiple subordinate servers, rollup reports reflect the storage information from the
subordinate server that most recently probed that entity.
In the Navigation Tree, you can use the Rollup Reports → Asset → Agents report to view
information about Data agents and Device agents that are associated with subordinate
servers in your environment. You can generate and sort the data within this report by clauses,
as follows:
By Agent: Shows agents sorted according to the name of the machine on which they are
installed.
By OS Type: Shows agents sorted according to the operating system under which they
run. Click the magnifying glass icon next to an operating system type to view more
detailed information about each of the agents running under that operating system.
By TPC Server: Shows agents sorted according to the subordinate server that manages
them.
For the TPC server rollup feature, all servers must be at TPC V3.3 versions older than V3.3
(for example, TPC V2.3, TPC V3.1) will not support the TPC server rollup function.
Perform the following steps on the master TPC server to add a subordinate TPC server:
1. Expand the Administrative Services → Data Sources portion of the Navigation Tree.
2. Click on TPC Servers
3. Click the Add TPC Server button.
When the Add TPC Server dialog appears, enter the following items:
Host Name or IP address: We recommend that you enter a fully qualified DNS name.
Host Device or Server Port: The default is port 9550.
Host Authentication Password: This is the TPC Host Authentication password of the
Device Server.
Optionally, input a Display Name.
Optionally, input a Description.
Select the Test TPC Server connectivity before adding check box on the Add TPC
Server dialog for a subordinate TPC server that is up and running and whose
authentication data matches that of the data provided in the Add TPC Server dialog.
Click the Save button on the Add TPC Server dialog
The TPC GUI sends the request to TPC master server to add the TPC subordinate server
and perform a check. The TPC master server verifies that the subordinate TPC Server is not
already being monitored. The TPC master server will verify that the login information for the
subordinate TPC server is correct by connecting the remote TPC subordinate server. When
these verification steps are completed, the TPC master server adds the subordinate TPC
server to its database and returns a success code to the TPC GUI. The panel in Figure 11-3
illustrates a successful ADD of a TPC subordinate server.
When a probe is run, the data server on the master TPC server authenticates with the device
servers on the subordinate TPC servers by using the host authentication passwords provided
by the user. It then makes Web service calls to the TPC Device Server on the subordinate.
Batch reports can be generated on any TPC Data agent that is being directly monitored by
the TPC master server from which the enterprise-wide rollup reports are generated. Adding
capabilities to the device server services allows other users of TPC to gain access to these
functions. For instance, workflows and the TPC CLI could very easily use the new capabilities
provided by the rollup component.
To probe subordinate TPC servers, you have to create a probe definition. By default, all the
newly registered subordinate servers are added in the default TPC server probe definition.
Prior to scheduling, a TPC Server probe verifies that at least one subordinate TPC server with
data is being monitored by the master TPC server.
To schedule a TPC server probe, perform these steps from the master TPC server:
1. Expand the Administrative Services ' IBM TotalStorage Productivity Center →
Monitoring →TPC Server Probes portion of the Navigation Tree.
2. Select All TPC Servers, as shown in Figure 11-4, and all the resources will be probed.
3. You also have the choice of selecting a specific TPC servers to probe, as well as the class
of information to pull from the probed servers (subsystem, fabric, computers, clusters,
tapes, databases). This information is stored in the master's local database.
The results of a successful TPC subordinate probe should similar to what is depicted in
Figure 11-5 depending the number of TPC subordinates selected and data gathered.
To view rollup capacity information according to storage subsystem, click Rollup Reports →
Capacity → Disk Capacity → By Storage System. The Selection page in Figure 11-6 is
displayed. Use the Available Columns and Included Columns list boxes to determine what
columns are displayed in a generated report as shown in Figure 11-6.
Click Filter to further filter the objects that appear in a report. Filters enable you to apply
general rules to the report based on the rows in that report. Figure 11-8 illustrates the Filter
panel.
In the next example we guide you through the generation of an Asset information report by
computer. To generate an enterprise rollup report, launch the TPC GUI for the master TPC
master server and logon to the TPC GUI as a TPC superuser or administrator:
1. In the main Navigation Tree, expand the IBM TotalStorage Productivity Center →
My Reports → Rollup Reports node in the TPC GUI.
2. Expand either Asset or Capacity
3. Expand and select one of the reports. (The Report Selection Panel is displayed). Notice
the information in the right hand column: Rollup Reports - Asset. Use these reports to view
detailed statistics about computers.
4. You have the ability to modify the selection and filter criteria for the report. Click the
Generate Report button.
5. The report is generated and displayed.
6. The device server on the subordinate server must be running in order for the master
server to pull data.
Figure 11-11 shows the installed agents that are installed and used to provide data for rollup
reports.
Figure 11-11 Rollup reports showing installed Agents and their versions
The VMWare VirtualCenter is the management application that is the central entry point for
the management and monitoring of multiple ESX Server in a data center. To utilize the
improved VMWare support, two data sources are required. A VMWare ESX server of
VMWare Virtual Infrastructure data source is needed. Also, a TPC Data agent is required on
each virtual machine you plan to monitor.
For more information about the ESX Server or VMWare VirtualCenter, see:
http://www.vmware.com
For a list of supported VMWare products and support guest operating systems, consult the
IBM TotalStorage Productivity Center support page:
http://www.ibm.com/servers/storage/support/software/tpc/
As shown in Figure 12-1, the data flow between the VMWare environment and the TPC
consists of the two different connections. The connection of the TPC to the VMWare Host
Agent of the VMWare ESX server via the new VMWare VI Data Source and the connection of
the TPC Data agents residing in the VMWare virtual machines inside the VMWare ESX
server.
Attention: There is no need to install a Data agent on the VMWare ESX server itself.
This feature is not supported.
This topic provides information on planning for VMWare Virtual Infrastructure configuration.
Before you can display reports or see the topology for VMWare Virtual Infrastructure, you
must complete the following general steps:
1. Import certificate. If the VMWare Virtual Infrastructure uses SSL certificates for
communication, you will have to use keytool to manually import the SSL certificates into a
truststore. Each Virtual Infrastructure data source provides an individual certificate. There
will be a default truststore registered in the Device Server’s system properties file. keytool
is a tool shipped with the Java run-time environment.
2. Add the VMWare VI data source. The data source can be a hypervisor (ESX Server or
VirtualCenter). This is the first step in getting information from VMWare Virtual
Infrastructure. Adding a VMWare data source is similar to adding a CIM agent or Data
agent.
3. Test the connection to the VMWare VI data source. This ensures that you can access
information from the VMWare data source.
4. Run a discovery job for the VMWare environment. The discovery is needed to retrieve
every ESX Server instance that is part of the Virtual Infrastructure that has been added.
The discovery mechanism is similar to a discovery for storage subsystems. Discovery jobs
can be scheduled and are performed on the complete list of known VMWare data sources.
5. Run a probe job for the ESX Server, hypervisor, and virtual machines. This step will get
the detailed information from the hypervisors and virtual machines for IBM TotalStorage
Productivity Center.
6. Configure alerts for VMWare. You can create alerts for the following alert conditions:
– Hypervisor discovered
– Hypervisor missing
– Virtual Machine added
– Virtual Machine deleted
7. Install the Data agent on each of the virtual machines you wish to monitor. For full
functionality, you need two data sources.
8. You will now be able to view VMWare reports and VMWare topology.
For reports, note the following considerations: You must probe both the ESX Server and
the Data agent in the virtual machines before you can generate accurate reports for disk
and file system capacity. For example, you have an ESX Server that has 100 GB and 60
GB is allocated to the virtual machine. The virtual machine uses 5 GB of space. Both the
ESX Server (H1) and the virtual machine (VC1) have been probed. You also have a
physical computer (PC1) that has been probed. The TOTAL capacity for the file system or
disk capacity row includes everything — virtualized disks and virtual machines as well as
non-virtualized disks and machines.
Use a current version of keytool, so regarding Figure 12-2, the keytool.exe marked on the
top does qualify.
The syntax to create the truststore for the TPC server is:
keytool -import -file <certificate-filename> -alias <server-name> -keystore
vmware.jks
4. Truststores are located in the Device Server configuration directory of your TPC server:
<TPC_install_directory>/device/conf
Copy the newly created certificate store / truststore to the Device Server configuration
directory of your TPC server.
The truststore will automatically be defined at service startup time as the following
property in the Device Server JVM:
javax.net.ssl.trustStore System
5. Add the VMWare VI data source. The data source can be a hypervisor (ESX Server or
VirtualCenter). This is the first step in getting information from VMWare Virtual
Infrastructure. Adding a VMWare data source is similar to adding a CIM agent or Data
agent. Go to Administrative Services → Data Sources → VMWare VI Data Source in
your TPC GUI and click Add VMWare VI Data Source (see Figure 12-5).
If the connection to your data source is successful, the information window will look similar
to Figure 12-7,
You must probe both the ESX Server and the Data agent in the virtual machines before
you can generate accurate reports for disk and file system capacity. For example, suppose
that you have an ESX Server that has 100 GB and 60 GB is allocated to the virtual
machine. The virtual machine uses 5 GB of space. Both the ESX Server (H1) and the
virtual machine (VC1) have been probed. You also have a physical computer (PC1) that
has been probed. The TOTAL capacity for the file system or disk capacity row includes
everything — virtualized disks and virtual machines as well as non-virtualized disks and
machines.
The solution to achieve a display export described here is one of many possible ways to do it.
Our servers and the environment we use are behind a firewall. It does not allow connections
to be made from the AIX server behind the firewall to the outside machines in front of the
firewall. Therefore we decided to implement the following solution, which is based on the use
of ssh, ssl, rpm, cygwin, and PuTTY. The solution is described utilizing an AIX server and a
Windows workstation. It will also work with other UNIX distributions and Linux if the involved
tools are applied properly.
cygwin installation
Download the setup.exe as follows:
1. On the Windows workstation you want to use to receive the X11 forwarding, double-click
setup.exe. The cygwin setup will start and welcome you with a panel similar to the one
shown in Figure A-1.
2. Click Next to continue with the installation. A new window similar to Figure A-2 will display
and ask you to choose your installation type. Assuming that this is the first time you are
installing cygwin, you select the option Install from Internet to access the product source
files from an online repository. Click Next.
3. The installation will ask you to choose an installation directory. Best practice is to leave the
options selectable here at their default value as shown in Figure A-3. Try to avoid adding
additional directory levels, different drive locations, or spaces in the directory name and
path. Click Next to continue.
This dialog lets you specify your type of Internet connection. If you use a proxy or other
special settings to connect to the internet, you can specify that in Figure A-5. Click Next to
continue.
If you use a firewall product, make sure that the installer is allow to connect to the Internet
(see Figure A-7).
This dialog lets you select the packages to install. Scroll through the packages and open
All → X11 → xwinwm (see Figure A-9). Make sure that it is selected for installation.
Additional required packages will automatically be installed. Click Next to continue.
After the installation is finished, you are prompted to choose whether to create a desktop icon
and a start menu icon. Select what fits you best and click Finish to end the installation
(see Figure A-11).
Congratulations, you have successfully installed cygwin. Click OK to end the installation
(see Figure A-12).
The second part of this solution is optional. To achieve X11 forwarding from the AIX system to
your Windows workstation, you need to open an ssh connection to the AIX system. This can
be done using cygwin or using PuTTY.
To open an ssh connection to the AIX server using cygwin, open a cygwin bash shell:
Start → All Programs → cygwin → Cygwin Bash Shell
From within the newly opened cygwin bash shell window, issue the command:
Xwin -multiwindow
To start the X11 server on your Windows workstation. The -multiwindow option to the XWin
command will make all X11 windows forwarded from any server to appear in their own,
separate Windows window. If you prefer to have all the X11 forwardings appear in one single
Windows window, just start XWin without any options.
From within the newly opened xterm window, enter the following command:
Start → Run...
and enter:
C:\cygwin\bin\run.exe -p /usr/X11R6/bin xterm -display 127.0.0.1:0.0 -ls
From within the newly created xterm window, enter the following commands:
xhost +
This command will disable security and allow all external hosts to forward X11 to your
machine.
ssh -X root@9.43.86.101
This command will create an ssh connection to the remote machine 9.43.86.101 as root user
and enable X11 forwarding through that tunnel.
After successfully connecting to the remote machine, start a graphical terminal window from
the remote host to verify that X11 forwarding is working by issuing the following command
(see Figure A-13):
xterm &
This will open a graphical terminal window from the remote machine exported to your local
workstation. Congratulations, you have successfully configured X11 forwarding.
PuTTY installation
If you want to use PuTTY to achieve X11 forwarding, you first need to install it. Download the
executable from:
http://www.chiark.greenend.org.uk/~sgtatham/putty/download.html
and put it into a directory on your workstation or download and use the PuTTY installer.
Either way, you will be able to start PuTTY by simply executing the command:
putty.exe
On the left side of the PuTTY windows, browse to Connection → SSH → X11 and check the
Enable X11 forwarding check box (see Figure A-15).
X11Forwarding yes
X11DisplayOffset 10
X11UseLocalhost yes
Carefully browse through the results of the command and identify the ssh daemon.
In our example the ssh daemon has the process number 197108. To restart our ssh daemon
issue:
kill -HUP 197108
You might have to disconnect your ssh session now and reconnect to enable the new settings
for your session.
Then, on the Windows workstation, start cygwin by clicking Start → All Programs →
cygwin → Cygwin Bash Shell
From within the newly opened cygwin bash shell window, issue Xwin -multiwindow to start the
X11 server on your Windows workstation. The -multiwindow option to the XWin command will
make all X11 windows forwarded from any server to appear in their own, separate Windows
window. If you prefer to have all the X11 forwardings appear in one single Windows window,
just start XWin without any options.
With the cygwin bash shell window still open and the X11 server on your Windows
workstation running, go back to the PuTTY window which should still be open and issue:
xterm &
Appendix B. Worksheets
This appendix contains worksheets that are meant for you to use during the planning and the
installation of the TotalStorage Productivity Center. The worksheets are meant to be
examples. Therefore you can decide whether you need to use them, for example, if you
already have all or most of the information collected somewhere.
If the tables are too small for your handwriting, or you want to store the information in an
electronic format, simply use a word processor or spreadsheet application, and use our
examples as a guide, to create your own installation worksheets.
You need one of the worksheets in the following sections for each machine where at least one
of the components or agents of Productivity Center will be installed. This is because you can
have multiple DB2 databases or logon accounts and you need to remember the IDs of each
DB2 individually.
Server information
Table B-1 contains detailed information about the servers that comprise the TotalStorage
Productivity Center environment.
Machine
Hostname
IP address ____.____.____.____
In Table B-2, simply mark whether a manager or a component will be installed on this
machine.
DB2
agentTrust.jks
Enter the user IDs and password that you used during the installation in Table B-4. Depending
on the selected managers and components, some of the lines are not used for this machine.
c
Host Authentication
Determine if there are firewalls in the IP path between the TotalStorage Productivity Center
server or servers and the devices, which might not allow the necessary communication. In the
first column of each table, enter as much information as possible to identify the devices later.
Important: Check the device support matrix for the associated CIM Agent.
For example, each time an I/O (a read or write) is issued to a volume, several counters (I/O
count, Bytes transferred) are incremented. If the counters are pulled at times T1 and T2, the
number of I/Os in the sample interval is obtained by subtracting the counters at time T1 from
the counters at time T2 (T2-T1). When this count is divided by the number of seconds
between T1 and T2, we obtain the I/O rate in I/Os/second for the sample interval (T1 to T2).
This is the technique, and is pretty simple for metrics like I/O rate, data rate, average transfer
size, and so forth. Other metrics, like Read hit ratios or Disk Utilization involve other
calculations involving sampled counters and times T1 and T2.
The counters in the firmware are usually unsigned 32 or 64 bit counters. Eventually, these
counters will “wrap,” meaning that the difference between the counters at T2 and T1 might be
difficult to interpret. The TPC Performance Manager attempts to adjust for such wraps during
its delta computations, but there might be unexpected wraps which can confuse the CIM
agent or the TPC Performance Manager. The TPC Performance Manger stores the deltas in
the database. Some counters are also stored in the TPC database, but the performance data
is mostly comprised of rates and other calculated metrics that depend on the counter deltas
and the sample interval, that is, the time between T1 and T2.
The primary and essential performance metrics are few and simple - for example, Read I/O
Rate, Write I/O Rate, Read Response Time and Write Response Time. Also important are
data rates and transfer sizes. Then come the cache behaviors, in the form of Read Hit Ratio
and Write Cache delays (percentages and rates). There are a myriad of additional metrics in
the TPC performance reports, but they should be used as adjuncts to the primary metrics,
sometimes helping to understand why the primary metrics have the values they do.
There are a very few metrics which measure other kinds of values. For example, the SVC
storage subsystem also reports the maximum read and write response times which occur
between times T1 and T2. Each time a sample of the counters is pulled, this kind of counter is
set back to zero. But the vast majority of counters are monotonically increasing, reset to zero
only by very particular circumstances (like hardware, software, or firmware resets).
The design of the TPC Performance Manager is such that several storage subsystems can be
included in a report (or individual subsystems by selection or filtering). But not all the metrics
apply to every subsystem of component. In these cases, a "-1" appears, indicating that no
data is expected for the metric in this particular case.
In the remainder of this section we look at the metrics that can be selected for each report.
We examine the reports in the order in which they appear in the TPC Navigation Tree.
Subsystem Component id
Read I/O Rate (normal) Average number of normal read operations per second for
the sample interval. Normal operations are not sequential,
hence random. This metric is only for ESS, DS8000,
DS6000
Read I/O Rate (sequential) Average number of sequential read operations per second
for the sample interval. Sequential I/O is detected by the
subsystem. This metric is only for ESS, DS8000, DS6000
Read I/O Rate (overall) Average number of read operations per second for the
sample interval. Applies to most subsystems.
Write I/O Rate (normal) Average number of normal write operations per second for
the sample interval. Normal operations are not sequential,
hence random. This metric is only for ESS, DS8000,
DS6000
Write I/O Rate (sequential) Average number of sequential write operations per second
for the sample interval. Sequential I/O is detected by the
subsystem. This metric is only for ESS, DS8000, DS6000
Write I/O Rate (overall) Average number of write operations per second for the
sample interval. Applies to most subsystems.
Total I/O Rate (normal) Average number of normal reads and writes per second for
the sample interval. Normal operations are not sequential,
hence random. This metric is only for ESS, DS8000,
DS6000
Total I/O Rate (sequential) Average number of sequential reads and writes per second
for the sample interval. Sequential I/O is detected by the
subsystem. This metric is only for ESS, DS8000, DS6000
Total I/O Rate (overall) Average number of reads and writes per second for the
sample interval. Applies to most subsystems.
Read Cache Hits Percentage Percentage of sequential reads that are cache hits in the
(sequential) sample interval. Only for ESS, DS8000, DS6000.
Read Cache Hits Percentage Percentage of reads during the sample interval that are
(overall) found in cache. This is an important metric.
Write Cache Hits Percentage Percentage of normal (random) writes that are handled in
(normal) cache. This number should be 100%. Only for ESS, DS8K,
DS6K
Write Cache Hits Percentage Percentage of sequential writes that are handled in cache.
(sequential) This number should be 100%. Only for ESS, DS8K, DS6K
Write Cache Hits Percentage (overall) Percentage of writes that are handled in cache. This number
should be 100% for most enterprise storage.
Total Cache Hits Percentage (normal) Percentage of normal reads and writes that are cache hits
during the sample interval.
Total Cache Hits Percentage Percentage of sequential reads and writes that are cache
(sequential) hits during the sample interval.
Total Cache Hits Percentage (overall) Weighted average of read cache hits and write cache hits.
Read Data Rate Average read data rate in megabytes per second during the
sample interval
Write Data Rate Average write data rate in megabytes per second during the
sample interval
Total Data Rate Average total (read + write) data rate in megabytes per
second during the sample interval.
Read Response Time Average response time in milliseconds for reads during the
sample interval. For this report, this is an average of read
hits in cache as well as read misses.
Write Response Time Average response time in milliseconds for writes during the
sample interval.
Overall Response Time Average response time in milliseconds for all I/O in the
sample interval, including both cache hits as well as misses
to backing storage if required.
Read Transfer Size Average transfer size in kilobytes for reads during the
sample interval
Write Transfer Size Average transfer size in kilobytes for writes during the
sample interval
Overall Transfer Size Average transfer size in kilobytes for all I/O during the
sample interval.
Record Mode Read I/O Rate This is the rate in I/O/sec for a special kind of read activity
detected by ESS, DS8K and DS6K. Only the requested data
is managed in cache rather than a full track or most of a track
of data.
Record Mode Read Cache Hit Read Hit percentage for the special class of reads
Percentage mentioned above. ESS, DS8k, DS6K only.
Cache to Disk Transfer Rate Average number of track transfers per second from cache to
disk during the sample interval.
Write-cache Delay Percentage Percentage of all I/O operations that were delayed due to
write-cache space constraints or other conditions during the
sample interval. Only writes can be delayed, but the
percentage is of all I/O.
Write-cache Delay I/O Rate The rate of I/O (actually writes) that are delayed during the
sample interval because of write cache.
Cache Holding Time The average number of seconds a piece of data will stay in
cache. This value is calculated using Little's Law, only for
DS8K, DS6K, and ESS.
Backend Read I/O Rate The average read rate in reads per second caused by read
misses. This is the read rate to the back-end storage for the
sample interval.
Backend Write I/O Rate The average write rate in writes per second caused by
front-end write activity. This is the write rate to the back-end
storage for the sample interval. These are logical writes and
the actual number of physical I/O operations depends on
whether the storage is RAID 5, RAID 10, or some other
architecture.
Total Backend I/O Rate The sum of Backend Read I/O Rate and Backend Write I/O
Rate over the sample interval.
Backend Read Data Rate Average number of megabytes per second read from
back-end storage during the sample interval.
Backend Write Data Rate Average number of megabytes per second written to
back-end storage during the sample interval.
Total Backend Data Rate Sum of the Backend Read and Write Data Rates for the
sample interval.
Backend Read Response Time Average response time in milliseconds for read operations to
the back-end storage.
Backend Write Response Time Average response time in milliseconds for write operations
to the back-end storage. This time can include several
physical I/O operations, depending on the type of RAID
architecture.
Overall Backend Response Time The weighted average of Backend read and write response
times during the sample interval.
Backend Read Transfer Size The average transfer size in kilobytes for reads to the
back-end storage during the sample interval.
Backend Write Transfer Size The average transfer size in kilobytes for data written to the
back-end storage during the sample interval.
Overall Backend Transfer Size Weighted average transfer size in kilobytes for backend
reads and writes during the sample interval.
Port Send I/O Rate The average rate per second for operations that send data
from an I/O port, typically to a server. This is typically a read
from the server's perspective.
Total Port I/O Rate Average read plus write I/O rate per second at the storage
port during the sample interval.
Port Send Data Rate The average data rate in megabytes per second for
operations that send data from an I/O port, typically to a
server.
Port Receive Data Rate The average data rate in megabytes per second for
operations where the storage port receives data, typically
from a server.
Total Port Data Rate Average read plus write data rate in megabytes per second
at the storage port during the sample interval
Port Send Response Time Average number of milliseconds that it took to service each
port send (server read) operation, for a particular port over
the sample interval.
Port Receive Response Time Average number of milliseconds that it took to service each
port receive (server write) operation, for a particular port
over the sample interval.
Total Port Response Time Weighted average port send and port receive time over the
sample interval.
Port Send Transfer Size Average size in kilobytes per Port Send operation during the
sample interval.
Port Receive Transfer Size Average size in kilobytes per Port Receive operation during
the sample interval.
Total Port Transfer Size Average size in kilobytes per port transfer during the sample
interval.
Read Queue Time For SVC, the average number of milliseconds that each
read operation during the sample interval spent on the
queue before being issued to the back-end storage device
Write Queue Time For SVC, the average number of milliseconds that each
write operation during the sample interval spent on the
queue before being issued to the back-end storage device
Overall Queue Time For SVC, the weighted average of Read Queue Time and
Write Queue Time during the sample interval.
Readahead Percentage of Cache For SVC, an obscure measurement of cache hits involving
Hits data that has been prestaged for one reason or another.
Dirty Write Percentage of Cache Hits For SVC, the percentage of write cache hits which modified
only data that was already marked "dirty" in the cache;
re-written data. This is an obscure measurement of how
effectively writes are coalesced before destaging.
Write Cache Overflow Percentage For SVC the percentage of write operations that were
delayed due to lack of write-cache space during the sample
interval.
Write Cache Overflow I/O Rate For SVC, the average rate per second of write operations
that were delayed due to lack of write-cache space during
the sample interval.
Write Cache Flush-through I/O Rate For SVC, the average rate per second of tracks processed in
Flush-through write mode during the sample interval.
Write Cache Write-through For SVC the percentage of write operations that were
Percentage processed in Write-through write mode during the sample
interval.
Write Cache Write-through I/O Rate For SVC, the average number of tracks per second that were
processed in Write-through write mode during the sample
interval.
CPU Utilization Percentage For SVC the average utilization of the cluster node
controllers during the sample interval.
Port to Host Send I/O Rate For SVC, the rate per second of port send to host (server)
during the sample interval.
Port to Host Receive I/O Rate For SVC, the rate per second of port receive operations from
host (server) during the sample interval.
Total Port to Host I/O Rate For SVC, total of port send and receive IO rate during the
sample interval.
Port to Disk Send I/O Rate For SVC, the rate per second of port send to back-end
storage during the sample interval.
Port to Disk Receive I/O Rate For SVC, the rate per second of port receive operations from
back-end storage during the sample interval.
Total Port to Disk I/O Rate For SVC, the sum of port to disk send and port to disk
receive rates during the sample interval.
Port to Local Node Send I/O Rate For SVC, the rate per second at which a port sends I/O to
other nodes in the local cluster during the sample interval.
Port to Local Node Receive I/O Rate For SVC, the rate at which a port receives I/O from other
nodes in the local cluster during the sample interval
Total Port to Local Node I/O Rate For SVC, the sum of port to local node send and receive
rates during the sample interval.
Port to Remote Node Send I/O Rate For SVC, the average number of exchanges (I/Os) per
second sent to nodes in the remote SVC cluster during the
sample interval. Typically some form of remote mirroring.
Port to Remote Node Receive I/O For SVC, the average number of exchanges (I/Os) per
Rate second received from nodes in the remote SVC cluster
during the sample interval. Typically some form of remote
mirroring.
Total Port to Remote Node I/O Rate For SVC, the sum of port to remote node send and receive
I/O per second during the sample interval.
Port to Host Send Data Rate For SVC, the megabytes per second of port send to host
(server) during the sample interval.
Port to Host Receive Data Rate For SVC, the megabytes per second of port receive
operations from host (server) during the sample interval.
Total Port to Host Data Rate For SVC, total of port send and receive megabytes per
second during the sample interval.
Port to Disk Receive Data Rate For SVC, the megabytes per second of port receive
operations from back-end storage during the sample interval
Total Port to Disk Data Rate For SVC, the sum of port to disk send and port to disk
receive megabytes per second during the sample interval.
Port to Local Node Send Data Rate For SVC, the megabytes per second at which a port sends
I/O to other nodes in the local cluster during the sample
interval.
Port to Local Node Receive Data For SVC, the megabytes per second at which a port receives
Rate I/O from other nodes in the local cluster during the sample
interval
Total Port to Local Node Data Rate For SVC, the sum of port to local node send and receive
megabytes per second during the sample interval.
Port to Remote Node Send Data Rate For SVC, the average number of megabytes per second
sent to nodes in the remote SVC cluster during the sample
interval. Typically some form of remote mirroring.
Port to Remote Node Receive Data For SVC, the average number of megabytes per second
Rate received from nodes in the remote SVC cluster during the
sample interval. Typically some form of remote mirroring.
Total Port to Remote Node Data Rate For SVC, the sum of port to remote node send and receive
megabytes per second during the sample interval.
Port to Local Node Send Response For SVC the average port service time in milliseconds for
Time these operations during the sample interval.
Port to Local Node Receive For SVC the average port service time in milliseconds for
Response Time this operation during the sample interval.
Overall Port to Local Node Response For SVC the average port service time in milliseconds for
Time these operations during the sample interval.
Port to Local Node Send Queue Time For SVC, the average time in msec waiting in queue before
these send operations are executed.
Port to Local Node Receive Queue For SVC, the average time in msec waiting in queue before
Time these receive operations are executed.
Overall Port to Local Node Queue For SVC, the average time in msec. waiting before these
Time port send or port receive operations are executed.
Port to Remote Node Send Response For SVC the average port service time in milliseconds for
Time these operations during the sample interval.
Port to Remote Node Receive For SVC the average port service time in milliseconds for
Response Time these operations during the sample interval.
Overall Port to Remote Node For SVC the average port service time in milliseconds for
Response Time these operations during the sample interval.
Port to Remote Node Send Queue For SVC, the average time in msec waiting in queue before
Time these send operations are executed.
Port to Remote Node Receive Queue For SVC, the average time in msec waiting in queue before
Time these send operations are executed.
Global Mirror Write I/O Rate For SVC, the rate in writes per second issued to the
secondary site for Global Mirror during the sample interval
Global Mirror Overlapping Write For SVC, the percentage of writes during the sample
Percentage interval, for which the write operations at the primary site for
Global Mirror have overlapping write domains.
Global Mirror Overlapping Write I/O For SVC, the average rate in writes per second during the
Rate sample interval, for which the write operations at the primary
site for Global Mirror have overlapping write domains.
Peak Read Response Time For SVC, the peak read response time in msec observed
during the sample interval. At the end of each sample
interval, this value is reset to zero.
Peak Write Response Time For SVC, the peak write response time in msec observed
during the sample interval. At the end of each sample
interval, this value is reset to zero.
Global Mirror Secondary Write Lag For SVC, the number of additional milliseconds it took to
service each secondary write operation for Global Mirror,
over and above the time needed to service the primary
writes during the sample interval.
By Controller report
This report is only for DS8000, DS6000, and ESS.
Critical and universal metrics are in bold. Less important metrics are in normal font, and
difficult metrics to interpret are in italics.
Subsystem Component id
Controller Component id
Read I/O Rate (normal) Average number of normal read operations per second for
the sample interval. Normal operations are not sequential,
hence random.
Read I/O Rate (sequential) Average number of sequential read operations per second
for the sample interval. Sequential IO is detected by the
subsystem.
Read I/O Rate (overall) Average number of read operations per second for the
sample interval.
Write I/O Rate (normal) Average number of normal write operations per second for
the sample interval. Normal operations are not sequential,
hence random.
Write I/O Rate (sequential) Average number of sequential write operations per second
for the sample interval. Sequential I/O is detected by the
subsystem.
Total I/O Rate (normal) Average number of normal reads and writes per second for
the sample interval. Normal operations are not sequential,
hence random.
Total I/O Rate (sequential) Average number of sequential reads and writes per second
for the sample interval. Sequential I/O is detected by the
subsystem.
Total I/O Rate (overall) Average number of reads and writes per second for the
sample interval
Read Cache Hit Percentage (normal) Percentage of normal (random) reads that are cache hits
during the sample interval.
Read Cache Hits Percentage Percentage of sequential reads that are cache hits in the
(sequential) sample interval.
Read Cache Hits Percentage Percentage of reads during the sample interval that are
(overall) found in cache. This is an important metric.
Write Cache Hits Percentage (normal) Percentage of normal (random) writes that are handled in
cache. This number should be 100%.
Write Cache Hits Percentage Percentage of sequential writes that are handled in cache.
(sequential) This number should be 100%.
Write Cache Hits Percentage (overall) Percentage of writes that are handled in cache. This number
should be 100% for most enterprise storage.
Total Cache Hits Percentage (normal) Weighted average of read cache hits and write cache hits.
Total Cache Hits Percentage Percentage of sequential reads and writes that are cache
(sequential) hits during the sample interval.
Total Cache Hits Percentage (overall) Weighted average of read cache hits and write cache hits.
Read Data Rate Average read data rate in megabytes per second during the
sample interval
Write Data Rate Average write data rate in megabytes per second during the
sample interval
Total Data Rate Average total (read + write) data rate in megabytes per
second during the sample interval.
Read Response Time Average response time in milliseconds for reads during the
sample interval. For this report, this is an average of read
hits in cache as well as read misses.
Write Response Time Average response time in milliseconds for writes during the
sample interval.
Overall Response Time Average response time in milliseconds for all IO in the
sample interval, including both cache hits as well as misses
to backing storage if required.
Read Transfer Size Average transfer size in kilobytes for reads during the
sample interval
Write Transfer Size Average transfer size in kilobytes for writes during the
sample interval
Record Mode Read I/O Rate This is the rate in I/O/sec for a special kind of read activity
detected by ESS, DS8K and DS6K. Only the requested data
is managed in cache rather than a full track or most of a track
of data.
Record Mode Read Cache Hit Read Hit percentage for the special class of reads
Percentage mentioned above.
Disk to Cache Transfer Rate Average number of track transfers per second from disk to
cache during the sample interval.
Cache to Disk Transfer Rate Average number of track transfers per second from cache to
disk during the sample interval.
Write-cache Delay Percentage Percentage of all I/O operations that were delayed due to
write-cache space constraints or other conditions during the
sample interval. Only writes can be delayed, but the
percentage is of all I/O. This is sometimes called NVS Full.
Write-cache Delay I/O Rate The rate of I/O (actually writes) that are delayed during the
sample interval because of write cache, sometimes called
NVS Full.
Cache Holding Time The average number of seconds a piece of data will stay in
cache. This value is calculated using Little's Law.
Backend Read I/O Rate The average read rate in reads per second caused by read
misses. This is the read rate to the back-end RAID arrays for
the sample interval.
Backend Write I/O Rate The average write rate in writes per second caused by
front-end write activity. This is the write rate to the back-end
storage for the sample interval. These are logical writes and
the actual number of physical I/O operations depends on
whether the storage is RAID 5, RAID 10, or some other
architecture.
Total Backend I/O Rate The sum of Backend Read I/O Rate and Backend Write I/O
Rate over the sample interval.
Backend Read Data Rate Average number of megabytes per second read from
back-end storage during the sample interval.
Backend Write Data Rate Average number of megabytes per second written to
back-end storage during the sample interval.
Total Backend Data Rate Sum of the Backend Read and Write Data Rates for the
sample interval.
Backend Read Response Time Average response time in milliseconds for read operations
to the back-end storage.
Backend Write Response Time Average response time in milliseconds for write operations
to the back-end storage. This time can include several
physical I/O operations, depending on the type of RAID
architecture.
Overall Backend Response Time The weighted average of Backend read and write response
times during the sample interval.
Backend Write Transfer Size The average transfer size in kilobytes for data written to the
back-end storage during the sample interval.
Overall Backend Transfer Size Weighted average transfer size in kilobytes for backend
reads and writes during the sample interval.
Critical and universal metrics are in bold. Less important ones in normal font and difficult
metrics to interpret are in italics.
Read I/O Rate (overall) Average number of read operations per second for the sample
interval. Applies to most subsystems.
Write I/O Rate (overall) Average number of write operations per second for the sample
interval. Applies to most subsystems.
Total I/O Rate (overall) Average number of reads and writes per second for the sample
interval. Applies to most subsystems.
Read Cache Hits Percentage Percentage of reads during the sample interval that are found
(overall) in cache. This is an important metric.
Read Data Rate Average read data rate in megabytes per second during the
sample interval
Write Data Rate Average write data rate in megabytes per second during the
sample interval
Total Data Rate Average total (read + write) data rate in megabytes per second
during the sample interval.
Read Response Time Average response time in milliseconds for reads during the
sample interval. For this report, this is an average of read hits
in cache as well as read misses.
Write Response Time Average response time in milliseconds for writes during the
sample interval.
Overall Response Time Average response time in milliseconds for all I/O in the sample
interval, including both cache hits as well as misses to backing
storage if required.
Write Transfer Size Average transfer size in kilobytes for writes during the sample
interval
Overall Transfer Size Average transfer size in kilobytes for all I/O during the sample
interval.
Disk to Cache Transfer Rate Average number of track transfers per second from disk to
cache during the sample interval.
Cache to Disk Transfer Rate Average number of track transfers per second from cache to
disk during the sample interval.
Write-cache Delay Percentage Percentage of all I/O operations that were delayed due to
write-cache space constraints or other conditions during the
sample interval. Only writes can be delayed, but the
percentage is of all I/O.
Write-cache Delay I/O Rate The rate of I/O (actually writes) that are delayed during the
sample interval because of write cache.
Backend Read I/O Rate The average read rate in reads per second caused by read
misses. This is the read rate to the back-end storage for the
sample interval.
Backend Write I/O Rate The average write rate in writes per second caused by
front-end write activity. This is the write rate to the back-end
storage for the sample interval. These are logical writes
Total Backend I/O Rate The sum of Backend Read I/O Rate and Backend Write I/O
Rate over the sample interval.
Backend Read Data Rate Average number of megabytes per second read from back-end
storage during the sample interval.
Backend Write Data Rate Average number of megabytes per second written to back-end
storage during the sample interval.
Total Backend Data Rate Sum of the Backend Read and Write Data Rates for the sample
interval.
Backend Read Response Time Average response time in milliseconds for read operations to
the back-end storage.
Backend Write Response Time Average response time in milliseconds for write operations to
the back-end storage. This time can include several physical
I/O operations, depending on the type of RAID architecture.
Overall Backend Response Time The weighted average of Backend read and write response
times during the sample interval.
Read Queue Time The average number of milliseconds that each read operation
during the sample interval spent on the queue before being
issued to the back-end storage device
Write Queue Time The average number of milliseconds that each write operation
during the sample interval spent on the queue before being
issued to the back-end storage device
Overall Queue Time The weighted average of Read Queue Time and Write Queue
Time during the sample interval.
Backend Write Transfer Size The average transfer size in kilobytes for data written to the
back-end storage during the sample interval.
Overall Backend Transfer Size Weighted average transfer size in kilobytes for backend reads
and writes during the sample interval.
Port Send I/O Rate The average rate per second for operations that send data from
an I/O port, typically to a server. This is typically a read from the
server's perspective.
Port Receive I/O Rate The average rate per second for operations where the storage
port receives data, typically from a server. This is typically a
write from the server's perspective.
Total Port I/O Rate Average read plus write I/O rate per second at the storage port
during the sample interval.
Port Send Data Rate The average data rate in megabytes per second for operations
that send data from an I/O port, typically to a server.
Port Receive Data Rate The average data rate in megabytes per second for operations
where the storage port receives data, typically from a server.
Total Port Data Rate Average read plus write data rate in megabytes per second at
the storage port during the sample interval
Readahead Percentage of Cache An obscure measurement of cache hits involving data that has
Hits been prestaged for one reason or another.
Dirty Write Percentage of Cache The percentage of write cache hits which modified only data
Hits that was already marked "dirty" in the cache; re-written data.
This is an obscure measurement of how effectively writes are
coalesced before destaging.
Write Cache Overflow Percentage For SVC the percentage of write operations that were delayed
due to lack of write-cache space during the sample interval.
Write Cache Overflow I/O Rate For SVC, the average rate per second of write operations that
were delayed due to lack of write-cache space during the
sample interval.
Write Cache Flush-through For SVC, the percentage of write operations that were
Percentage processed in Flush-through write mode during the sample
interval.
Write Cache Flush-through I/O For SVC, the average rate per second of tracks processed in
Rate Flush-through write mode during the sample interval.
Write Cache Write-through For SVC the percentage of write operations that were
Percentage processed in Write-through write mode during the sample
interval.
Write Cache Write-through I/O Rate For SVC, the average number of tracks per second that were
processed in Write-through write mode during the sample
interval.
CPU Utilization Percentage The average utilization of the node controllers in this I/O group
during the sample interval.
Port to Host Send I/O Rate For SVC, the rate per second of port send to host (server)
during the sample interval.
Total Port to Host I/O Rate For SVC, total of port send and receive I/O rate during the
sample interval.
Port to Disk Send I/O Rate For SVC, the rate per second of port send to back-end storage
during the sample interval.
Port to Disk Receive I/O Rate For SVC, the rate per second of port receive operations from
back-end storage during the sample interval.
Total Port to Disk I/O Rate For SVC, the sum of port to disk send and port to disk receive
rates during the sample interval.
Port to Local Node Send I/O Rate For SVC, the rate per second at which a port sends I/O to other
nodes in the local cluster during the sample interval.
Port to Local Node Receive I/O For SVC, the rate at which a port receives I/O from other nodes
Rate in the local cluster during the sample interval
Total Port to Local Node I/O Rate For SVC, the sum of port to local node send and receive rates
during the sample interval.
Port to Remote Node Send I/O Rate For SVC, the average number of exchanges (I/Os) per second
sent to nodes in the remote SVC cluster during the sample
interval. Typically some form of remote mirroring.
Port to Remote Node Receive I/O For SVC, the average number of exchanges (I/Os) per second
Rate received from nodes in the remote SVC cluster during the
sample interval. Typically some form of remote mirroring.
Total Port to Remote Node I/O Rate For SVC, the sum of port to remote node send and receive I/O
per second during the sample interval.
Port to Host Send Data Rate For SVC, the megabytes per second of port send to host
(server) during the sample interval.
Port to Host Receive Data Rate For SVC, the megabytes per second of port receive operations
from host (server) during the sample interval.
Total Port to Host Data Rate For SVC, total of port send and receive megabytes per second
during the sample interval.
Port to Disk Send Data Rate For SVC, the megabytes per second of port send to back-end
storage during the sample interval.
Port to Disk Receive Data Rate For SVC, the megabytes per second of port receive operations
from back-end storage during the sample interval
Total Port to Disk Data Rate For SVC, the sum of port to disk send and port to disk receive
megabytes per second during the sample interval.
Port to Local Node Send Data Rate For SVC, the megabytes per second at which a port sends I/O
to other nodes in the local cluster during the sample interval.
Port to Local Node Receive Data For SVC, the megabytes per second at which a port receives
Rate I/O from other nodes in the local cluster during the sample
interval
Total Port to Local Node Data Rate For SVC, the sum of port to local node send and receive
megabytes per second during the sample interval.
Port to Remote Node Receive Data For SVC, the average number of megabytes per second
Rate received from nodes in the remote SVC cluster during the
sample interval. Typically some form of remote mirroring.
Total Port to Remote Node Data For SVC, the sum of port to remote node send and receive
Rate megabytes per second during the sample interval.
Port to Local Node Send Response For SVC the average port service time in milliseconds for these
Time operations during the sample interval.
Port to Local Node Receive For SVC the average port service time in milliseconds for this
Response Time operation during the sample interval.
Overall Port to Local Node For SVC the average port service time in milliseconds for these
Response Time operations during the sample interval.
Port to Local Node Send Queue For SVC, the average time in msec waiting in queue before
Time these send operations are executed.
Port to Local Node Receive Queue For SVC, the average time in msec waiting in queue before
Time these receive operations are executed.
Overall Port to Local Node Queue For SVC, the average time in msec. waiting before these port
Time send or receive operations are executed.
Port to Remote Node Send For SVC the average port service time in milliseconds for these
Response Time operations during the sample interval.
Port to Remote Node Receive For SVC the average port service time in milliseconds for these
Response Time operations during the sample interval.
Overall Port to Remote Node For SVC the average port service time in milliseconds for these
Response Time operations during the sample interval.
Port to Remote Node Send Queue For SVC, the average time in msec waiting in queue before
Time these send operations are executed.
Port to Remote Node Receive For SVC, the average time in msec waiting in queue before
Queue Time these send operations are executed.
Overall Port to Remote Node For SVC, the average time in msec waiting in queue before
Queue Time these send operations are executed.
Global Mirror Write I/O Rate For SVC, the rate in writes per second issued to the secondary
site for Global Mirror during the sample interval
Global Mirror Overlapping Write For SVC, the percentage of writes during the sample interval,
Percentage for which the write operations at the primary site for Global
Mirror have overlapping write domains.
Global Mirror Overlapping Write I/O For SVC, the average rate in writes per second during the
Rate sample interval, for which the write operations at the primary
site for Global Mirror have overlapping write domains.
Peak Read Response Time For SVC, the peak read response time in msec observed
during the sample interval. At the end of each sample interval,
this value is reset to zero.
Peak Write Response Time For SVC, the peak write response time in msec observed
during the sample interval. At the end of each sample interval,
this value is reset to zero.
By Node report
This is an SVC specific report
Critical and universal metrics are in bold. Less important ones in normal font and difficult
metrics to interpret are in italics.
Read I/O Rate (overall) Average number of read operations per second for the
sample interval. Applies to most subsystems.
Write I/O Rate (overall) Average number of write operations per second for the
sample interval. Applies to most subsystems.
Total I/O Rate (overall) Average number of reads and writes per second for the
sample interval. Applies to most subsystems.
Read Cache Hits Percentage Percentage of reads during the sample interval that are
(overall) found in cache. This is an important metric.
Write Cache Hits Percentage (overall) Percentage of writes that are handled in cache. This
number should be very nearly 100%
Total Cache Hits Percentage (overall) Weighted average of read cache hits and write cache hits.
Read Data Rate Average read data rate in megabytes per second during the
sample interval
Write Data Rate Average write data rate in megabytes per second during the
sample interval
Total Data Rate Average total (read + write) data rate in megabytes per
second during the sample interval.
Read Response Time Average response time in milliseconds for reads during the
sample interval. For this report, this is an average of read
hits in cache as well as read misses.
Write Response Time Average response time in milliseconds for writes during the
sample interval.
Overall Response Time Average response time in milliseconds for all I/O in the
sample interval, including both cache hits as well as misses
to backing storage if required.
Read Transfer Size Average transfer size in kilobytes for reads during the
sample interval
Write Transfer Size Average transfer size in kilobytes for writes during the
sample interval
Disk to Cache Transfer Rate Average number of track transfers per second from disk to
cache during the sample interval.
Cache to Disk Transfer Rate Average number of track transfers per second from cache
to disk during the sample interval.
Write-cache Delay Percentage Percentage of all I/O operations that were delayed due to
write-cache space constraints or other conditions during the
sample interval. Only writes can be delayed, but the
percentage is of all I/O.
Write-cache Delay I/O Rate The rate of I/O (actually writes) that are delayed during the
sample interval because of write cache.
Backend Read I/O Rate The average read rate in reads per second caused by read
misses. This is the read rate to the back-end storage for the
sample interval.
Backend Write I/O Rate The average write rate in writes per second caused by
front-end write activity. This is the write rate to the back-end
storage for the sample interval. These are logical writes
Total Backend I/O Rate The sum of Backend Read I/O Rate and Backend Write I/O
Rate over the sample interval.
Backend Read Data Rate Average number of megabytes per second read from
back-end storage during the sample interval.
Backend Write Data Rate Average number of megabytes per second written to
back-end storage during the sample interval.
Total Backend Data Rate Sum of the Backend Read and Write Data Rates for the
sample interval.
Backend Read Response Time Average response time in milliseconds for read operations
to the back-end storage.
Backend Write Response Time Average response time in milliseconds for write operations
to the back-end storage. This time can include several
physical I/O operations, depending on the type of RAID
architecture.
Overall Backend Response Time The weighted average of Backend read and write response
times during the sample interval.
Read Queue Time The average number of milliseconds that each read
operation during the sample interval spent on the queue
before being issued to the back-end storage device
Write Queue Time The average number of milliseconds that each write
operation during the sample interval spent on the queue
before being issued to the back-end storage device
Overall Queue Time The weighted average of Read Queue Time and Write
Queue Time during the sample interval.
Backend Read Transfer Size The average transfer size in kilobytes for reads to the
back-end storage during the sample interval.
Backend Write Transfer Size The average transfer size in kilobytes for data written to the
back-end storage during the sample interval.
Port Send I/O Rate The average rate per second for operations that send data
from an I/O port, typically to a server. This is typically a read
from the server's perspective.
Port Receive I/O Rate The average rate per second for operations where the
storage port receives data, typically from a server. This is
typically a write from the server's perspective.
Total Port I/O Rate Average read plus write I/O rate per second at the storage
port during the sample interval.
Port Send Data Rate The average data rate in megabytes per second for
operations that send data from an I/O port, typically to a
server.
Port Receive Data Rate The average data rate in megabytes per second for
operations where the storage port receives data, typically
from a server.
Total Port Data Rate Average read plus write data rate in megabytes per second
at the storage port during the sample interval
Readahead Percentage of Cache Hits An obscure measurement of cache hits involving data that
has been prestaged for one reason or another.
Dirty Write Percentage of Cache Hits The percentage of write cache hits which modified only data
that was already marked "dirty" in the cache; re-written
data. This is an obscure measurement of how effectively
writes are coalesced before destaging.
Write Cache Overflow Percentage For SVC the percentage of write operations that were
delayed due to lack of write-cache space during the sample
interval.
Write Cache Overflow I/O Rate For SVC, the average rate per second of write operations
that were delayed due to lack of write-cache space during
the sample interval.
Write Cache Flush-through For SVC, the percentage of write operations that were
Percentage processed in Flush-through write mode during the sample
interval.
Write Cache Flush-through I/O Rate For SVC, the average rate per second of tracks processed
in Flush-through write mode during the sample interval.
Write Cache Write-through For SVC the percentage of write operations that were
Percentage processed in Write-through write mode during the sample
interval.
Write Cache Write-through I/O Rate For SVC, the average number of tracks per second that
were processed in Write-through write mode during the
sample interval.
CPU Utilization Percentage The average utilization of the node controllers in this I/O
group during the sample interval.
Port to Host Send I/O Rate For SVC, the rate per second of port send to host (server)
during the sample interval.
Port to Host Receive I/O Rate For SVC, the rate per second of port receive operations
from host (server) during the sample interval.
Port to Disk Send I/O Rate For SVC, the rate per second of port send to back-end
storage during the sample interval.
Port to Disk Receive I/O Rate For SVC, the rate per second of port receive operations
from back-end storage during the sample interval.
Total Port to Disk I/O Rate For SVC, the sum of port to disk send and port to disk
receive rates during the sample interval.
Port to Local Node Send I/O Rate For SVC, the rate per second at which a port sends I/O to
other nodes in the local cluster during the sample interval.
Port to Local Node Receive I/O Rate For SVC, the rate at which a port receives I/O from other
nodes in the local cluster during the sample interval
Total Port to Local Node I/O Rate For SVC, the sum of port to local node send and receive
rates during the sample interval.
Port to Remote Node Send I/O Rate For SVC, the average number of exchanges (I/Os) per
second sent to nodes in the remote SVC cluster during the
sample interval. Typically some form of remote mirroring.
Port to Remote Node Receive I/O For SVC, the average number of exchanges (I/Os) per
Rate second received from nodes in the remote SVC cluster
during the sample interval. Typically some form of remote
mirroring.
Total Port to Remote Node I/O Rate For SVC, the sum of port to remote node send and receive
I/O per second during the sample interval.
Port to Host Send Data Rate For SVC, the megabytes per second of port send to host
(server) during the sample interval.
Port to Host Receive Data Rate For SVC, the megabytes per second of port receive
operations from host (server) during the sample interval.
Total Port to Host Data Rate For SVC, total of port send and receive megabytes per
second during the sample interval.
Port to Disk Send Data Rate For SVC, the megabytes per second of port send to
back-end storage during the sample interval.
Port to Disk Receive Data Rate For SVC, the megabytes per second of port receive
operations from back-end storage during the sample
interval
Total Port to Disk Data Rate For SVC, the sum of port to disk send and port to disk
receive megabytes per second during the sample interval.
Port to Local Node Send Data Rate For SVC, the megabytes per second at which a port sends
I/O to other nodes in the local cluster during the sample
interval.
Port to Local Node Receive Data Rate For SVC, the megabytes per second at which a port
receives I/O from other nodes in the local cluster during the
sample interval
Total Port to Local Node Data Rate For SVC, the sum of port to local node send and receive
megabytes per second during the sample interval.
Port to Remote Node Receive Data For SVC, the average number of megabytes per second
Rate received from nodes in the remote SVC cluster during the
sample interval. Typically some form of remote mirroring.
Total Port to Remote Node Data Rate For SVC, the sum of port to remote node send and receive
megabytes per second during the sample interval.
Port to Local Node Send Response For SVC the average port service time in milliseconds for
Time these operations during the sample interval.
Port to Local Node Receive Response For SVC the average port service time in milliseconds for
Time this operation during the sample interval.
Overall Port to Local Node Response For SVC the average port service time in milliseconds for
Time these operations during the sample interval.
Port to Local Node Send Queue Time For SVC, the average time in msec waiting in queue before
these send operations are executed.
Port to Local Node Receive Queue For SVC, the average time in msec waiting in queue before
Time these receive operations are executed.
Overall Port to Local Node Queue For SVC, the average time in msec. waiting before these
Time port send or receive operations are executed.
Port to Remote Node Send Response For SVC the average port service time in milliseconds for
Time these operations during the sample interval.
Port to Remote Node Receive For SVC the average port service time in milliseconds for
Response Time these operations during the sample interval.
Overall Port to Remote Node For SVC the average port service time in milliseconds for
Response Time these operations during the sample interval.
Port to Remote Node Send Queue For SVC, the average time in msec waiting in queue before
Time these send operations are executed.
Port to Remote Node Receive Queue For SVC, the average time in msec waiting in queue before
Time these send operations are executed.
Overall Port to Remote Node Queue For SVC, the average time in msec waiting in queue before
Time these send operations are executed.
Global Mirror Write I/O Rate For SVC, the rate in writes per second issued to the
secondary site for Global Mirror during the sample interval
Global Mirror Overlapping Write For SVC, the percentage of writes during the sample
Percentage interval, for which the write operations at the primary site for
Global Mirror have overlapping write domains.
Global Mirror Overlapping Write I/O For SVC, the average rate in writes per second during the
Rate sample interval, for which the write operations at the
primary site for Global Mirror have overlapping write
domains.
Peak Read Response Time For SVC, the peak read response time in msec observed
during the sample interval. At the end of each sample
interval, this value is reset to zero.
Global Mirror Secondary Write Lag For SVC, the number of additional milliseconds it took to
service each secondary write operation for Global Mirror,
over and above the time needed to service the primary
writes during the sample interval.
By Array report
This report is for DS8000, DS6000, and ESS only.
Critical and universal metrics are in bold. Less important ones in normal font and difficult
metrics to interpret are in italics.
Subsystem Component id
Read I/O Rate (normal) Average number of normal read operations per second for
the sample interval. Normal operations are not sequential,
hence random. This metric is only for ESS, DS8000,
DS6000
Read I/O Rate (sequential) Average number of sequential read operations per second
for the sample interval. Sequential I/O is detected by the
subsystem. This metric is only for ESS, DS8000, DS6000
Read I/O Rate (overall) Average number of read operations per second for the
sample interval. Applies to most subsystems.
Write I/O Rate (normal) Average number of normal write operations per second for
the sample interval. Normal operations are not sequential,
hence random. This metric is only for ESS, DS8000,
DS6000
Write I/O Rate (sequential) Average number of sequential write operations per
second for the sample interval. Sequential I/O is detected
by the subsystem. This metric is only for ESS, DS8000,
DS6000
Write I/O Rate (overall) Average number of write operations per second for the
sample interval. Applies to most subsystems.
Total I/O Rate (normal) Average number of normal reads and writes per second
for the sample interval. Normal operations are not
sequential, hence random. This metric is only for ESS,
DS8000, DS6000
Total I/O Rate (sequential) Average number of sequential reads and writes per
second for the sample interval. Sequential I/O is detected
by the subsystem. This metric is only for ESS, DS8000,
DS6000
Total I/O Rate (overall) Average number of reads and writes per second for the
sample interval. Applies to most subsystems.
Read Cache Hits Percentage Percentage of sequential reads that are cache hits in the
(sequential) sample interval. Only for ESS, DS8000, DS6000.
Read Cache Hits Percentage (overall) Percentage of reads during the sample interval that are
found in cache. This is an important metric.
Write Cache Hits Percentage (normal) Percentage of normal (random) writes that are handled in
cache. This number should be 100%. Only for ESS,
DS8K, DS6K
Write Cache Hits Percentage Percentage of sequential writes that are handled in cache.
(sequential) This number should be 100%. Only for ESS, DS8K, DS6K
Write Cache Hits Percentage (overall) Percentage of writes that are handled in cache. This
number should be 100% for most enterprise storage.
Total Cache Hits Percentage (normal) Percentage of normal reads and writes that are cache hits
during the sample interval.
Total Cache Hits Percentage Percentage of sequential reads and writes that are cache
(sequential) hits during the sample interval.
Total Cache Hits Percentage (overall) Weighted average of read cache hits and write cache hits.
Read Data Rate Average read data rate in megabytes per second during
the sample interval
Write Data Rate Average write data rate in megabytes per second during
the sample interval
Total Data Rate Average total (read + write) data rate in megabytes per
second during the sample interval.
Read Response Time Average response time in milliseconds for reads during
the sample interval. For this report, this is an average of
read hits in cache as well as read misses.
Write Response Time Average response time in milliseconds for writes during
the sample interval.
Overall Response Time Average response time in milliseconds for all I/O in the
sample interval, including both cache hits as well as
misses to backing storage if required.
Read Transfer Size Average transfer size in kilobytes for reads during the
sample interval
Write Transfer Size Average transfer size in kilobytes for writes during the
sample interval
Overall Transfer Size Average transfer size in kilobytes for all I/O during the
sample interval.
Record Mode Read I/O Rate This is the rate in I/O/sec for a special kind of read activity
detected by ESS, DS8K and DS6K. Only the requested
data is managed in cache rather than a full track or most
of a track of data.
Record Mode Read Cache Hit Read Hit percentage for the special class of reads
Percentage mentioned above. ESS, DS8k, DS6K only.
Cache to Disk Transfer Rate Average number of track transfers per second from cache
to disk during the sample interval.
Write-cache Delay Percentage Percentage of all I/O operations that were delayed due to
write-cache space constraints or other conditions during
the sample interval. Only writes can be delayed, but the
percentage is of all I/O.
Write-cache Delay I/O Rate The rate of I/O (actually writes) that are delayed during the
sample interval because of write cache.
Backend Read I/O Rate The average read rate in reads per second caused by
read misses. This is the read rate to the back-end storage
for the sample interval.
Backend Write I/O Rate The average write rate in writes per second caused by
front-end write activity. This is the write rate to the
back-end storage for the sample interval. These are
logical writes and the actual number of physical I/O
operations depends on whether the storage is RAID 5,
RAID 10, or some other architecture.
Total Backend I/O Rate The sum of Backend Read I/O Rate and Backend Write
I/O Rate over the sample interval.
Backend Read Data Rate Average number of megabytes per second read from
back-end storage during the sample interval.
Backend Write Data Rate Average number of megabytes per second written to
back-end storage during the sample interval.
Total Backend Data Rate Sum of the Backend Read and Write Data Rates for the
sample interval.
Backend Read Response Time Average response time in milliseconds for read operations
to the back-end storage.
Backend Write Response Time Average response time in milliseconds for write operations
to the back-end storage. This time can include several
physical I/O operations, depending on the type of RAID
architecture.
Overall Backend Response Time The weighted average of Backend read and write
response times during the sample interval.
Backend Read Transfer Size The average transfer size in kilobytes for reads to the
back-end storage during the sample interval.
Backend Write Transfer Size The average transfer size in kilobytes for data written to
the back-end storage during the sample interval.
Overall Backend Transfer Size Weighted average transfer size in kilobytes for backend
reads and writes during the sample interval.
Disk Utilization Percentage Average disk utilization during the sample interval. This is
also the utilization of the raid array, since the activity is
uniform across the array.
Sequential I/O Percentage Percentage of the I/O during the sample interval which the
storage believes to be sequential. This is detected by the
storage algorithms.
Critical and universal metrics are in bold. Less important ones in normal font and difficult
metrics to interpret are in italics.
Read I/O Rate (overall) Average number of read operations per second for the
sample interval. Applies to most subsystems.
Write I/O Rate (overall) Average number of write operations per second for the
sample interval. Applies to most subsystems.
Total I/O Rate (overall) Average number of reads and writes per second for the
sample interval. Applies to most subsystems.
Read Data Rate Average read data rate in megabytes per second during
the sample interval
Write Data Rate Average write data rate in megabytes per second during
the sample interval
Total Data Rate Average total (read + write) data rate in megabytes per
second during the sample interval.
Read Response Time Average response time in milliseconds for reads during the
sample interval. For this report, this is an average of read
hits in cache as well as read misses.
Write Response Time Average response time in milliseconds for writes during the
sample interval.
Overall Response Time Average response time in milliseconds for all I/O in the
sample interval, including both cache hits as well as
misses to backing storage if required.
Read Transfer Size Average transfer size in kilobytes for reads during the
sample interval
Write Transfer Size Average transfer size in kilobytes for writes during the
sample interval
Overall Transfer Size Average transfer size in kilobytes for all I/O during the
sample interval.
Backend Read I/O Rate The average read rate in reads per second caused by read
misses. This is the read rate to the back-end storage for the
sample interval.
Backend Write I/O Rate The average write rate in writes per second caused by
front-end write activity. This is the write rate to the
back-end storage for the sample interval. These are logical
writes
Backend Read Data Rate Average number of megabytes per second read from
back-end storage during the sample interval.
Backend Write Data Rate Average number of megabytes per second written to
back-end storage during the sample interval.
Total Backend Data Rate Sum of the Backend Read and Write Data Rates for the
sample interval.
Backend Read Response Time Average response time in milliseconds for read operations
to the back-end storage.
Backend Write Response Time Average response time in milliseconds for write operations
to the back-end storage. This time can include several
physical I/O operations, depending on the type of RAID
architecture.
Overall Backend Response Time The weighted average of Backend read and write response
times during the sample interval.
Read Queue Time The average number of milliseconds that each read
operation during the sample interval spent on the queue
before being issued to the back-end storage device
Write Queue Time The average number of milliseconds that each write
operation during the sample interval spent on the queue
before being issued to the back-end storage device
Overall Queue Time The weighted average of Read Queue Time and Write
Queue Time during the sample interval.
Backend Read Transfer Size The average transfer size in kilobytes for reads to the
back-end storage during the sample interval.
Backend Write Transfer Size The average transfer size in kilobytes for data written to the
back-end storage during the sample interval.
Overall Backend Transfer Size Weighted average transfer size in kilobytes for backend
reads and writes during the sample interval.
Critical and universal metrics are in bold. Less important ones in normal font and difficult
metrics to interpret are in italics.
Volume Volume Id
Read I/O Rate (normal) Average number of normal read operations per second for
the sample interval. Normal operations are not sequential,
hence random. This metric is only for ESS, DS8000,
DS6000
Read I/O Rate (sequential) Average number of sequential read operations per second
for the sample interval. Sequential I/O is detected by the
subsystem. This metric is only for ESS, DS8000, DS6000
Read I/O Rate (overall) Average number of read operations per second for the
sample interval. Applies to most subsystems.
Write I/O Rate (normal) Average number of normal write operations per second for
the sample interval. Normal operations are not sequential,
hence random. This metric is only for ESS, DS8000,
DS6000
Write I/O Rate (sequential) Average number of sequential write operations per second
for the sample interval. Sequential I/O is detected by the
subsystem. This metric is only for ESS, DS8000, DS6000
Write I/O Rate (overall) Average number of write operations per second for the
sample interval. Applies to most subsystems.
Total I/O Rate (normal) Average number of normal reads and writes per second for
the sample interval. Normal operations are not sequential,
hence random. This metric is only for ESS, DS8000,
DS6000
Total I/O Rate (sequential) Average number of sequential reads and writes per
second for the sample interval. Sequential I/O is detected
by the subsystem. This metric is only for ESS, DS8000,
DS6000
Total I/O Rate (overall) Average number of reads and writes per second for the
sample interval. Applies to most subsystems.
Read Cache Hit Percentage (normal) Percentage of normal (random) reads that are cache hits
during the sample interval. Only for ESS, DS8000,
DS6000
Read Cache Hits Percentage Percentage of sequential reads that are cache hits in the
(sequential) sample interval. Only for ESS, DS8000, DS6000.
Read Cache Hits Percentage (overall) Percentage of reads during the sample interval that are
found in cache. This is an important metric.
Write Cache Hits Percentage Percentage of sequential writes that are handled in cache.
(sequential) This number should be 100%. Only for ESS, DS8K, DS6K
Write Cache Hits Percentage (overall) Percentage of writes that are handled in cache. This
number should be 100% for most enterprise storage.
Total Cache Hits Percentage (normal) Percentage of normal reads and writes that are cache hits
during the sample interval.
Total Cache Hits Percentage Percentage of sequential reads and writes that are cache
(sequential) hits during the sample interval.
Total Cache Hits Percentage (overall) Weighted average of read cache hits and write cache hits.
Read Data Rate Average read data rate in megabytes per second during
the sample interval
Write Data Rate Average write data rate in megabytes per second during
the sample interval
Total Data Rate Average total (read + write) data rate in megabytes per
second during the sample interval.
Read Response Time Average response time in milliseconds for reads during the
sample interval. For this report, this is an average of read
hits in cache as well as read misses.
Write Response Time Average response time in milliseconds for writes during
the sample interval.
Overall Response Time Average response time in milliseconds for all I/O in the
sample interval, including both cache hits as well as
misses to backing storage if required.
Read Transfer Size Average transfer size in kilobytes for reads during the
sample interval
Write Transfer Size Average transfer size in kilobytes for writes during the
sample interval
Overall Transfer Size Average transfer size in kilobytes for all I/O during the
sample interval.
Record Mode Read I/O Rate This is the rate in I/O/sec for a special kind of read activity
detected by ESS, DS8K and DS6K. Only the requested
data is managed in cache rather than a full track or most of
a track of data.
Record Mode Read Cache Hit Read Hit percentage for the special class of reads
Percentage mentioned above. ESS, DS8k, DS6K only.
Disk to Cache Transfer Rate Average number of track transfers per second from disk to
cache during the sample interval.
Cache to Disk Transfer Rate Average number of track transfers per second from cache
to disk during the sample interval.
Write-cache Delay Percentage Percentage of all I/O operations that were delayed due to
write-cache space constraints or other conditions during
the sample interval. Only writes can be delayed, but the
percentage is of all I/O.
Readahead Percentage of Cache Hits For SVC, an obscure measurement of cache hits involving
data that has been prestaged for one reason or another.
Dirty Write Percentage of Cache Hits For SVC, the percentage of write cache hits which
modified only data that was already marked "dirty" in the
cache; re-written data. This is an obscure measurement of
how effectively writes are coalesced before destaging.
Write Cache Overflow Percentage For SVC the percentage of write operations that were
delayed due to lack of write-cache space during the
sample interval.
Write Cache Overflow I/O Rate For SVC, the average rate per second of write operations
that were delayed due to lack of write-cache space during
the sample interval.
Write Cache Flush-through Percentage For SVC, the percentage of write operations that were
processed in Flush-through write mode during the sample
interval.
Write Cache Flush-through I/O Rate For SVC, the average rate per second of tracks processed
in Flush-through write mode during the sample interval.
Write Cache Write-through Percentage For SVC the percentage of write operations that were
processed in Write-through write mode during the sample
interval.
Write Cache Write-through I/O Rate For SVC, the average number of tracks per second that
were processed in Write-through write mode during the
sample interval.
Global Mirror Write I/O Rate For SVC, the rate in writes per second issued to the
secondary site for Global Mirror during the sample interval
Global Mirror Overlapping Write For SVC, the percentage of writes during the sample
Percentage interval, for which the write operations at the primary site
for Global Mirror have overlapping write domains.
Global Mirror Overlapping Write I/O For SVC, the average rate in writes per second during the
Rate sample interval, for which the write operations at the
primary site for Global Mirror have overlapping write
domains.
Peak Read Response Time For SVC, the peak read response time in msec observed
during the sample interval. At the end of each sample
interval, this value is reset to zero.
Peak Write Response Time For SVC, the peak write response time in msec observed
during the sample interval. At the end of each sample
interval, this value is reset to zero.
Global Mirror Secondary Write Lag For SVC, the number of additional milliseconds it took to
service each secondary write operation for Global Mirror,
over and above the time needed to service the primary
writes during the sample interval.
Critical and universal metrics are in bold. Less important ones in normal font and difficult
metrics to interpret are in italics.
Read I/O Rate (overall) Average number of read operations per second for the
sample interval. Applies to most subsystems.
Write I/O Rate (overall) Average number of write operations per second for the
sample interval. Applies to most subsystems.
Total I/O Rate (overall) Average number of reads and writes per second for the
sample interval. Applies to most subsystems.
Backend Read Data Rate Average number of megabytes per second read from
back-end storage during the sample interval.
Backend Write Data Rate Average number of megabytes per second written to
back-end storage during the sample interval.
Total Backend Data Rate Sum of the Backend Read and Write Data Rates for the
sample interval.
Backend Read Response Time Average response time in milliseconds for read operations
to the back-end storage.
Backend Write Response Time Average response time in milliseconds for write operations
to the back-end storage. This time can include several
physical I/O operations, depending on the type of RAID
architecture.
Overall Backend Response Time The weighted average of Backend read and write
response times during the sample interval.
Read Queue Time The average number of milliseconds that each read
operation during the sample interval spent on the queue
before being issued to the back-end storage device
Write Queue Time The average number of milliseconds that each write
operation during the sample interval spent on the queue
before being issued to the back-end storage device
Overall Queue Time The weighted average of Read Queue Time and Write
Queue Time during the sample interval.
Backend Read Transfer Size The average transfer size in kilobytes for reads to the
back-end storage during the sample interval.
Backend Write Transfer Size The average transfer size in kilobytes for data written to
the back-end storage during the sample interval.
Overall Backend Transfer Size Weighted average transfer size in kilobytes for backend
reads and writes during the sample interval.
Port Port ID
Port Send I/O Rate Average number of I/O operations per second for send
operations, for a particular port during the sample
interval
Port Receive I/O Rate Average number of I/O operations per second for
receive operations, for a particular port during the
sample interval
Total Port I/O Rate Average number of I/O operations per second for send
and receive operations, for a particular port during the
sample interval
Port Send Data Rate Average number of megabytes per second that were
transferred for send (server read) operations, for a
particular port during the sample interval
Port Receive Data Rate Average number of megabytes per second that were
transferred for receive (server write) operations, for a
particular port during the sample interval
Total Port Data Rate Average number of megabytes per second for send and
receive operations during the sample interval.
Port Send Response Time Average number of milliseconds that it took to service
each send (server read) operation during the sample
interval
Port Receive Response Time Average number of milliseconds that it took to service
each receive (server write) operation during the sample
interval
Total Port Response Time Average number of milliseconds that it took to service
each send and receive operation during the sample
interval
Port Send Transfer Size Average number of KB sent per I/O by a particular port
Port Receive Transfer Size Average number of KB received per I/O by a particular
port during the sample interval
Port to Host Send I/O Rate The rate per second of port send to host (server) during
the sample interval.
Port to Host Receive I/O Rate The rate per second of port receive operations from host
(server) during the sample interval.
Total Port to Host I/O Rate The total of port send and receive I/O rate during the
sample interval.
Port to Disk Receive I/O Rate For SVC, the rate per second of port receive operations
from back-end storage during the sample interval.
Total Port to Disk I/O Rate For SVC, the sum of port to disk send and port to disk
receive rates during the sample interval.
Port to Local Node Send I/O Rate For SVC, the rate per second at which a port sends I/O
to other nodes in the local cluster during the sample
interval.
Port to Local Node Receive I/O Rate For SVC, the rate at which a port receives I/O from other
nodes in the local cluster during the sample interval
Total Port to Local Node I/O Rate For SVC, the sum of port to local node send and receive
rates during the sample interval.
Port to Remote Node Send I/O Rate For SVC, the average number of exchanges (I/Os) per
second sent to nodes in the remote SVC cluster during
the sample interval. Typically some form of remote
mirroring.
Port to Remote Node Receive I/O Rate For SVC, the average number of exchanges (I/Os) per
second received from nodes in the remote SVC cluster
during the sample interval. Typically some form of
remote mirroring.
Total Port to Remote Node I/O Rate For SVC, the sum of port to remote node send and
receive I/O per second during the sample interval.
Port to Host Send Data Rate For SVC, the megabytes per second of port send to host
(server) during the sample interval.
Port to Host Receive Data Rate For SVC, the megabytes per second of port receive
operations from host (server) during the sample
interval.
Total Port to Host Data Rate For SVC, total of port send and receive megabytes per
second during the sample interval.
Port to Disk Send Data Rate For SVC, the megabytes per second of port send to
back-end storage during the sample interval.
Port to Disk Receive Data Rate For SVC, the megabytes per second of port receive
operations from back-end storage during the sample
interval
Total Port to Disk Data Rate For SVC, the sum of port to disk send and port to disk
receive megabytes per second during the sample
interval.
Port to Local Node Send Data Rate For SVC, the megabytes per second at which a port
sends I/O to other nodes in the local cluster during the
sample interval.
Port to Local Node Receive Data Rate For SVC, the megabytes per second at which a port
receives I/O from other nodes in the local cluster during
the sample interval
Total Port to Local Node Data Rate For SVC, the sum of port to local node send and receive
megabytes per second during the sample interval.
Port to Remote Node Receive Data Rate For SVC, the average number of megabytes per second
received from nodes in the remote SVC cluster during
the sample interval. Typically some form of remote
mirroring.
Total Port to Remote Node Data Rate For SVC, the sum of port to remote node send and
receive megabytes per second during the sample
interval.
Port Port ID
Port Send Packet Rate Average number of packets per second for send
operations, for a particular port during the sample interval.
Port Receive Packet Rate Average number of packets per second for receive
operations, for a particular port during the sample interval.
Total Port Packet Rate Average number of packets per second for send and
receive operations, for a particular port during the sample
interval.
Port Send Data Rate Average number of megabytes (2^20 bytes) per second
that were transferred for send (write) operations, for a
particular port during the sample interval.
Port Receive Data Rate Average number of megabytes (2^20 bytes) per second
that were transferred for receive (read) operations, for a
particular port during the sample interval.
Total Port Data Rate Average number of megabytes (2^20 bytes) per second
that were transferred for send and receive operations, for a
particular port during the sample interval.
Port Peak Send Data Rate Peak number of megabytes (2^20 bytes) per second that
were sent by a particular port during the sample interval.
Port Peak Receive Data Rate Peak number of megabytes (2^20 bytes) per second that
were received by a particular port during the sample
interval.
Port Send Packet Size Average number of KB sent per packet by a particular port
during the sample interval.
Error Frame Rate The average number of frames per second that were
received in error during the sample interval.
Dumped Frame Rate The average number of frames per second that were lost
due to a lack of available host buffers during the sample
interval.
Link Failure Rate The average number of link errors per second during the
sample interval.
Loss of Sync Rate The average number of times per second that
synchronization was lost during the sample interval.
Loss of Signal Rate The average number of times per second that the signal
was lost during the sample interval.
CRC Error Rate The average number of frames received per second in
which the CRC in the frame did not match the CRC
computed by the receiver during the sample interval.
Short Frame Rate The average number of frames received per second that
were shorter than 28 octets (24 header + 4 CRC) not
including any SOF/EOF bytes during the sample interval.
Long Frame Rate The average number of frames received per second that
were longer than 2140 octets (24 header + 4 CRC + 2112
data) not including any SOF/EOF bytes during the sample
interval.
Encoding Disparity Error Rate The average number of disparity errors received per
second during the sample interval.
Discarded Class3 Frame Rate The average number of class-3 frames per second that
were discarded during the sample interval.
F-BSY Frame Rate The average number of F-BSY frames per second during
the sample interval.
F-RJT Frame Rate The average number of F-RJT frames per second during
the sample interval.
The publications listed in this section are considered particularly suitable for a more detailed
discussion of the topics covered in this book.
IBM Redbooks
For information about ordering these publications, see “How to get Redbooks” on page 349.
Note that some of the documents referenced here might be available in softcopy only.
IBM TotalStorage Productivity Center: The Next Generation, SG24-7194
Other publications
These publications are also relevant as further information sources:
IBM TotalStorage Productivity Center for Data Installation and Configuration Guide for
V2.3, GC32-1727
IBM TotalStorage Productivity Center Version 3.3 User’s Guide, GC32-1775
Online resources
These Web sites are also relevant as further information sources:
TPC V3.3 support products list:
http://www-1.ibm.com/support/docview.wss?rs=597&uid=ssg1S1003019/
Prerequisites:
http://www-304.ibm.com/jct01004c/systems/support/supportsite.wss/supportresourc
es?taskind=3&brandind=5000033&familyind=5329731
developerWorks Web site:
http://www.ibm.com/developerworks/java/jdk/linux/
VMWare site:
http://www.vmware.com
N
I N Series support 157
IBM FAStT 309 NAS
IBM TS3310 Tape Library 215 filer login ID 166
IBMCDB common agent database 116 monitoring options 159
install sequence 18 Unix proxy agent requirements 159
installation Windows proxy agent 160
Internet Information Services 21 NAS CIFS share 162
installation licenses 58 NAS device
Internet Information Services 21 license 165
Intrepid 10000 Director 13 Probe job 171
IP address 310 quotas 188
scan and probe 169
NAS devices
K Discovery 162
keytool command 280
manage through Discovery 175
TPC dashboard 171
L NAS filer
license NAS devices 165 asset information 184
Linux install 73 CIFS share 161
Agent Manger 95 member of Windows domain 160
database user information 99 SNMP community name 164
DB2 UDB connection 98 Topology Viewer 182
db2profile script 95 NAS filer data
IBMCDB database 98 TPC dashboard 180
Linux platform NAS filesystems reports 185
DB2 upgrade 136 NAS support
DB2 V8 75 SNMP queries 158
List DB2 databases 139 NetApp Data ONTAP 182
log files NetApp NAS support 157
CLI 71 NETBIOS 20
Device server 71
GUI 71
P
Password gathering 129
M Path Planner 9, 246
master TPC server 266 SDD 246
McDATA interface performance counters 316
configuring 196 performance metrics 316
Direct Connection mode 191 -1 value 316
EFCM Proxy mode 191 essential 316
TPC connection 198 pin list persistence 5, 224
McDATA OPENconnectors SMI-S Interface 190 pin lists
EFCM proxy mode 191 multiple users 224
supported platforms 190 pinning 224
McData OPENconnectors SMI-S Interface planner wizards 9
user guide 193 probe 269
McData provider 198 Probe job status 171
Index 353
VMWare Virtual Infrastructrue
planning 279
VMWare virtual machine
Fabric agent 279
VMWare VirtualCenter 278
Volume Performance Advisor 8, 246
Volume Placement Advisor 233
Volume Planner 8, 246
W
Windows Proxy Agent 160
workload profile 257
X
X11 forwarding 295
putty installation 304
Z
Zone Planner 9, 247
zone planning 260
TotalStorage
Productivity Center V3.3
Update Guide ®