Documente Academic
Documente Profesional
Documente Cultură
ibm.com/redbooks
International Technical Support Organization IBM TotalStorage Productivity Center V2.3: Getting Started December 2005
SG24-6490-01
Note: Before using this information and the product it supports, read the information in Notices on page xiii.
Second Edition (December 2005) This edition applies to Version 2, Release 3 of IBM TotalStorage Productivity Center (product number 5608-UC1, 5608-UC3, 5608-UC4, 5608-UC5.
Copyright International Business Machines Corporation 2005. All rights reserved. Note to U.S. Government Users Restricted Rights -- Use, duplication or disclosure restricted by GSA ADP Schedule Contract with IBM Corp.
Contents
Notices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xiii Trademarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xiv Preface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xv The team that wrote this redbook. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xv Become a published author . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xvii Comments welcome. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xvii Part 1. IBM TotalStorage Productivity Center foundation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1 Chapter 1. IBM TotalStorage Productivity Center overview . . . . . . . . . . . . . . . . . . . . . . 3 1.1 Introduction to IBM TotalStorage Productivity Center . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 1.1.1 Standards organizations and standards . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 1.2 IBM TotalStorage Open Software family . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5 1.3 IBM TotalStorage Productivity Center . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6 1.3.1 Data subject matter expert: TotalStorage Productivity Center for Data . . . . . . . . . 7 1.3.2 Fabric subject matter expert: Productivity Center for Fabric . . . . . . . . . . . . . . . . . . 9 1.3.3 Disk subject matter expert: TotalStorage Productivity Center for Disk . . . . . . . . . 12 1.3.4 Replication subject matter expert: Productivity Center for Replication . . . . . . . . . 14 1.4 IBM TotalStorage Productivity Center . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16 1.4.1 Productivity Center for Disk and Productivity Center for Replication . . . . . . . . . . 17 1.4.2 Event services. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23 1.5 Taking steps toward an On Demand environment . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24 Chapter 2. Key concepts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.1 IBM TotalStorage Productivity Center architecture . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.1.1 Architectural overview diagram . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.1.2 Architectural layers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.1.3 Relationships between the managers and components . . . . . . . . . . . . . . . . . . . . 2.1.4 Collecting data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2 Standards used in IBM TotalStorage Productivity Center . . . . . . . . . . . . . . . . . . . . . . . 2.2.1 ANSI standards . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2.2 Web-Based Enterprise Management . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2.3 Storage Networking Industry Association . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2.4 Simple Network Management Protocol. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2.5 Fibre Alliance MIB . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.3 Service Location Protocol (SLP) overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.3.1 SLP architecture . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.3.2 Common Information Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.4 Component interaction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.4.1 CIMOM discovery with SLP . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.4.2 How CIM Agent works. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.5 Tivoli Common Agent Services . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.5.1 Tivoli Agent Manager . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.5.2 Common Agent . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27 28 28 29 31 32 34 34 34 35 36 37 38 38 47 49 49 50 51 53 53
Part 2. Installing the IBM TotalStorage Productivity Center base product suite . . . . . . . . . . . . . . . . . 55 Chapter 3. Installation planning and considerations. . . . . . . . . . . . . . . . . . . . . . . . . . . 57
iii
3.1 Configuration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.2 Installation prerequisites . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.2.1 TCP/IP ports used by TotalStorage Productivity Center . . . . . . . . . . . . . . . . . . . . 3.2.2 Default databases created during the installation . . . . . . . . . . . . . . . . . . . . . . . . . 3.3 Our lab setup environment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.4 Pre-installation check list. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.5 User IDs and security . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.5.1 User IDs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.5.2 Increasing user security . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.5.3 Certificates and key files . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.5.4 Services and service accounts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.6 Starting and stopping the managers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.7 Windows Management Instrumentation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.8 World Wide Web Publishing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.9 Uninstalling Internet Information Services. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.10 Installing SNMP. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.11 IBM TotalStorage Productivity Center for Fabric. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.11.1 The computer name . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.11.2 Database considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.11.3 Windows Terminal Services . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.11.4 Tivoli NetView . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.11.5 Personal firewall . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.11.6 Changing the HOSTS file . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.12 IBM TotalStorage Productivity Center for Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.12.1 Server recommendations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.12.2 Supported subsystems and databases . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.12.3 Security considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.12.4 Creating the DB2 database. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
58 58 59 62 62 64 65 65 68 69 69 70 70 73 73 73 75 75 75 75 76 77 77 78 78 78 79 81
Chapter 4. Installing the IBM TotalStorage Productivity Center suite . . . . . . . . . . . . . 83 4.1 Installing the IBM TotalStorage Productivity Center . . . . . . . . . . . . . . . . . . . . . . . . . . . 84 4.1.1 Considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 84 4.2 Prerequisite Software Installation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 85 4.2.1 Best practices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 85 4.2.2 Installing prerequisite software . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 85 4.3 Suite installation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 110 4.3.1 Best practices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 110 4.3.2 Installing the TotalStorage Productivity Center suite . . . . . . . . . . . . . . . . . . . . . 110 4.3.3 IBM TotalStorage Productivity Center for Disk and Replication Base. . . . . . . . . 125 4.3.4 IBM TotalStorage Productivity Center for Disk . . . . . . . . . . . . . . . . . . . . . . . . . . 140 4.3.5 IBM TotalStorage Productivity Center for Replication. . . . . . . . . . . . . . . . . . . . . 146 4.3.6 IBM TotalStorage Productivity Center for Fabric. . . . . . . . . . . . . . . . . . . . . . . . . 157 4.3.7 IBM TotalStorage Productivity Center for Data . . . . . . . . . . . . . . . . . . . . . . . . . . 171 Chapter 5. CIMOM install and configuration. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.2 Planning considerations for Service Location Protocol . . . . . . . . . . . . . . . . . . . . . . . . 5.2.1 Considerations for using SLP DAs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.2.2 SLP configuration recommendation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.3 General performance guidelines . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.4 Planning considerations for CIMOM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.4.1 CIMOM configuration recommendations. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.5 Installing CIM agent for ESS . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 191 192 192 192 193 194 194 195 196
iv
5.5.1 ESS CLI Install . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.5.2 DS CIM Agent install . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.5.3 Post Installation tasks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.6 Configuring the DS CIM Agent for Windows . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.6.1 Registering DS Devices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.6.2 Registering ESS Devices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.6.3 Register ESS server for Copy services. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.6.4 Restart the CIMOM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.6.5 CIMOM user authentication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.7 Verifying connection to the ESS . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.7.1 Problem determination . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.7.2 Confirming the ESS CIMOM is available . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.7.3 Setting up the Service Location Protocol Directory Agent . . . . . . . . . . . . . . . . . 5.7.4 Configuring TotalStorage Productivity Center for SLP discovery . . . . . . . . . . . . 5.7.5 Registering the DS CIM Agent to SLP . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.7.6 Verifying and managing CIMOMs availability. . . . . . . . . . . . . . . . . . . . . . . . . . . 5.8 Installing CIM agent for IBM DS4000 family . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.8.1 Verifying and Managing CIMOM availability . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.9 Configuring CIMOM for SAN Volume Controller . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.9.1 Adding the SVC TotalStorage Productivity Center for Disk user account. . . . . . 5.9.2 Registering the SAN Volume Controller host in SLP . . . . . . . . . . . . . . . . . . . . . 5.10 Configuring CIMOM for TotalStorage Productivity Center for Disk summary . . . . . . 5.10.1 SLP registration and slptool . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.10.2 Persistency of SLP registration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.10.3 Configuring slp.reg file . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
196 202 211 212 212 213 214 215 215 216 219 220 221 223 224 224 225 233 234 235 241 241 242 243 243
Part 3. Configuring the IBM TotalStorage Productivity Center. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 245 Chapter 6. Configuring IBM TotalStorage Productivity Center for Disk. . . . . . . . . . . 6.1 Productivity Center for Disk Discovery summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.2 SLP DA definition . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.2.1 Verifying and managing CIMOMs availability . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.3 Disk and Replication Manager remote GUI. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6.3.1 Installing Remote Console for Performance Manager function. . . . . . . . . . . . . . 6.3.2 Launching Remote Console for TotalStorage Productivity Center . . . . . . . . . . . 247 248 248 256 259 270 277
Chapter 7. Configuring TotalStorage Productivity Center for Replication . . . . . . . . 279 7.1 Installing a remote GUI and CLI . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 280 Chapter 8. Configuring IBM TotalStorage Productivity Center for Data . . . . . . . . . . 8.1 Configuring the CIM Agents . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.1.1 CIM and SLP interfaces within Data Manager . . . . . . . . . . . . . . . . . . . . . . . . . . 8.1.2 Configuring CIM Agents . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.1.3 Setting up a disk alias . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.2 Setting up the Web GUI . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.2.1 Using IBM HTTP Server . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.2.2 Using Internet Information Server . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.2.3 Configuring the URL in Fabric Manager . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.3 Installing the Data Manager remote console. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.4 Configuring Data Manager for Databases. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.5 Alert Disposition . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 289 290 290 290 293 295 295 299 303 304 313 316
Chapter 9. Configuring IBM TotalStorage Productivity Center for Fabric . . . . . . . . . 319 9.1 TotalStorage Productivity Center component interaction . . . . . . . . . . . . . . . . . . . . . . 320
Contents
9.1.1 IBM TotalStorage Productivity Center for Disk and Replication Base. . . . . . . . . 9.1.2 SNMP . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.1.3 Tivoli Provisioning Manager . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.2 Post-installation procedures . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.2.1 Installing Productivity Center for Fabric Agent . . . . . . . . . . . . . . . . . . . . . . . . . 9.2.2 Installing Productivity Center for Fabric Remote Console . . . . . . . . . . . . . . . . 9.3 Configuring IBM TotalStorage Productivity Center for Fabric . . . . . . . . . . . . . . . . . . . 9.3.1 Configuring SNMP . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.3.2 Configuring the outband agents . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.3.3 Checking inband agents . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.3.4 Performing an initial poll and setting up the poll interval . . . . . . . . . . . . . . . . . . . Chapter 10. Deployment of agents . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10.1 Installing the agents . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10.2 Data Agent installation using the installer . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10.3 Deploying the agent . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
320 320 320 321 321 331 342 342 346 348 349 351 352 354 361
Part 4. Using the IBM TotalStorage Productivity Center . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 373 Chapter 11. Using TotalStorage Productivity Center for Disk. . . . . . . . . . . . . . . . . . . 11.1 Productivity Center common base: Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.2 Launching TotalStorage Productivity Center. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.3 Exploiting Productivity Center common base . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.3.1 Launch Device Manager . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.4 Performing volume inventory . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.5 Changing the display name of a storage device . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.6 Working with ESS . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.6.1 ESS Volume inventory . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.6.2 Assigning and unassigning ESS Volumes . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.6.3 Creating new ESS volumes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.6.4 Launch device manager for an ESS device . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.7 Working with DS8000 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.7.1 DS8000 Volume inventory . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.7.2 Assigning and unassigning DS8000 Volumes . . . . . . . . . . . . . . . . . . . . . . . . . 11.7.3 Creating new DS8000 volumes. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.7.4 Launch device manager for an DS8000 device . . . . . . . . . . . . . . . . . . . . . . . . 11.8 Working with SAN Volume Controller . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.8.1 Working with SAN Volume Controller MDisks. . . . . . . . . . . . . . . . . . . . . . . . . . 11.8.2 Creating new MDisks on supported storage devices . . . . . . . . . . . . . . . . . . . . 11.8.3 Create and view SAN Volume Controller VDisks . . . . . . . . . . . . . . . . . . . . . . . 11.9 Working with DS4000 family or FAStT storage . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.9.1 Working with DS4000 or FAStT volumes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.9.2 Creating DS4000 or FAStT volumes. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.9.3 Assigning hosts to DS4000 and FAStT Volumes . . . . . . . . . . . . . . . . . . . . . . . 11.9.4 Unassigning hosts from DS4000 or FAStT volumes. . . . . . . . . . . . . . . . . . . . . 11.9.5 Volume properties . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.10 Event Action Plan Builder . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.10.1 Applying an Event Action Plan to a managed system or group . . . . . . . . . . . 11.10.2 Exporting and importing Event Action Plans . . . . . . . . . . . . . . . . . . . . . . . . . . Chapter 12. Using TotalStorage Productivity Center Performance Manager . . . . . . 12.1 Exploiting Performance Manager . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.1.1 Performance Manager GUI . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.1.2 Performance Manager data collection . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vi
IBM TotalStorage Productivity Center V2.3: Getting Started
375 376 376 377 378 378 382 383 384 385 387 388 389 390 392 393 394 396 396 399 402 406 407 409 413 414 415 416 421 423 427 428 429 429
12.1.3 Using IBM Director Scheduler function . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.1.4 Reviewing data collection task status . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.1.5 Managing Performance Manager Database . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.1.6 Performance Manager gauges . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.1.7 ESS thresholds . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.1.8 Data collection for SAN Volume Controller . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.1.9 SAN Volume Controller thresholds . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.1.10 Data collection for the DS6000 and DS8000 . . . . . . . . . . . . . . . . . . . . . . . . . 12.1.11 DS6000 and DS8000 thresholds . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.2 Exploiting gauges . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.2.1 Before you begin . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.2.2 Creating gauges: an example . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.2.3 Zooming in on the specific time period . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.2.4 Modify gauge to view array level metrics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.2.5 Modify gauge to review multiple metrics in same chart. . . . . . . . . . . . . . . . . . . 12.3 Performance Manager command line interface . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.3.1 Performance Manager CLI commands . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.3.2 Sample command outputs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.4 Volume Performance Advisor (VPA) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.4.1 VPA introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.4.2 The provisioning challenge . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.4.3 Workload characterization and workload profiles . . . . . . . . . . . . . . . . . . . . . . . 12.4.4 Workload profile values. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.4.5 How the Volume Performance Advisor makes decisions . . . . . . . . . . . . . . . . . 12.4.6 Enabling the Trace Logging for Director GUI Interface . . . . . . . . . . . . . . . . . . . 12.4.7 Getting started . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12.4.8 Creating and managing workload profiles. . . . . . . . . . . . . . . . . . . . . . . . . . . . . Chapter 13. Using TotalStorage Productivity Center for Data . . . . . . . . . . . . . . . . . . 13.1 TotalStorage Productivity Center for Data overview . . . . . . . . . . . . . . . . . . . . . . . . . 13.1.1 Business purpose of TotalStorage Productivity Center for Data. . . . . . . . . . . . 13.1.2 Components of TotalStorage Productivity Center for Data . . . . . . . . . . . . . . . . 13.1.3 Security considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.2 Functions of TotalStorage Productivity Center for Data . . . . . . . . . . . . . . . . . . . . . . 13.2.1 Basic menu displays . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.2.2 Discover and monitor Agents, disks, filesystems, and databases . . . . . . . . . . 13.2.3 Reporting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.2.4 Alerts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.2.5 Chargeback: Charging for storage usage . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.3 OS Monitoring . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.3.1 Navigation tree . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.3.2 Groups . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.3.3 Discovery . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.3.4 Pings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.3.5 Probes. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.3.6 Profiles . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.3.7 Scans . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.4 OS Alerts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.4.1 Alerting navigation tree . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.4.2 Computer Alerts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.4.3 Filesystem Alerts. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.4.4 Directory Alerts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.4.5 Alert logs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
435 437 439 443 457 460 461 463 466 467 468 468 471 471 474 475 475 477 478 478 478 479 479 480 481 482 508 521 522 522 522 523 523 524 526 529 532 533 533 534 535 540 542 545 547 552 555 558 560 562 563 564
Contents
vii
13.5 Policy management. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.5.1 Quotas . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.5.2 Network Appliance Quotas . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.5.3 Constraints . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.5.4 Filesystem extension and LUN provisioning . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.5.5 Scheduled Actions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.6 Database monitoring . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.6.1 Groups . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.6.2 Probes. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.6.3 Profiles . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.6.4 Scans . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.7 Database Alerts. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.7.1 Instance Alerts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.7.2 Database-Tablespace Alerts. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.7.3 Table Alerts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.7.4 Alert log . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.8 Databases policy management . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.8.1 Network Quotas . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.8.2 Instance Quota . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.8.3 Database Quota . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.9 Database administration samples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.9.1 Database up . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.9.2 Database utilization. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.9.3 Need for reorganization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.10 Data Manager reporting capabilities . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.10.1 Major reporting categories . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.11 Using the standard reporting functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.11.1 Asset Reporting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.11.2 Storage Subsystems Reporting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.11.3 Availability Reporting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.11.4 Capacity Reporting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.11.5 Usage Reporting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.11.6 Usage Violation Reporting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.11.7 Backup Reporting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.12 TotalStorage Productivity Center for Data ESS Reporting . . . . . . . . . . . . . . . . . . . 13.12.1 ESS Reporting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.13 IBM Tivoli Storage Resource Manager top 10 reports . . . . . . . . . . . . . . . . . . . . . . 13.13.1 ESS used and free storage . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.13.2 ESS attached hosts report . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.13.3 Computer Uptime Reporting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.13.4 Growth in storage used and number of files . . . . . . . . . . . . . . . . . . . . . . . . . . 13.13.5 Incremental backup trends . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.13.6 Database reports against DBMS size . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.13.7 Database instance storage report . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.13.8 Database reports size by instance and by computer . . . . . . . . . . . . . . . . . . . 13.13.9 Locate the LUN on which a database is allocated . . . . . . . . . . . . . . . . . . . . . 13.13.10 Finding important files on your systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.14 Creating customized reports . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.14.1 System Reports . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.14.2 Reports owned by a specific username . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.14.3 Batch Reports . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.15 Setting up a schedule for daily reports . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13.16 Setting up a reports Web site . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . viii
IBM TotalStorage Productivity Center V2.3: Getting Started
565 565 570 570 576 582 583 584 585 586 587 588 588 588 589 589 589 590 591 591 591 591 591 591 592 593 594 595 604 604 605 607 610 627 634 634 653 653 656 657 659 661 665 667 667 669 672 683 683 686 688 697 698
13.17 Charging for storage usage. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 700 Chapter 14. Using TotalStorage Productivity Center for Fabric . . . . . . . . . . . . . . . . . 14.1 NetView navigation overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.1.1 NetView interface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.1.2 Maps and submaps. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.1.3 NetView window structure. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.1.4 NetView Explorer . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.1.5 NetView Navigation Tree . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.1.6 Object selection and NetView properties . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.1.7 Object symbols . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.1.8 Object status . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.1.9 Status propagation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.1.10 NetView and Productivity Center for Fabric integration . . . . . . . . . . . . . . . . . 14.2 Walk-through of Productivity Center for Fabric . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.2.1 Device Centric view. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.2.2 Host Centric view . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.2.3 SAN view . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.2.4 Launching element managers. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.2.5 Explore view . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.3 Topology views . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.3.1 SAN view . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.3.2 Device Centric View . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.3.3 Host Centric View . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.3.4 iSCSI discovery. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.3.5 MDS 9000 discovery . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.4 SAN menu options . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.4.1 SAN Properties . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.5 Application launch . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.5.1 Native support . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.5.2 NetView support for Web interfaces . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.5.3 Launching TotalStorage Productivity Center for Data. . . . . . . . . . . . . . . . . . . . 14.5.4 Other menu options . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.6 Status cycles . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.7 Practical cases . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.7.1 Cisco MDS 9000 discovery . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.7.2 Removing a connection on a device running an inband agent . . . . . . . . . . . . . 14.7.3 Removing a connection on a device not running an agent . . . . . . . . . . . . . . . . 14.7.4 Powering off a switch . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.7.5 Running discovery on a RNID-compatible device. . . . . . . . . . . . . . . . . . . . . . . 14.7.6 Outband agents only . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.7.7 Inband agents only . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.7.8 Disk devices discovery . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.7.9 Well placed agent strategy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.8 Netview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.8.1 Reporting overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.8.2 SNMP and MIBs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.9 NetView setup and configuration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.9.1 Advanced Menu . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.9.2 Copy Brocade MIBs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.9.3 Loading MIBs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.10 Historical reporting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.10.1 Creating a Data Collection . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 703 704 704 704 704 705 707 707 709 709 711 711 712 713 714 714 723 725 725 727 731 732 733 734 735 735 739 740 740 742 742 743 745 745 747 750 752 756 758 760 762 764 766 767 767 769 769 770 771 774 775
Contents
ix
14.10.2 Database maintenance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.10.3 Troubleshooting the Data Collection daemon . . . . . . . . . . . . . . . . . . . . . . . . . 14.10.4 NetView Graph Utility . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.11 Real-time reporting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.11.1 MIB Tool Builder . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.11.2 Displaying real-time data. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.11.3 SmartSets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.11.4 SmartSets and Data Collections . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.11.5 Seed file . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.12 Productivity Center for Fabric and iSCSI . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.13 What is iSCSI? . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.14 How does iSCSI work? . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.15 Productivity Center for Fabric and iSCSI . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.15.1 Functional description . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.15.2 iSCSI discovery. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.16 ED/FI - SAN Error Predictor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.16.1 Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.16.2 Error processing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.16.3 Configuration for ED/FI - SAN Error Predictor . . . . . . . . . . . . . . . . . . . . . . . . 14.16.4 Using ED/FI. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14.16.5 Searching for the faulted device on the topology map . . . . . . . . . . . . . . . . . . 14.16.6 Removing notifications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Chapter 15. Using TotalStorage Productivity Center for Replication. . . . . . . . . . . . . 15.1 TotalStorage Productivity Center for Replication overview . . . . . . . . . . . . . . . . . . . . 15.1.1 Supported Copy Services . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15.1.2 Replication session . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15.1.3 Storage group . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15.1.4 Storage pools . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15.1.5 Relationship of group, pool, and session . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15.1.6 Copyset and sequence concepts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15.2 Exploiting Productivity Center for replication. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15.2.1 Before you start. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15.2.2 Adding a replication device . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15.2.3 Creating a storage group . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15.2.4 Modifying a storage group . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15.2.5 Viewing storage group properties . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15.2.6 Deleting a storage group. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15.2.7 Creating a storage pool. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15.2.8 Modifying a storage pool . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15.2.9 Deleting a storage pool . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15.2.10 Viewing storage pool properties . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15.2.11 Creating storage paths . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15.2.12 Point-in-Time Copy - creating a session . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15.2.13 Creating a session - verifying source-target relationship . . . . . . . . . . . . . . . . 15.2.14 Continuous Synchronous Remote Copy - creating a session. . . . . . . . . . . . . 15.2.15 Managing a Point-in-Time copy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15.2.16 Managing a Continuous Synchronous Remote Copy . . . . . . . . . . . . . . . . . . . 15.3 Using Command Line Interface (CLI) for replication . . . . . . . . . . . . . . . . . . . . . . . . . 15.3.1 Session details . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15.3.2 Starting a session . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15.3.3 Suspending a session . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15.3.4 Terminating a session . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
783 784 784 786 787 791 794 802 805 810 811 811 812 813 813 814 814 816 818 820 822 825 827 828 828 830 831 831 832 833 834 834 834 838 841 842 843 844 847 848 849 850 852 856 861 866 873 884 886 888 892 893
Chapter 16. Hints, tips, and good-to-knows . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16.1 SLP configuration recommendation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16.1.1 SLP registration and slptool . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16.2 Tivoli Common Agent Services . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16.2.1 Locations of configured user IDs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16.2.2 Resource Manager registration. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16.2.3 Tivoli Agent Manager status . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16.2.4 Registered Fabric Agents . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16.2.5 Registered Data Agents . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16.3 Launchpad . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16.3.1 Launchpad installation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16.3.2 Launchpad customization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16.4 Remote consoles . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16.5 Verifying whether a port is in use . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16.6 Manually removing old CIMOM entries . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16.7 Collecting logs for support. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16.7.1 IBM Director logfiles . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16.7.2 Using Event Action Plans . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16.7.3 Following Discovery using Windows raswatch utility . . . . . . . . . . . . . . . . . . . . 16.7.4 DB2 database checking . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16.7.5 IBM WebSphere tracing and logfile browsing . . . . . . . . . . . . . . . . . . . . . . . . . . 16.8 SLP and CIM Agent problem determination . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16.8.1 Enabling SLP tracing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16.8.2 Device registration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16.9 Replication Manager problem determination . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16.9.1 Diagnosing an indications problem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16.9.2 Restarting the replication environment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16.10 Enabling trace logging. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16.10.1 Enabling WebSphere Application Server trace . . . . . . . . . . . . . . . . . . . . . . . . 16.11 ESS user authentication problem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16.12 SVC Data collection task failure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Chapter 17. Database management and reporting . . . . . . . . . . . . . . . . . . . . . . . . . . . 17.1 DB2 database overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17.2 Database purging in TotalStorage Productivity Center . . . . . . . . . . . . . . . . . . . . . . . 17.2.1 Performance manager database panel. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17.3 IBM DB2 tool suite . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17.3.1 Command Line Tools . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17.3.2 Development Tools . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17.3.3 General Administration Tools . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17.3.4 Monitoring Tools . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17.4 DB2 Command Center overview. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17.4.1 Command Center navigation example . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17.5 DB2 Command Center custom report example . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17.5.1 Extracting LUN data report . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17.5.2 Command Center report . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17.6 Exporting collected performance data to a file . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17.6.1 Control Center. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17.6.2 Data extraction tools, tips and reporting methods. . . . . . . . . . . . . . . . . . . . . . . 17.7 Database backup and recovery overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17.8 Backup example . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
899 900 901 901 901 902 902 904 906 906 907 909 911 911 911 917 917 921 921 922 927 928 929 930 930 931 931 931 932 940 940 943 944 944 945 948 948 950 950 951 952 952 956 956 959 976 976 979 984 988
xi
Server information. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . User IDs and passwords for key files and installation. . . . . . . . . . . . . . . . . . . . . . . . . . Storage device information . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . IBM TotalStorage Enterprise Storage Server . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . IBM FAStT . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . IBM SAN Volume Controller . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Related publications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . IBM Redbooks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Other publications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Online resources . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . How to get IBM Redbooks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Help from IBM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
992 993 994 994 995 996 997 997 997 997 998 998
Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 999
xii
Notices
This information was developed for products and services offered in the U.S.A. IBM may not offer the products, services, or features discussed in this document in other countries. Consult your local IBM representative for information on the products and services currently available in your area. Any reference to an IBM product, program, or service is not intended to state or imply that only that IBM product, program, or service may be used. Any functionally equivalent product, program, or service that does not infringe any IBM intellectual property right may be used instead. However, it is the user's responsibility to evaluate and verify the operation of any non-IBM product, program, or service. IBM may have patents or pending patent applications covering subject matter described in this document. The furnishing of this document does not give you any license to these patents. You can send license inquiries, in writing, to: IBM Director of Licensing, IBM Corporation, North Castle Drive Armonk, NY 10504-1785 U.S.A. The following paragraph does not apply to the United Kingdom or any other country where such provisions are inconsistent with local law: INTERNATIONAL BUSINESS MACHINES CORPORATION PROVIDES THIS PUBLICATION "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESS OR IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF NON-INFRINGEMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Some states do not allow disclaimer of express or implied warranties in certain transactions, therefore, this statement may not apply to you. This information could include technical inaccuracies or typographical errors. Changes are periodically made to the information herein; these changes will be incorporated in new editions of the publication. IBM may make improvements and/or changes in the product(s) and/or the program(s) described in this publication at any time without notice. Any references in this information to non-IBM Web sites are provided for convenience only and do not in any manner serve as an endorsement of those Web sites. The materials at those Web sites are not part of the materials for this IBM product and use of those Web sites is at your own risk. IBM may use or distribute any of the information you supply in any way it believes appropriate without incurring any obligation to you. Information concerning non-IBM products was obtained from the suppliers of those products, their published announcements or other publicly available sources. IBM has not tested those products and cannot confirm the accuracy of performance, compatibility or any other claims related to non-IBM products. Questions on the capabilities of non-IBM products should be addressed to the suppliers of those products. This information contains examples of data and reports used in daily business operations. To illustrate them as completely as possible, the examples include the names of individuals, companies, brands, and products. All of these names are fictitious and any similarity to the names and addresses used by an actual business enterprise is entirely coincidental. COPYRIGHT LICENSE: This information contains sample application programs in source language, which illustrates programming techniques on various operating platforms. You may copy, modify, and distribute these sample programs in any form without payment to IBM, for the purposes of developing, using, marketing or distributing application programs conforming to the application programming interface for the operating platform for which the sample programs are written. These examples have not been thoroughly tested under all conditions. IBM, therefore, cannot guarantee or imply reliability, serviceability, or function of these programs. You may copy, modify, and distribute these sample programs in any form without payment to IBM for the purposes of developing, using, marketing, or distributing application programs conforming to IBM's application programming interfaces.
xiii
Trademarks
The following terms are trademarks of the International Business Machines Corporation in the United States, other countries, or both:
AIX Cloudscape DB2 DB2 Universal Database e-business on demand Enterprise Storage Server Eserver Eserver FlashCopy IBM ibm.com iSeries MVS Netfinity NetView OS/390 Predictive Failure Analysis pSeries QMF Redbooks Redbooks (logo) S/390 Sequent ThinkPad Tivoli Enterprise Tivoli Enterprise Console Tivoli TotalStorage WebSphere xSeries z/OS zSeries 1-2-3
The following terms are trademarks of other companies: Java and all Java-based trademarks and logos are trademarks or registered trademarks of Sun Microsystems, Inc. in the United States, other countries, or both. Microsoft, Windows, Windows NT, and the Windows logo are trademarks of Microsoft Corporation in the United States, other countries, or both. Intel, Intel Inside (logos), MMX, and Pentium are trademarks of Intel Corporation in the United States, other countries, or both. UNIX is a registered trademark of The Open Group in the United States and other countries. Linux is a trademark of Linus Torvalds in the United States, other countries, or both. Other company, product, and service names may be trademarks or service marks of others.
xiv
Preface
IBM TotalStorage Productivity Center is a suite of infrastructure management software that can centralize, automate, and simplify the management of complex and heterogeneous storage environments. It can help reduce the effort of managing complex storage infrastructures, improve storage capacity utilization, and improve administration efficiency. IBM TotalStorage Productivity Center allows you to respond to on demand storage needs and brings together, in a single point, the management of storage devices, fabric, and data. This IBM Redbook is intended for administrators and users who are installing and using IBM TotalStorage Productivity Center V2.3. It provides an overview of the product components and functions. We describe the hardware and software environment required, provide a step-by-step installation procedure, and offer customization and usage hints and tips. This book is not a replacement for the existing IBM Redbooks, or product manuals, that detail the implementation and configuration of the individual products that make up the IBM TotalStorage Productivity Center, or the products as they may have been called in previous versions. We refer to those books as appropriate throughout this book.
xv
Thanks to the following people for their contributions to this project: Sangam Racherla Yvonne Lyon ITSO, San Jose Center Bob Haimowitz ITSO, Raleigh Center Diana Duan Tina Dunton Nancy Hobbs Paul Lee Thiha Than Miki Walter IBM San Jose Martine Wedlake IBM Beaverton Ryan Darris IBM Tucson Doug Dunham Tivoli Storage SWAT Team Mike Griese Technical Support Marketing Lead, Rochester Curtis Neal Scott Venuti Open System Demo Center, San Jose
xvi
Comments welcome
Your comments are important to us! We want our Redbooks to be as helpful as possible. Send us your comments about this or other Redbooks in one of the following ways: Use the online Contact us review redbook form found at:
ibm.com/redbooks
Mail your comments to: IBM Corporation, International Technical Support Organization Dept. QXXE Building 80-E2 650 Harry Road San Jose, California 95120-6099
Preface
xvii
xviii
Part 1
Part
Chapter 1.
Key standards for Storage Management are: Distributed Management Task Force (DMTF) Common Information Model (CIM) Standards. This includes the CIM Device Model for Storage, which at the time of writing was Version 2.7.2 for the CIM schema. Storage Networking Industry Association (SNIA) Storage Management Initiative Specification (SMI-S).
Previously we discussed the next steps or entry points into an On Demand environment. The IBM software products which represent these entry points and which comprise the IBM TotalStorage Open Software Family is shown in Figure 1-3.
Taking a closer look at storage infrastructure management (see Figure 1-4), we focus on four subject matter experts to empower the storage administrators to effectively do their work. Data subject matter expert San Fabric subject matter expert Disk subject matter expert Replication subject matter expert
1.3.1 Data subject matter expert: TotalStorage Productivity Center for Data
The Data subject matter expert has intimate knowledge of how storage is used, for example whether the data is used by a file system or a database application. Figure 1-5 on page 8 shows the role of the Data subject matter expert which is filled by the TotalStorage Productivity Center for Data (formerly the IBM Tivoli Storage Resource Manager).
Figure 1-5 Monitor and Configure the Storage Infrastructure Data area
Heterogeneous storage infrastructures, driven by growth in file and database data, consume increasing amounts of administrative time, as well as actual hardware resources. IT managers need ways to make their administrators more efficient and more efficiently utilize their storage resources. Tivoli Storage Resource Manager gives storage administrators the automated tools they need to manage their storage resources more cost-effectively. TotalStorage Productivity Center for Data allows you to identify different classes of data, report how much space is being consumed by these different classes, and take appropriate actions to keep the data under control. Features of the TotalStorage Productivity Center for Data are: Automated identification of the storage resources in an infrastructure and analysis of how effectively those resources are being used. File-system and file-level evaluation uncovers categories of files that, if deleted or archived, can potentially represent significant reductions in the amount of data that must be stored, backed up and managed. Automated control through policies that are customizable with actions that can include centralized alerting, distributed responsibility and fully automated response. Predict future growth and future at-risk conditions with historical information. Through monitoring and reporting, TotalStorage Productivity Center for Data helps the storage administrator prevent outages in the storage infrastructure. Armed with timely information, the storage administrator can take action to keep storage and data available to the application. TotalStorage Productivity Center for Data also helps to make the most efficient use of storage budgets, by allowing administrators to use their existing storage more efficiently, and more accurately predict future storage growth.
TotalStorage Productivity Center for Data monitors storage assets, capacity, and usage across an enterprise. TotalStorage Productivity Center for Data can look at: Storage from a host perspective: Manage all the host-attached storage, capacity and consumption attributed to file systems, users, directories, and files Storage from an application perspective: Monitor and manage the storage activity inside different database entities including instance, tablespace, and table Storage utilization and provide chargeback information.
Architecture
The TotalStorage Productivity Center for Data server system manages a number of Agents, which can be servers with storage attached, NAS systems, or database application servers. Information is collected from the Agents and stored in a database repository. The stored information can then be displayed from a native GUI client or browser interface anywhere in the network. The GUI or browser interface gives access to the other functions of TotalStorage Productivity Center for Data, including creating and customizing of a large number of different types of reports and setting up alerts. With TotalStorage Productivity Center for Data, you can: Monitor virtually any host Monitor local, SAN-attached and Network Attached Storage from a browser anywhere on the network For more information refer to the redbook IBM Tivoli Storage Resource Manager: A Practical Introduction, SG24-6886.
Optimized Storage Resource Utilization by reporting on storage network performance Enhanced Storage Personnel Productivity - Tivoli SAN Manager creates a single point of control, administration and security for the management of heterogeneous storage networks Figure 1-6 describes the requirements that must be addressed by the Fabric subject matter expert.
Figure 1-6 Monitor and Configure the Storage Infrastructure Fabric area
TotalStorage Productivity Center for Fabric monitors and manages switches and hubs, storage and servers in a Storage Area Network. TotalStorage Productivity Center for Fabric can be used for both online monitoring and historical reporting. TotalStorage Productivity Center for Fabric: Manages fabric devices (switches) through outband management. Discovers many details about a monitored server and its local storage through an Agent loaded onto a SAN-attached host (Managed Host). Monitors the network and collects events and traps Launches vendor-provided specific SAN element management applications from the TotalStorage Productivity Center for Fabric Console. Discovers and manages iSCSI devices. Provides a fault isolation engine for SAN problem determination (ED/FI - SAN Error Predictor) TotalStorage Productivity Center for Fabric is compliant with the standards relevant to SAN storage and management.
10
11
In outband discovery, all communications occur over the IP network: TotalStorage Productivity Center for Fabric requests information over the IP network from a switch using SNMP queries on the device. The device returns the information toTotalStorage Productivity Center for Fabric, also over the IP network. Inband discovery is the process of discovering information about the SAN, including topology and attribute data, through the Fibre Channel data paths. In inband discovery, both the IP and Fibre Channel networks are used: TotalStorage Productivity Center for Fabric requests information (via the IP network) from a Tivoli SAN Manager agent installed on a Managed Host. That agent requests information over the Fibre Channel network from fabric elements and end points in the Fibre Channel network. The agent returns the information to TotalStorage Productivity Center for Fabric over the IP network. TotalStorage Productivity Center for Fabric collects, co-relates and displays information from all devices in the storage network, using both the IP network and the Fibre Channel network. If the Fibre Channel network is unavailable for any reason, monitoring can still continue over the IP network.
1.3.3 Disk subject matter expert: TotalStorage Productivity Center for Disk
The Disk subject matter experts job allows you to manage the disk systems. It will discover and classify all disk systems that exist and draw a picture of all discovered disk systems. The Disk subject matter expert provides the ability to monitor, configure, create disks and do LUN masking of disks. It also does performance trending and performance threshold I/O analysis for both real disks and virtual disks. It also does automated status and problem alerts via SNMP. This role is filled by the TotalStorage Productivity Center for Disk (formerly the IBM TotalStorage Multiple Device Manager Performance Manager component). The requirements addressed by the Disk subject matter expert are shown in Figure 1-7 on page 13. The disk systems monitoring and configuration needs must be covered by a comprehensive management tool like the TotalStorage Productivity Center for Disk.
12
Figure 1-7 Monitor and configure the Storage Infrastructure Disk area
The TotalStorage Productivity Center for Disk provides the raw capabilities of initiating and scheduling performance data collection on the supported devices, of storing the received performance statistics into database tables for later use, and of analyzing the stored data and generating reports for various metrics of the monitored devices. In conjunction with data collection, the TotalStorage Productivity Center for Disk is responsible for managing and monitoring the performance of the supported storage devices. This includes the ability to configure performance thresholds for the devices based on performance metrics, the generation of alerts when these thresholds are exceeded, the collection and maintenance of historical performance data, and the creation of gauges, or performance reports, for the various metrics to display the collected historical data to the end user. The TotalStorage Productivity Center for Disk enables you to perform sophisticated performance analysis for the supported storage devices.
Functions
TotalStorage Productivity Center for Disk provides the following functions: Collect data from devices The Productivity Center for Disk collects data from the IBM TotalStorage Enterprise Storage Server (ESS), SAN Volume Controller (SVC), DS400 family and SMI-S enabled devices. Each Performance Collector collects performance data from one or more storage groups, all of the same device type (for example, ESS or SAN Volume Controller). Each Performance Collection has a start time, a stop time, and a sampling frequency. The performance sample data is stored in DB2 database tables. Configure performance thresholds You can use the Productivity Center for Disk to set performance thresholds for each device type. Setting thresholds for certain criteria enables Productivity Center for Disk to notify you when a certain threshold has been exceeded, so that you to take action before a critical event occurs.
13
You can specify what action should be taken when a threshold-exceeded condition occurs. The action may be to log the occurrence or to trigger an event. The threshold settings can vary by individual device. Monitor performance metrics across storage subsystems from a single console Receive timely alerts to enable event action based on customer policies View performance data from the Productivity Center for Disk database You can view performance data from the Productivity Center for Disk database in both graphical and tabular forms. The Productivity Center for Disk allows a TotalStorage Productivity Center user to access recent performance data in terms of a series of values of one or more metrics, associated with a finite set of components per device. Only recent performance data is available for gauges. Data that has been purged from the database cannot be viewed. You can define one or more gauges by selecting certain gauge properties and saving them for later referral. Each gauge is identified through a user-specified name, and once defined, a gauge can be started, which means it is then displayed in a separate window of the TotalStorage Productivity Center GUI. You can have multiple gauges active at the same time. Gauge definition will be accomplished through a wizard, to aid in entering a valid set of gauge properties. Gauges are saved in the Productivity Center for Disk database and retrieved upon request. When you request data pertaining to a defined gauge, the Performance Manager builds a query to the database, retrieves and formats the data and returns it to you. Once started, a gauge is displayed in its own window, and displays all available performance data for the specified initial date/time range. The date/time range can be changed after the initial gauge widow is displayed. Focus on storage optimization through identification of best LUN The Volume Performance Advisor is an automated tool to help the storage administrator pick the best possible placement of a new LUN to be allocated, that is, the best placement from a performance perspective. It also uses the historical performance statistics collected from the supported devices, to locate unused storage capacity on the SAN that exhibits the best (estimated) performance characteristics. Allocation optimization involves several variables which are user controlled, such as required performance level and the time of day/week/month of prevalent access. This function is fully integrated with the Device Manager function, this is so that when a new LUN is added, for example, to the ESS, the Performance Manager can seamlessly select the best possible LUN. For detailed information about how to use the functions of the TotalStorage Productivity Center for Disk refer to Chapter 11, Using TotalStorage Productivity Center for Disk on page 375.
14
Productivity Center for Replication will start up all replication pairs and monitor them to completion. If any of the replication pairs fail, meaning the application is out of sync, the Productivity Center for Replication will suspend them until the problem is resolved, resync them and resume the replication. The Productivity Center for Replication provides complete management of the replication process. The requirements addressed by the Replication subject matter expert are shown Figure 1-8. Replication in a complex environment needs to be addressed by a comprehensive management tool like the TotalStorage Productivity Center for Replication.
Figure 1-8 Monitor and Configure the Storage Infrastructure Replication area
Functions
Data replication is the core function required for data protection and disaster recovery. It provides advanced copy services functions for supported storage subsystems on the SAN. Replication Manager administers and configures the copy services functions and monitors the replication actions. Its capabilities consist of the management of two types of copy services: the Continuous Copy (also known as Peer-to-Peer, PPRC, or Remote Copy), and the Point-in-Time Copy (also known as FlashCopy). At this time TotalStorage Productivity Center for Replication supports the IBM TotalStorage ESS. Productivity Center for Replication includes support for replica sessions, which ensures that data on multiple related heterogeneous volumes is kept consistent, provided that the underlying hardware supports the necessary primitive operations. Productivity Center for Replication also supports the session concept, such that multiple pairs are handled as a consistent unit, and that Freeze-and-Go functions can be performed when errors in mirroring occur. Productivity Center for Replication is designed to control and monitor the copy services operations in large-scale customer environments.
15
Productivity Center for Replication provides a user interface for creating, maintaining, and using volume groups and for scheduling copy tasks. The User Interface populates lists of volumes using the Device Manager interface. Some of the tasks you can perform with Productivity Center for Replication are: Create a replication group. A replication group is a collection of volumes grouped together so that they can be managed concurrently. Set up a Group for replication. Create, save, and name a replication task. Schedule a replication session with the user interface: Create Session Wizard. Select Source Group. Select Copy Type. Select Target Pool. Save Session.
Start a replication session A user can also perform these tasks with the Productivity Center for Replication command-line interface. For more information about the Productivity Center for Replication functions refer to Chapter 15, Using TotalStorage Productivity Center for Replication on page 827.
16
The IBM TotalStorage Productivity Center establishes the foundation for IBMs e-business On Demand technology. We need the function in an On Demand environment to provide IT resources On Demand - when the resources are needed by an application to support the customers business process. Of course, we are able to provide resources or remove resources today but the question is how. The process is expensive and time consuming. The IBM TotalStorage Productivity Center is the basis for the provisioning of storage resources to make the e-business On Demand environment a reality. In the future there will be more automation required to handle the hugh amount work in the provisioning area, more automation like the BM TotalStorage Productivity Center launch pad provides. Automation means workflow. Workflow is the key to getting work automated. IBM has a long history and investment in building workflow engines and work flows. Today IBM is using the IBM Tivoli Intelligent Orchestrator and IBM Tivoli Provisioning Manager to satisfy the resource requests in the e-business on demand environment in the server arena. The IBM Tivoli Intelligent Orchestrator and The IBM Tivoli Provisioning Manager provide the provisioning in the e-business On Demand environment.
1.4.1 Productivity Center for Disk and Productivity Center for Replication
The Productivity Center for Disk and Productivity Center for Replication is software that has been designed to enable administrators to manage SANs and storage from a single console (Figure 1-10 on page 18). This software solution is designed specifically for managing networked storage components based on the SMI-S, including: IIBM TotalStorage SAN Volume Controller IBM TotalStorage Enterprise Storage Server (ESS) IBM TotalStorage Fibre Array Storage Technology (FAStT) IBM TotalStorage DS4000 series SMI enabled device
17
Productivity Center for Disk and Productivity Center for Replication are built on IBM Director, a comprehensive server management solution. Using Director with the multiple device management solution enables administrators to consolidate the administration of IBM storage subsystems and provide advanced storage management functions (including replication and performance management) across multiple IBM storage subsystems. It interoperates with SAN Management and Enterprise System Resource Manager (ESRM) products from IBM, includingTotalStorage Productivity Center for Data and SAN Management products from other vendors. In a SAN environment, multiple devices work together to create a storage solution. The Productivity Center for Disk and Productivity Center for Replication provides integrated administration, optimization, and replication features for interacting SAN devices, including the SAN Volume Controller and DS4000 Family devices. It provides an integrated view of the underlying system so that administrators can drill down through the virtualized layers to easily perform complex configuration tasks and more productively manage the SAN infrastructure. Because the virtualization layers support advanced replication configurations, the Productivity Center for Disk and Productivity Center for Replication products offer features that simplify the configuration, monitoring, and control of disaster recovery and data migration solutions. In addition, specialized performance data collection, analysis, and optimization features are provided. As the SNIA standards mature, the Productivity Center view will be expanded to include CIM-enabled devices from other vendors, in addition to IBM storage. Figure 1-11 on page 19 provides an overview of Productivity Center for Disk and Productivity Center for Replication.
18
Device Manager
IBM Director
WebSphere Application Server DB2
The Productivity Center for Disk and Productivity Center for Replication provides support for configuration, tuning, and replication of the virtualized SAN. As with the individual devices, the Productivity Center for Disk and Productivity Center for Replication layers are open and can be accessed via a GUI, CLI, or standards-based Web Services. Productivity Center for Disk and Productivity Center for Replication provide the following functions: Device Manager - Common function provided when you install the base prerequisite products for either Productivity Center for Disk or Productivity Center for Replication Performance Manager - provided by Productivity Center for Disk Replication Manager - provided by Productivity Center for Replication
Device Manager
The Device Manager is responsible for the discovery of supported devices; collecting asset, configuration, and availability data from the supported devices; and providing a limited topography view of the storage usage relationships between those devices. The Device Manager builds on the IBM Director discovery infrastructure. Discovery of storage devices adheres to the SNIA SMI-S specification standards. Device Manager uses the Service Level Protocol (SLP) to discover SMI-S enabled devices. The Device Manager creates managed objects to represent these discovered devices. The discovered managed objects are displayed as individual icons in the Group Contents pane of the IBM Director Console as shown in Figure 1-12 on page 20.
19
Device Manager provides a subset of configuration functions for the managed devices, primarily LUN allocation and assignment. Its function includes certain cross-device configuration, as well as the ability to show and traverse inter-device relationships. These services communicate with the CIM Agents that are associated with the particular devices to perform the required configuration. Devices that are not SMI-S compliant are not supported. The Device Manager also interacts and provides some SAN management functionality when IBM Tivoli SAN Manager is installed. The Device Manager health monitoring keeps you aware of hardware status changes in the discovered storage devices. You can drill down to the status of the hardware device, if applicable. This enables you to understand which components of a device are malfunctioning and causing an error status for the device.
SAN Management
When a supported SAN Manager is installed and configured, the Device Manager leverages the SAN Manager to provide enhanced function. Along with basic device configuration functions such as LUN creation, allocation, assignment, and deletion for single and multiple devices, basic SAN management functions such as LUN discovery, allocation, and zoning are provided in one step. IBM TotalStorage Productivity Center for Fabric (formerly IBM Tivoli SAN Manager) is currently the supported SAN Manager. The set of SAN Manager functions that will be exploited are: The ability to retrieve the SAN topology information, including switches, hosts, ports, and storage devices The ability to retrieve and to modify the zoning configuration on the SAN The ability to register for event notification, to ensure that Productivity Center for Disk is aware when the topology or zoning changes as new devices are discovered by the SAN Manager, and when hosts' LUN configurations change
20
Functions
Collect data from devices The Performance Manager collects data from the IBM TotalStorage Enterprise Storage Server (ESS), IBM TotalStorage SAN Volume Controller (SVC), IBM TotalStorage DS4000 series, IBM TotalStorage DS6000 and IBM TotalStorage DS8000 series and SMI-S enabled devices. The performance collection task collects performance data from one or more storage groups, all of the same device type (for example, ESS or SVC). Each performance collection task has a start time, a stop time, and a sampling frequency. The performance sample data is stored in DB2 database tables. Configure performance thresholds You can use the Performance Manager to set performance thresholds for each device type. Setting thresholds for certain criteria enables Performance Manager to notify you when a certain threshold has been exceeded, so that you can take action before a critical event occurs. You can specify what action should be taken when a threshold-exceeded condition occurs. The action may be to log the occurrence or to trigger an event. The threshold settings can vary by individual device. The eligible metrics for threshold checking are fixed for each storage device. If the threshold metrics are modified by the user, the modifications are accepted immediately and applied to checking being performed by active performance collection tasks. Examples of threshold metrics include: Disk utilization value Average cache hold time Percent of sequential I/Os I/O rate NVS full value Virtual disk I/O rate Managed disk I/O rate There is a user interface that supports threshold settings, enabling a user to: Modify a threshold property for a set of devices of like type. Modify a threshold property for a single device. Reset a threshold property to the IBM-recommended value (if defined) for a set of devices of like type. IBM-recommended critical and warning values will be provided for all thresholds known to indicate potential performance problems for IBM storage devices.
21
Reset a threshold property to the IBM-recommended value (if defined) for a single device. Show a summary of threshold properties for all of the devices of like type. View performance data from the Performance Manager database.
Gauges
The Performance Manager supports a performance-type gauge. The performance-type gauge presents sample-level performance data. The frequency at which performance data is sampled on a device depends on the sampling frequency that you specify when you define the performance collection task. The maximum and minimum values of the sampling frequency depend on the device type. The static display presents historical data over time. The refreshable display presents near real-time data from a device that is currently collecting performance data. The Performance Manager enables a Productivity Center for Disk user to access recent performance data in terms of a series of values of one or more metrics associated with a finite set of components per device. Only recent performance data is available for gauges. Data that has been purged from the database cannot be viewed. You can define one or more gauges by selecting certain gauge properties and saving them for later referral. Each gauge is identified through a user-specified name and, when defined, a gauge can be started, which means that it is then displayed in a separate window of the Productivity Center GUI. You can have multiple gauges active at the same time. Gauge definition is accomplished through a wizard to aid in entering a valid set of gauge properties. Gauges are saved in the Productivity Center for Disk database and retrieved upon request. When you request data pertaining to a defined gauge, the Performance Manager builds a query to the database, retrieves and formats the data, and returns it to you. When started, a gauge is displayed in its own window, and it displays all available performance data for the specified initial date/time range. The date/time range can be changed after the initial gauge window is displayed. For performance-type gauges, if a metric selected for display is associated with a threshold enabled for checking, the current threshold properties are also displayed in the gauge window and are updated each time the gauge data is refreshed.
22
Database information function Due to the amount of data collected by the Performance Manager function provided by Productivity Center for Disk, the database should be monitored to prevent it from running out of space. The database information function returns the database % full. This function can be invoked from either the Web user interface or the CLI.
23
An event is an occurrence of a predefined condition relating to a specific managed object that identifies a change in a system process or a device. The notification of that change can be generated and tracked (for example, notification that a Productivity Center component is not available). Productivity Center for Disk and Productivity Center for Replication take full advantage of, and build upon, the IBM Director Event Services. The IBM Director includes sophisticated event-handling support. Event Action Plans can be set up that specify what steps, if any, should be taken when particular events occur in the environment. Director Event Management encompasses the following concepts: Events can be generated by any managed object. IBM Director receives such events and calls appropriate internal event handlers that have been registered. Actions are user-configured steps to be taken for a particular event or type of event. There can be zero or more actions associated with a particular action plan. System administrators can create their own actions by customizing particular predefined actions. Event Filters are a set of characteristics or criteria that determine whether an incoming event should be acted on. Event Action Plans are associations of one or more event filters with one or more actions. Event Action Plans become active when you apply them to a system or a group of systems. The IBM Director Console includes an extensive set of GUI panels, called the Event Action Plan Builder, that enable the user to create action plans and event filters. Event Filters can be configured using the Event Action Plan Builder and set up with a variety of criteria, such as event types, event severities, day and time of event occurrence, and event categories. This allows control over exactly what action plans are invoked for each specific event. Productivity Center provides extensions to the IBM Director event management support. It takes full advantage of the IBM Director built-in support for event logging and viewing. It generates events that will be externalized. Action plans can be created based on filter criteria for these events. The default action plan is to log all events in the event log. It creates additional event families, and event types within those families, that will be listed in the Event Action Plan Builder. Event actions that enable Productivity Center functions to be exploited from within action plans will be provided. An example is the action to indicate the amount of historical data to be kept.
24
An On Demand operating environment must be: Flexible Self-managing Scalable Economical Resilient Based on open standards The move to an On Demand storage environment is an evolving one, it does not happen all at once. There are several next steps that you may take to move to the On Demand environment. Constant changes to the storage infrastructure (upgrading or changing hardware for example) can be addressed by virtualization which provides flexibility by hiding the hardware and software from users and applications. Empower administrators with automated tools for managing heterogeneous storage infrastructures. and eliminate human error. Control storage growth with automated identification and movement of low-activity or inactive data to a hierarchy of lower-cost storage. Manage cost associated with capturing point-in-time copies of important data for regulatory or bookkeeping requirements by maintaining this inactive data in a hierarchy of lower-cost storage. Ensure recoverability through the automated creation, tracking and vaulting of reliable recovery points for all enterprise data. The ultimate goal to eliminate human errors by preparing for Infrastructure Orchestration software that can be used to automate workflows. No matter which steps you take to an On Demand environment there will be results. The results will be improved application availability, optimized storage resource utilization, and enhanced storage personnel productivity.
25
26
Chapter 2.
Key concepts
There are certain industry standards and protocols that are the basis of the IBM TotalStorage Productivity Center. The understanding of these concepts is important for installing and customizing the IBM TotalStorage Productivity Center. In this chapter, we describe the standards on which the IBM TotalStorage Productivity Center is built, as well as the methods of communication used to discover and manage storage devices. We also discuss communication between the various components of the IBM TotalStorage Productivity Center. To help you understand these concepts, we provide diagrams to show the relationship and interaction of the various elements in the IBM TotalStorage Productivity Center environment.
27
IBM TotalStorage Productivity Center and Tivoli Provisioning Manager are presented as building blocks in the diagram. Both of the products are not a single application but a complex environment by themselves. The diagram also shows the different methods used to collect information from multiple systems to give an administrator the necessary views on the environment, for example: Software clients (agents) Standard interfaces and protocols (for example, Simple Network Management Protocol (SNMP), Common Information Model (CIM) Agent) Proprietary interfaces (for only a few devices) In addition to the central data collection, Productivity Center provides a single point of control for a storage administrator, even though each manager still comes with its own interface. A program called the Launchpad is provided to start the individual applications from a central dashboard.
28
The Tivoli Provisioning Manager relies on Productivity Center to make provisioning possible.
Layer two consists of the individual components that are part of the product suite: IBM TotalStorage Productivity Center for Disk IBM TotalStorage Productivity Center for Replication IBM TotalStorage Productivity Center for Fabric IBM TotalStorage Productivity Center for Data Throughout this redbook, these products are referred to as managers or components. Layer three includes all the prerequisite components, for example IBM DB2, IBM WebSphere, IBM Director, IBM Tivoli NetView, and Tivoli Common Agent Services. IBM TotalStorage Productivity Center for Fabric can be installed on a full version of WebSphere Application Server or on the embedded WebSphere Application Server, which is shipped with Productivity Center for Fabric. Installation on a full version of WebSphere Application Server is used when other components of TotalStorage Productivity Center are installed on the same logical server. IBM TotalStorage Productivity Center for Fabric can utilize an existing IBM Tivoli Netview installation or can be installed along with it. Note: Each of the manager and prerequisite components can be drilled down even further, but in this book we go into this detail only where necessary. The only exception is Tivoli Common Agent Services, which is a new underlying service in the Tivoli product family.
29
Product A product is something that is available to be ordered. The individual products that are
included in IBM TotalStorage Productivity Center are introduced in Chapter 1, IBM TotalStorage Productivity Center overview on page 3.
Components
Products (licensed software packages) and prerequisite software applications are in general called components. Some of the components are internal, meaning that, from the installation and configuration point of view, they are somewhat transparent. External components have to be separately installed. We usually use the term components for the following applications: IBM Director (external, used by Disk and Replication Manager) IBM DB2 (external, used by all managers) IBM WebSphere Application Server (external, used by Disk and Replication Manager, used by Fabric Manager if installed on the same logical server) Embedded WebSphere Application Server (internal, used by Fabric Manager) Tivoli NetView (internal, used by Fabric Manager) Tivoli Common Agent Services (external, used by Data and Fabric Manager) Not all of the internal components are always shown in the diagrams and lists in this book. The term subcomponent is used to emphasize that a certain component (the subcomponent) belongs to or is used by another component. For example, a Resource Manager is a subcomponent of the Fabric or Data Manager.
Managers The managers are the central components of the IBM TotalStorage Productivity Center
environment. They may share some of the prerequisite components. For example, IBM DB2 and IBM WebSphere are used by different managers. In this book, we sometimes use the following terms: Disk Manager for Productivity Center for Disk Replication Manager for Productivity Center for Replication Data Manager for Productivity Center for Data Fabric Manager for Productivity Center for Fabric In addition, we use the term manager for the Agent Manager for Tivoli Agent Manager component, because the name of the component already includes that term.
Agents The agents are not shown in the diagram in Figure 2-2 on page 29, but they have an
important role in the IBM TotalStorage Productivity Center environment. There are two types of agents: Common Information Model (CIM) Agents and agents that belong to one of the managers: CIM Agents: Agents that offer a CIM interface for management applications, for example, for IBM TotalStorage DS8000 and DS6000 series storage systems, IBM TotalStorage Enterprise Storage Server (ESS), SAN (Storage Area Network) Volume Controller, and DS4000 Storage Systems formerly known as FAStT (Fibre Array Storage Technology) Storage Systems Agents that belong to one of the managers: Data Agents: Agents to collect data for the Data Manager Fabric Agents: Agents that are used by the Fabric Manager for inband SAN data discovery and collection 30
IBM TotalStorage Productivity Center V2.3: Getting Started
In addition to these agents, the Service Location Protocol (SLP) also use the term agent for these components: User Agent Service Agent Directory Agent
Elements
We use the generic term element whenever we do not differentiate between components and managers.
All these components can also run on one machine. In this case all managers and IBM Director will share the same DB2 installation and all managers and IBM Tivoli Agent Manager will share the same WebSphere installation.
31
32
The discovery paths are shown in parentheses in the diagram in Figure 2-4.
33
T11 committee
Since the 1970s, the objective of the ANSI T11 committee is to define interface standards for high-performance and mass storage applications. Since that time, the committee has completed work on three projects: High-Performance Parallel Interface (HIPPI) Intelligent Peripheral Interface (IPI) Single-Byte Command Code Sets Connection (SBCON) Currently the group is working on Fibre Channel (FC) and Storage Network Management (SM) standards.
34
The WBEM architecture defines the following elements: CIM Client The CIM Client is a management application similar to IBM TotalStorage Productivity Center that uses CIM to manage devices. A CIM Client can reside anywhere in the network, because it uses HTTP to talk to CIM Object Managers and Agents. CIM Managed Object A CIM Managed Object is a hardware or software component that can be managed by a management application using CIM. CIM Agent The CIM Agent is embedded into a device or it can be installed on the server using the CIM provider as the translator of devices proprietary commands to CIM calls, and interfaces with the management application (the CIM Client). The CIM Agent is linked to one device. CIM Provider A CIM Provider is the element that translates CIM calls to the device-specific commands. It is like a device driver. A CIM Provider is always closely linked to a CIM Object Manager or CIM Agent. CIM Object Manager A CIM Object Manager (CIMOM) is a part of the CIM Server that links the CIM Client to the CIM Provider. It enables a single CIM Agent to talk to multiple devices. CIM Server A CIM Server is the software that runs the CIMOM and the CIM provider for a set of devices. This approach is used when the devices do not have an embedded CIM Agent. This term is often not used. Instead people often use the term CIMOM when they really mean the CIM Server.
35
The idea behind SMI-S is to standardize the management interfaces so that management applications can use these and provide cross device management. This means that a newly introduced device can be immediately managed as it conforms to the standards. SMI-S extends CIM and WBEM with the following features: A single management transport Within the WBEM architecture, the CIM-XML over HTTP protocol was selected for this transport in SMI-S. A complete, unified, and rigidly specified object model SMI-S defines profiles and recipes within the CIM that enables a management client to reliably use a component vendors implementation of the standard, such as the control of LUNs and zones in the context of a SAN. Consistent use of durable names As a storage network configuration evolves and is re-configured, key long-lived resources, such as disk volumes, must be uniquely and consistently identified over time. Rigorously documented client implementation considerations SMI-S provides client developers with vital information for traversing CIM classes within a device or subsystem and between devices and subsystems such that complex storage networking topologies can be successfully mapped and reliably controlled. An automated discovery system SMI-S compliant products, when introduced in a SAN environment, automatically announce their presence and capabilities to other constituents using SLP (see 2.3.1, SLP architecture on page 38). Resource locking SMI-S compliant management applications from multiple vendors can exist in the same storage device or SAN and cooperatively share resources through a lock manager. The models and protocols in the SMI-S implementation are platform-independent, enabling application development for any platform, and enabling them to run on different platforms. The SNIA also provides interoperability tests which help vendors to test their applications and devices if they conform to the standard. Managers or components that use this standard include: IBM TotalStorage Productivity Center for Disk IBM TotalStorage Productivity Center for Replication IBM TotalStorage Productivity Center for Data
36
Device management
An SNMP manager can read information from an SNMP agent to monitor a device. Therefore the device needs to be polled on an interval basis. The SNMP manager can also change the configuration of a device, by setting certain values to corresponding variables. Managers or components that use these standards include the IBM TotalStorage Productivity Center for Fabric.
Traps
A device can also be set up to send a notification to the SNMP manager (this is called a trap) to asynchronously inform this SNMP manager of a status change. Depending on the existing environment and organization, it is likely that your environment already has an SNMP management application in place. The managers or components that use this standard are: IBM TotalStorage Productivity Center for Fabric (sending and receiving of traps) IBM TotalStorage Productivity Center for Data can be set up to send traps, but does not receive traps IBM TotalStorage Productivity Center for Disk and IBM TotalStorage Productivity Center for Replication events can be sent as SNMP traps by utilizing the IBM Director infrastructure.
37
Note: This MIB is not part of IBM TotalStorage Productivity Center. To learn more about Fibre Alliance and MIB, refer to the following Web sites:
http://www.fibrealliance.org http://www.fibrealliance.org/fb/mib_intro.htm
38
The SA can run in the same process or in a different process as the service itself. In either case, the SA supports registration and de-registration requests for the service (as shown in the right part of Figure 2-5). The service registers itself with the SA during startup, and removes the registration for itself during shutdown. In addition, every service registration is associated with a life-span value, which specifies the time that the registration will be active. In the left part of the diagram, you can see the interaction between a UA and the SA.
A service is required to reregister itself periodically, before the life-span of its previous registration expires. This ensures that expired registration entries are not kept. For instance, if a service becomes inactive without removing the registration for itself, that old registration is removed automatically when its life span expires. The maximum life span of a registration is 65535 seconds (about 18 hours).
39
The SLP UA follows the multicast convergence algorithm and sends repeated multicast messages until no new replies are received. The resulting set of discovered services, including their service URL and any service attributes, are returned to the client application or user. The client application or user is then responsible for contacting the individual services, as needed, using the services URL (see Figure 2-6).
An SLP UA is not required to discover all matching services that exist in the network, but only enough of them to provide useful results. This restriction is mainly due to the transmission size limits for UDP packets. They can be exceeded when there are many registered services or when the registered services have lengthy URLs or a large number of attributes. However, in most modern SLP implementations, the UAs can recognize truncated service replies and establish TCP connections to retrieve all of the information of the registered services. With this type of UA and SA implementation, the only exposure that remains is when there are too many SAs within the multicast range. This can cut short the multicast convergence mechanism. This exposure can be mitigated by the SLP administrator by setting up one or more SLP DAs.
Directory Agent
The SLP DA is an optional component of SLP that collects and caches network service broadcasts. The DA is primarily used to simplify SLP administration and to improve SLP performance. You can consider the SLP DA as an intermediate tier in the SLP architecture. It is placed between the UAs and the SAs so that both UAs and SAs communicate only with the DA instead of with each other. This eliminates a large portion of the multicast request or reply traffic in the network. It also protects the SAs from being overwhelmed by too many service requests if there are many UAs in the environment.
40
Figure 2-7 shows the interactions of the SLP UAs and SAs in an environment with SLP DAs.
Figure 2-7 SLP User Agent interactions with User Agent and Service Agent
When SLP DAs are present, the behavior of both SAs and UAs changes significantly. When an SA is first initializing, it performs a DA discovery using a multicast service request. It also specifies the special, reserved service type service:directory-agent. This process is also called active DA discovery. It is achieved through the same mechanism as any other discovery using SLP. Similarly, in most cases, an SLP UA also performs active DA discovery using multicasting when it first starts. However, if the SLP UA is statically configured with one or more DA addresses, it uses those addresses instead. If it is aware of one or more DAs, either through static configuration or active discovery, it sends unicast service requests to those DAs instead of multicasting to SAs. The DA replies with unicast service replies, providing the requested service URLs and attributes. Figure 2-8 shows the interactions of UAs and SAs with DAs, during active DA discovery.
41
The SLP DA functions similarly to an SLP SA, receiving registration and deregistration requests, and responding to service requests with unicast service replies. There are a couple of differences, where DAs provide more functionality than SAs. One area, mentioned previously, is that DAs respond to service requests of the service:directory-agent service type with a DA advertisement response message, passing back a service URL containing the DAs IP address. This allows SAs and UAs to perform active discovery on DAs. One other difference is that when a DA first initializes, it sends a multicast DA advertisement message to advertise its services to any existing SAs (and UAs) that may already be active in the network. UAs can optionally listen for, and SAs are required to listen for, such advertisement messages. This listening process is also sometimes called passive DA discovery. When the SA finds a new DA through passive DA discovery, it sends registration requests for all its currently registered services to that new DA. Figure 2-9 shows the interactions of DAs with SAs and UAs, during passive DA discovery.
42
SLP communication
SLP uses three methods to send messages across an IP network: unicast, broadcast, or multicast. Data can be sent to one single destination (unicast) or to multiple destinations that are listening at the same time (multicast). The difference between a multicast and a broadcast is quite important. A broadcast addresses all stations in a network. Multicast messages are only used by those stations that are members of a multicast group (that have joined a multicast group).
Unicast
The most common communication method, unicast, requires that a sender of a message identifies one and only one target of that message. The target IP address is encoded within the message packet, and is used by the routers along the network path to route the packet to the proper destination. If a sender wants to send the same message to multiple recipients, then multiple messages must be generated and placed in the network, one message per recipient. When there are many potential recipients for a particular message, then this places an unnecessary strain on the network resources, since the same data is duplicated many times, where the only difference is the target IP address encoded within the messages.
Broadcast
In cases where the same message must be sent to many targets, broadcast is a much better choice than unicast, since it puts much less strain in the network. Broadcasting uses a special IP address, 255.255.255.255, which indicates that the message packet is intended to be sent to all nodes in a network. As a result, the sender of a message needs to generate only a single copy of that message, and can still transmit it to multiple recipients, that is to all members of the network. The routers multiplex the message packet, as it is sent along all possible routes in the network to reach all possible destinations. This puts much less strain on the network bandwidth, since only a single message stream enters the network, as opposed to one message stream per recipient. However, it puts much more strain on the individual nodes (and routers) in the network, since every node receives the message, even though most likely not every node is interested in the message. This means that those members of the network that were not the intended recipients, who receive the message anyway, must receive the unwanted message and discard it. Due to this inefficiency, in most network configurations, routers are configured to not forward any broadcast traffic. This means that any broadcast messages can only reach nodes on the same subnet as the sender.
Multicast
The ability of the SLP to automatically discover services that are available in the network, without a lot of setup or configuration, depends in a large part on the use of IP multicasting. IP multicasting is a broad subject in itself, and only a brief and simple overview is provided here.
Chapter 2. Key concepts
43
Multicasting can be thought of as more sophisticated broadcast, which aims to solve some of the inefficiencies inherent in the broadcasting mechanism. With multicasting, again the sender of a message has to generate only a single copy of the message, saving network bandwidth. However unlike broadcasting, with multicasting, not every member of the network receives the message. Only those members who have explicitly expressed an interest in the particular multicast stream receive the message. Multicasting introduces a concept called a multicast group, where each multicast group is associated with a specific IP address. A particular network node (host) can join one or more multicast groups, which notifies the associated router or routers that there is an interest in receiving multicast streams for those groups. When the sender, who does not necessarily have to be part of the same group, sends messages to a particular multicast group, that message is routed appropriately to only those subnets, which contain members of that multicast group. This avoids flooding the entire network with the message, as is the case for broadcast traffic.
Multicast addresses
The Internet Assigned Numbers Authority (IANA), which controls the assignment of IP addresses, has assigned the old Class D IP address range to be used for IP multicasting. Of this entire range, which extends from 224.0.0.0 to 239.255.255.255, the 224.0.0.* addresses are reserved for router management and communication. Some of the 224.0.1.* addresses are reserved for particular standardized multicast applications. Each of the remaining addresses corresponds to a particular general purpose multicast group. The Service Location Protocol uses address 239.255.255.253 for all its multicast traffic. The port number for SLP is 427, for both unicast and multicast.
Configuration recommendations
Ideally, after IBM TotalStorage Productivity Center is installed, it would discover all storage devices that it can physically reach over the IP network. However in most situations, this is not the case. This is primarily due to the previously mentioned limitations of multicasting and the fact that the majority of routers have multicasting disabled by default. As a result, in most cases without any additional configuration, IBM TotalStorage Productivity Center discovers only those storage devices that reside in its own subnet, but no more. The following sections provide some configuration recommendations to enable TotalStorage Productivity Center to discover a larger set of storage devices.
Router configuration
The vast majority of the intelligence that allows multicasting to work is implemented in the router operating system software. As a result, it is necessary to properly configure the routers in the network to allow multicasting to work effectively. Unfortunately, there is a dizzying array of protocols and algorithms which can be used to configure particular routers to enable multicasting. These are the most common ones: Internet Group Management Protocol (IGMP) is used to register individual hosts in particular multicast groups, and to query group membership on particular subnets. Distance Vector Multicast Routing Protocol (DVMRP) is a set of routing algorithms that use a technique called Reverse Path Forwarding to decide how multicast packets are to be routed in the network. Protocol-Independent Multicast (PIM) comes in two varieties: dense mode (PIM-DM) and sparse mode (PIM-SM). They are optimized to networks where either a large percentage of nodes require multicast traffic (dense), or a small percentage require the traffic (sparse).
44
Multicast Open Shortest Path First (MOSPF) is an extension of OSPF, a link-state unicast routing protocol that attempts to find the shortest path between any two networks or subnets to provide the most optimal routing of packets. The routers of interest are all those which are associated with subnets that contain one or more storage devices which are to be discovered and managed by TotalStorage Productivity Center. You can configure the routers in the network to enable multicasting in general, or at least to allow multicasting for the SLP multicast address, 239.255.255.253, and port, 427. This is the most generic solution and permits discovery to work the way that it was intended by the designers of SLP. To properly configure your routers for multicasting, refer to your router manufacturers reference and configuration documentation. Although older hardware may not support multicasting, all modern routers do. However, in most cases, multicast support is disabled by default, which means that multicast traffic is sent only among the nodes of a subnet but is not forwarded to other subnets. For SLP, this means that service discovery is limited to only those agents which reside in the same subnet.
Firewall configuration
In the case where one or more firewalls are used between TotalStorage Productivity Center and the storage devices that are to be managed, the firewalls need to be configured to pass traffic in both directions, as SLP communication is two way. This means that when TotalStorage Productivity Center, for example, queries an SLP DA that is behind a firewall for the registered services, the response will not use an already opened TCP/IP session but will establish another connection in the direction from the SLP DA to the TotalStorage Productivity Center. For this reason, port 427 should be opened in both directions, otherwise the response will not be received and TotalStorage Productivity Center will not recognize services offered by this SLP DA.
SLP DA configuration
If router configuration is not feasible, another technique is to use SLP DAs to circumvent the multicast limitations. Since with statically configured DAs, all service requests are unicast instead of multicast by the UA, it is possible to simply configure one DA for each subnet that contains storage devices which are to be discovered by TotalStorage Productivity Center. One DA is sufficient for each of such subnets, although more can be configured without harm, perhaps for reasons of fault tolerance. Each of these DAs can discover all services within its own subnet, but no other services outside its own subnet. To allow Productivity Center to discover all of the devices, you must statically configure it with the addresses of each of these DAs. You accomplish this using the IBM Director GUIs Discovery Preference panel. From the MDM SLP Configuration tab, you can enter a list of DA addresses. As described previously, Productivity Center unicasts service requests to each of these statically configured DAs, but also multicasts service requests on the local subnet on which Productivity Center is installed. Figure 2-10 on page 46 displays a sample environment where DAs have been used to bridge the multicast gap between subnets in this manner. Note: At this time, you cannot set up IBM TotalStorage Productivity Center for Data to use remote DAs such as Productivity Center for Disk and Productivity Center for Replication. You need to define all remote CIM Agents by creating a new entry in the CIMOM Login panel or you can register remote services in DA which resides in local subnet. Refer to Configuring the CIM Agents on page 290 for detailed information.
45
You can easily configure an SLP DA by changing the configuration of the SLP SA included as part of an existing CIM Agent installation. This causes the program that normally runs as an SLP SA to run as an SLP DA instead. The procedure to perform this configuration is explained in 6.2, SLP DA definition on page 248. Note that the change from SA to DA does not affect the CIMOM service of the subject CIM Agent, which continues to function as normal, sending registration and de-registration commands to the DA directly.
46
47
The CIM Agent or CIMOM translates a proprietary management interface to the CIM interface. The CIM Agent for the IBM TotalStorage ESS includes a CIMOM inside it. In the future, more and more devices will be native CIM compliant, and will therefore have a built-in Agent as shown in the Embedded Model in Figure 2-11. When widely adopted, SMI-S will streamline the way that the entire storage industry deals with management. Management application developers will no longer have to integrate incompatible feature-poor interfaces into their products. Component developers will no longer have to push their unique interface functionality to application developers. Instead, both will be better able to concentrate on developing features and functions that have value to end-users. Ultimately, faced with reduced costs for management, end-users will be able to adopt storage-networking technology faster and build larger, more powerful networks.
Agent
1 1
0..n
Proprietary
Proprietary
Device or Subsystem
Device or Subsystem
Proxy Model
Embedded Model
Proxy Model
48
The CIM Agent includes a CIMOM, which adapts various devices using a plug-in called a provider. The CIM Agent can work as a proxy or can be embedded in storage devices. When the CIM Agent is installed as a proxy, the IBM CIM Agent can be installed on the same server that supports the device user interface.
49
Lock Manager
SA SA 0..n UA
Directory Manager
DA 0..n
0..n
Proprietary
Proprietary
Device or Subsystem
Device or Subsystem
Proxy Model
Embedded Model
Proxy Model
50
51
The Common Agent provides the platform for the application specific agents. Depending on the tasks for which a subagent is used, the Common Agent is installed on the customers application servers, desktop PCs, or notebooks. Note: In different documentation, Readme files, directory and file names, you also see the terms Common Endpoint, Endpoint, or simply EP. This always refers to the Common Agent, which is part of the Tivoli Common Agent Services. The Common Agent talks to the application specific subagent, with the Agent Manager and the Resource Manager, but the actual system level functions are invoked by the subagent. The information that the subagent collects is sent directly to the Resource Manager by using the applications native protocol. This is enabled to have down-level agents in the same environment, as the new agents that are shipped with the IBM TotalStorage Productivity Center. Certificates are used to validate if a requester is allowed to establish a communication. Demo keys are supplied to quickly set up and configure a small environment, since every installation CD uses the same certificates, this is not secure. If you want to use Tivoli Common Agent Services in a production environment, we recommend that you use your own keys that can be created during the Tivoli Agent Manager installation. One of the most important certificates is stored in the agentTrust.jks file. The certificate can also be created during the installation of Tivoli Agent Manager. If you do not use the demo certificates, you need to have this file available during the installation of the Common Agent and the Resource Manager. This file is locked with a password (the agent registration password) to secure the access to the certificates. You can use the ikeyman utility in the java\jre subdirectory to verify your password. 52
IBM TotalStorage Productivity Center V2.3: Getting Started
53
When you install the software, the agent has to register with the Tivoli Agent Manager. During this procedure, you need to specify the registration port on the manager (by default 9511). Furthermore, you need to specify an agent registration password. This registration is performed by the Common Agent, which is installed automatically if not already installed. If the subagent is deployed from a central location, the port 9510 is by default used by the installer (running on the central machine), to communicate with the Common Agent to download and install the code. When this method is used, no password or certificate is required, because these were already provided during the Common Agent installation on the machine. If you choose to use your own certificate during the Tivoli Agent Manager installation, you need to supply it for the Common Agent installation.
54
Part 2
Part
55
56
Chapter 3.
57
3.1 Configuration
You can install the storage management components of IBM TotalStorage Productivity Center on a variety of platforms. However, for the IBM TotalStorage Productivity Center suite, when all four manager components are installed on the same system, the only common platforms for the managers are: Windows 2000 Server with Service Pack 4 Windows 2000 Advanced Server Windows 2003 Enterprise Edition Note: Refer to the following Web site for the updated support summaries, including specific software, hardware, and firmware levels supported:
http://www.storage.ibm.com/software/index.html
If you are using the storage provisioning workflows, you must install IBM TotalStorage Productivity Center for Disk and TotalStorage Productivity Center for Replication and IBM TotalStorage Productivity Center for Fabric on the same machine. Because of processing requirements, we recommend that you install IBM Tivoli Provisioning Manager on a separate Windows machine.
Hardware
The following hardware is required: Dual Pentium 4 or Intel Xeon 2.4 GHz or faster processors 4 GB of DRAM Network connectivity Subsystem Device Driver (SDD), for IBM TotalStorage Productivity Center for Fabric (optional) 5 GB available disk space.
Database
You must comply with the following database requirements: The installation of DB2 Version 8.2 is part of the Prerequisite Software Installer and is required by all the managers. Other databases that are supported are: For IBM TotalStorage Productivity Center for Fabric: IBM Cloudscape 5.1.60 (provided on the CD) Microsoft SQL Server Version 7.0, 2000 Oracle 8i, 9i, 9i V2 Sybase SQL Server (Adaptive Server Enterprise) Version 12.5 or higher IBM Cloudscape 5.1.60 (provided on the CD) For IBM TotalStorage Productivity Center for Data:
58
9513
80
59
60
61
System 1
The name of our first system was Colorado. The following applications were installed on this system: DB2 IBM Director WebSphere Application Server WebSphere Application Server update Tivoli Agent Manager IBM TotalStorage Productivity Center for Disk and Replication Base IBM TotalStorage Productivity Center for Disk IBM TotalStorage Productivity Center for Replication
62
IBM TotalStorage Productivity Center for Data IBM TotalStorage Productivity Center for Fabric
System 2
The name of our second system was Gallium. The following applications were installed on this server: Data Agent
System 3
The name of our third system was PQDISRV. The following applications were installed on this server: DB2 Application software
CIMOM system 1
Our first CIMOM server was named TPCMAN. On this server, we installed ESS CLI ESS CIMOM LSI Provider (FAStT CIMOM)
CIMOM system 3
Our third CIMOM system was named SVCCON. We installed the following applications on this server: SAN Volume Controller (SVC) Console SVC CIMOM
Networking
We used the following switches for networking: IBM Ethernet 10/100 24 Port switch 2109 F16 Fiber switch
Storage devices
We employed the following storage devices: IBM TotalStorage Enterprise Storage Server (ESS) 800 and F20 DS8000 DS6000 DS4000 IBM SVC Figure 3-1 on page 64 shows a diagram of our lab setup environment.
63
ESS
XXX.YYY.6.29 XXX.YYY.6.26
XXX.YYY.ZZZ.34
XXX.YYY.ZZZ.35
XXX.YYY.ZZZ.73
Colorado Server
W2K -> IBM TotalStorage Productivity Center for Disk, and Replication -> IBM TotalStorage Productivity Center for Fabric
Gallium Server
W2K -> Tivoli Agent Manager -> IBM TotalStorage Productivity Center for Data
PQDISRV Server
W2K
Faroe Server
W2K
->Application Server
XXX.YYY.ZZZ.49
XXX.YYY.ZZZ.36
XXX.YYY.ZZZ.100
XXX.YYY.ZZZ.69
FAStT 700
XXX.YYY.ZZZ.202 XXX.YYY.ZZZ.203
4. Install and configure Simple Network Management Protocol (SNMP) (Fabric requirement). 5. Identify any firewalls and obtain the required authorization. 6. Obtain the static IP addresses that will be used for the TotalStorage Productivity Center servers.
64
Create a token object Increase quotas Replace a process-level token Log on as a service Debug programs
Table 3-8 shows the user IDs that are used in a TotalStorage Productivity Center environment. It provides information about the Windows group to which the user ID must belong, whether it is a new user ID that is created during the installation, and when the user ID is used.
Table 3-8 User IDs used in a IBM TotalStorage Productivity Center environment Element Suite Installer DB2 User ID Administrator db2admina New user No Yes, will be created No Windows DB2 management and Windows Service Account DirAdmin or DirSuper Windows Service Account Type Group or groups Usage
tpcadmina
Windows
65
User ID managerb
Usage Used during the registration of a Resource Manager to the Agent Manager Used to authenticate agents and lock the certificate key files Windows Service Account
Common Agent (see also Common Agent on page 67) Common Agent
AgentMgrb
itcauserb
TotalStorage Productivity Center universal user Tivoli NetView IBM WebSphere Host Authentication
tpccimoma
Windows
DirAdmin
This ID is used to accomplish connectivity with the managed devices. For example, this ID has to be set up on the CIM Agents. See Fabric Manager User IDs on page 68 See Fabric Manager User IDs on page 68 See Fabric Manager User IDs on page 68
a. This account can have any name you choose. b. This account name cannot be changed during the installation. c. The DB2 administrator user ID and password are used here. See Fabric Manager User IDs on page 68.
Granting privileges
Grant privileges to the user ID used to install the IBM TotalStorage Productivity Center for Disk and Replication Base, IBM TotalStorage Productivity Center for Disk, and the IBM TotalStorage Productivity Center for Replication. These user rights are governed by the local security policy and are not initially set as the defaults for administrators. They may not be in effect when you log on as the local administrator. If the IBM TotalStorage Productivity Center installation program does not detect the required user rights for the logged on user name, the program can optionally set them. The program can set the local security policy settings to assign these user rights. Alternatively, you can manually set them prior to performing the installation. To manually set these privileges, follow these steps: 1. Click Start Settings Control Panel. 2. Double-click Administrative Tools. 3. Double-click Local Security Policy. 4. The Local Security Settings window opens. Expand Local Policies. Then double-click User Rights Assignments to see the policies in effect on your system. For each policy added to the user, perform the following steps: a. Highlight the policy to be selected. b. Double-click the policy and look for the users name in the Assigned To column of the Local Security Policy Setting window to verify the policy setting. Ensure that the Local Policy Setting and the Effective Policy Setting options are selected.
66
c. If the user name does not appear in the list for the policy, you must add the policy to the user. Perform the following steps to add the user to the list: i. In the Local Security Policy Setting window, click Add. ii. In the Select Users or Groups window, under the Name column, highlight the user of group. iii. Click Add to place the name in the lower window. iv. Click OK to add the policy to the user or group. 5. After you set these user rights, either by using the installation program or manually, log off the system and then log on again for the user rights to take effect. 6. Restart the installation program to continue with the IBM TotalStorage Productivity Center for Disk and Replication Base.
IBM Director
With Version 4.1, you no longer need to create an internal user account. All user IDs must be operating system accounts and members of one of the following groups: DirAdmin or DirSuper groups (Windows), diradmin, or dirsuper groups (Linux) Administrator or Domain Administrator groups (Windows), root (Linux) In addition, a host authentication password is used to allow managed hosts and remote consoles to communicate with IBM Director.
Resource Manager
The user ID and password (default is manager and password) for the Resource Manager is stored in the AgentManager\config\Authorization.xml file on the Agent Manager. Since this is used only during the initial registration of a new Resource Manager, there is no problem with changing the values at any time. You can find a detailed procedure on how to change this in the Installation and Planning Guides of the corresponding manager. You can have multiple Resource Manager user IDs if you want to separate the administrators for the different managers, for example for IBM TotalStorage Productivity Center for Data and IBM TotalStorage Productivity Center for Fabric.
Common Agent
Each time the Common Agent is started, this context and password are used to validate the registration of the agent with the Tivoli Agent Manager. Furthermore the password is used to lock the certificate key files (agentTrust.jks). The default password is changeMe, but you should change the password when you install the Tivoli Agent Manager. The Tivoli Agent Manager stores this password in the AgentManager.properties file. If you start with the defaults, but want to change the password later, all the agents have to be changed. A procedure to change the password is available in the Installation and Planning Guides of the corresponding managers (at this time Data or Fabric). Since the password is used to lock the certificate files, you must also apply this change to Resource Managers.
67
The Common Agent user ID AgentMgr is not a user ID, but rather the context in which the agent is registered at the Tivoli Agent Manager. There is no need to change this, so we recommend that you accept the default.
68
Common Agent Productivity Center for Data Productivity Center for Fabric
69
Comment
70
3. In the Windows Management Instrumentation Properties window, go down to Service status and click the Stop button (Figure 3-3). Wait for the service to stop.
71
4. After the service is stopped, in the Windows Management Instrumentation Properties window, change the Startup type to Disabled (Figure 3-4) and click OK.
5. After disabling the service, it may start again. If so, go back and stop the service again. The service should now be stopped and disabled as shown in Figure 3-5.
Important: After the Prerequisite Software installation completes. You must enable the Windows Management Instrumentation service before installing the suite. To enable the service, change the Startup type from Disabled (see Figure 3-4) to Automatic.
72
73
e. Click the Security tab (see Figure 3-7). f. Select Accept SNMP packets from any host. g. Click OK.
11.After you set the public community name, restart the SNMP community service. 74
IBM TotalStorage Productivity Center V2.3: Getting Started
75
Cloudscape database
If you install TotalStorage Productivity Center for Fabric and specify the Cloudscape database, you need the following user IDs and passwords: Agent manager name or IP address and password Common agent password to register with the Agent Manager Resource manager user ID and password to register with the Agent Manager WebSphere administrative user ID and password host authentication password Tivoli NetView password only
DB2 database
If you install IBM TotalStorage Productivity Center for Fabric and specify the DB2 database, you need the following user IDs and passwords: Agent manager name or IP address and password Common agent password to register with the Agent Manager Resource manager user ID and password to register with the Agent Manager DB2 administrator user ID and password DB2 user ID and password WebSphere administrative user ID and password Host authentication password only Tivoli NetView password only Note: If you are running Windows 2000, when the IBM TotalStorage Productivity Center for Fabric installation program asks for an existing user ID for WebSphere, that user ID must act as part of the operating system user.
WebSphere
To change the WebSphere user ID and password, follow this procedure: 1. Open the install_location\apps\was\properties\soap.client.props file.
76
2. Modify the following entries: com.ibm.SOAP. login Userid=user_ID (enter a value for user_ID) com.ibm.SOAP. login Password=password (enter a value for password) 3. Save the file. 4. Run the following script:
ChangeWASAdminPass.bat user_ID password install_dir
Here user_ID is the WebSphere user ID and password is the password. install_dir is the directory where the manager is installed and is optional. For example, install_dir is c:\Program Files\IBM\TPC\Fabric\manager\bin\W32-ix86.
Security considerations
You set up security by using certificates. There are demonstration certificates or you can generate new certificates. This option is specified when you installed the Agent Manager. See Figure 4-26 on page 104. We recommend that you generate new certificates. If you used the demonstration certificates, continue with the installation. If you generated new certificates, follow this procedure: 1. Copy the manager CD image to your computer. 2. Copy the agentTrust.jks file from the Agent Manager (AgentManager/certs directory) to the /certs directory of the manager CD image. This overwrites the existing agentTrust.jks file. 3. You can write a new CD image with the new file or keep this image on your computer and point the Suite Installer to the directory when requested.
77
# # # # # # # # # #
The IP address and the host name should be separated by at least one space. Additionally, comments (such as these) may be inserted on individual lines or following the machine name denoted by a '#' symbol. For example: 102.54.94.97 38.25.63.10 rhino.acme.com x.acme.com # source server # x client host
Note: Host names are case sensitive, which is limitation within WebSphere. Check your host name.
78
RDBMS support
Data Manager currently supports the monitoring of the following relational database management systems (RDBMS): Microsoft SQL Server 7.0, 2000 Oracle 8i, 9i, 9i V2, 10G Sybase SQL Server 11.0.9 and higher DB2 Universal Database (UDB) 7.1, 7.2, 8.1, 8.2 (64-bit UDB DB2 instances are supported)
79
User levels
There are two levels of users within IBM TotalStorage Productivity Center for Data: non-administrator users and administrator users. The level of users determine how they use IBM TotalStorage Productivity Center for Data. Non-administrator users View the data collected by IBM TotalStorage Productivity Center for Data. Create, generate, and save reports. IBM TotalStorage Productivity Center for Data administrators. These users can: Create, modify, and schedule Pings, Probes, and Scans Create, generate, and save reports Perform administrative tasks and customize the IBM TotalStorage Productivity Center for Data environment Create Groups, Profiles, Quotas, and Constraints Set alerts Important: Security is set up by using the certificates. You can use the demonstration certificates or you can generate new certificates. It is recommended that you generate new certificates when you install the Agent Manager.
Certificates
If you generated new certificates, follow this procedure: 1. Copy the CD image to your computer. 2. Copy the agentTrust.jks file from the Agent Manager directory AgentManager/certs to the CommonAgent\certs directory of the manager CD image. This overwrites the existing agentTrust.jks file. You can write a new CD image with the new file or keep this image on your computer and point the Suite Installer to the directory when requested. Important: Before installing IBM TotalStorage Productivity Center for Data, define the group within your environment that will have administrator rights within Data Manager. This group must exist on the same machine where you are installing the Server component. During the installation, you are prompted to enter the name of this group.
80
81
3. In the window that opens (Figure 3-9), complete the required database name information. We used the database name of TPCDATA. Click Finish to complete the database creation.
82
Chapter 4.
83
4.1.1 Considerations
You may want to use IBM TotalStorage Productivity Center for Disk to manage the IBM TotalStorage Enterprise Storage Server (ESS), DS8000, DS6000, Storage Area Network (SAN) Volume Controller (SVC), IBM TotalStorage Fibre Array Storage Technology (FAStT), or DS4000 storage subsystems. In this case, you must install the prerequisite input/output (I/O) Subsystem Licensed Internal Code (SLIC) and Common Information Model (CIM) Agent for the devices. See Chapter 6, Configuring IBM TotalStorage Productivity Center for Disk on page 247, for more information. If you are installing the CIM Agent for the ESS, or the DS8000 or DS6000 you must install it on a separate machine. TotalStorage Productivity Center 2.3 does not support Linux on zSeries or on S/390. Nor does IBM TotalStorage Productivity Center support Windows domains.
84
85
3. The Prerequisite Software Installer wizard welcome pane in Figure 4-2 opens. Click Next. The Software License Agreement panel is then displayed. Read the terms of the license agreement. If you agree with the terms of the license agreement select the I accept the terms in the license agreement radio button and click Next to continue.
4. The prerequisite operating system check panel in Figure 4-3 on page 87 opens. When it completes successfully click Next.
86
5. The Tivoli Common Directory location panel (Figure 4-4) opens and prompts for a location for the log files. Accept the default location or enter a different location. Click Next to continue.
87
6. The product selection panel (Figure 4-5) opens. To install the entire TotalStorage Productivity Center suite, check the boxes next to DB2, WebSphere, and Agent Manager.
7. The DB2 Universal Database panel (Figure 4-6) opens. Select Enterprise Server Edition and click Next to continue.
88
Note: After clicking Next (Figure 4-6), if you see the panel in Figure 4-7, you must first stop and disable the Windows Management Instrumentation service before continuing with the installation. See Figure 3.7 on page 70 for detailed instructions.
89
8. The DB2 user name and password panel (Figure 4-8) opens. If the DB2 user name exists on the system, the correct password must be entered or the DB2 installation will fail. If the DB2 user name does not exist it will be created by the DB2 install. In our installation we accepted the default user name and entered a unique password. Click Next to continue.
90
9. The Target Directory Confirmation panel (Figure 4-9) opens. Accept the default target directories for DB2 installation or enter a different location. Click Next.
10.The select the languages panel (Figure 4-10) opens. This installs the languages selected for DB2. Select your desired language(s). Click Next.
91
11.The Preview Prerequisite Software Information panel (Figure 4-11) opens. Review the information and click Next.
12.The WebSphere Application Server system prerequisites check panel (Figure 4-12) opens. When the check completes successfully click Next.
92
13.The installation options panel (Figure 4-13) opens. Select the type of installation you wish to perform. The rest of this section guides you through Unattended Installation. Unattended Installation guides you through copying all installation images to a central location called the installation image depot. Once the copies are completed, the component installations proceed with no further intervention needed. Attended Installation prompts you to enter the location of each install image as needed. Click Next to continue.
93
14.The install image depot location panel opens (see Figure 4-14). Enter the location where all installation images are to be copied. Click Next.
15.You are first prompted for the location of the DB2 installation image (see Figure 4-15). Browse to the installation image and select the path to the installation files or insert the install CD and click Copy.
94
16.After the DB2 installation image is copied to the install image depot, you are prompted for the location of the WebSphere installation image (see Figure 4-16). Browse to the installation image and select the path to the installation files or insert the install CD and click Copy.
17.After the WebSphere installation image is copied, you are prompted for the location of the WebSphere Cumulative fix 3 installation image (see Figure 4-17). Browse to the installation image and select the path to the installation files or insert the install CD and click Copy.
95
18.When an install image has been successfully copied to the Install Image Depot, a green check mark appears to the right of the prerequisite. After all the prerequisite software images are successfully copied to the install image depot (Figure 4-18), click Next.
96
19.The installation of DB2, WebSphere, and the WebSphere Fix Pack begins. When a prerequisite is successfully installed, a green check mark appears to its left. If the installation of a prerequisite fails, a red X appears to the left. If a prerequisite installation fails, exit the installer, check the logs to determine and correct the problem, and restart the installer. When the installation completes successfully (see Figure 4-19), click Next.
97
20.The Agent Manager Registry Information panel opens. Select the type of database, specify the database name, and choose a local or remote database. The default DB2 database name is IBMCDB. For a local database connection, the DB2 database will be created if it does not exist. We recommend that you take the default database name for a local database. Click Next to continue (see Figure 4-20). Attention: For a remote database connection, the database specified below must exist. Refer to 3.12.4, Creating the DB2 database on page 81 for information on how to create a database in DB2.
98
21.The Database Connection Information panel in Figure 4-21 opens. Specify the location of the database software directory (for DB2, the default install location is C:\Program Files\IBM\SQLLIB), the database user name and password. You must specify the database host name and port if you are using a remote database. Click Next to continue.
99
Note: For a remote database connection the database specified in Figure 4-20 on page 98 must exist. If the database does not exist, you will see the error message shown in Figure 4-22. Refer to 3.12.4, Creating the DB2 database on page 81 for information on how to create a database in DB2.
100
22.A panel opens prompting for a location to install Tivoli Agent Manager (see Figure 4-23). Accept the default location or enter a different location. Click Next to continue.
101
23.The WebSphere Application Server Information panel (Figure 4-24) opens. This panel lets you specify the host name or IP address, and the cell and node names on which to install the Agent Manager. If you specify a host name, use the fully qualified host name. For example, specify HELIUM.almaden.ibm.com. If you use the IP address, use a static IP address. This value is used in the URLs for all Agent Manager services. We recommend that you use the fully qualified host name, not the IP address of the Agent Manager server. Typically the cell and node name are both the same as the host name of the computer. If WebSphere was installed before you started the Agent Manager installation wizard, you can look up the cell and node name values in the %WebSphere Application Server_INSTALL_ROOT%\bin\SetupCmdLine.bat file. You can also specify the ports used by the Agent Manager. We recommend that you accept the defaults. Registration Port: The default is 9511 for the server-side Secure Sockets Layer (SSL). Secure Port: The default is 9512 for client authentication, two-way SSL. Public Port: The default is 9513. If you are using WebSphere network deployment or a customized deployment, make sure that the cell and node names are correct. For more information about WebSphere deployment, see your WebSphere documentation. After filling in the required information in the WebSphere Application Server Information panel, click Next.
102
Note: If an IP address is entered in the WebSphere Application Server Information panel shown in Figure 4-24, the next panel (see Figure 4-25) explains why a host name is recommended. Click Back to use a host name or click Next to use the IP address.
103
24.The Security Certificates panel (Figure 4-26) opens. Specify whether to create new certificates or to use the demonstration certificates. In a typical production environment, you would create new certificates. The ability to use demonstration certificates is provided as a convenience for testing and demonstration purposes. Make a selection and click Next.
104
25.The Security Certificate Settings panel (see Figure 4-27) opens. Specify the certificate authority name, security domain, and agent registration password. The agent registration password is used to register the agents. You must provide this password when you install the agents. This password also sets the Agent Manager key store and trust store files. Record this password, it will be used again in the installation process. The domain name is used in the right-hand portion of the distinguished name (DN) of every certificate issued by the Agent Manager. It is the name of the security domain defined by the Agent Manager. Typically, this value is the registered domain name or contains the registered domain name. For example, for the computer system myserver.ibm.com, the domain name is ibm.com. This value must be unique in your environment. If you have multiple Agent Managers installed, this value must be different on each Agent Manager. The default agent registration password is changeMe and it is case sensitive. Click Next to continue.
105
26.The User input summary panel for Agent Manager (see Figure 4-28) opens. Review the information and click Next.
106
27.The summary information for Agent Manager panel (see Figure 4-29) opens. Click Next.
107
28.You will see a panel indicating the status of the Agent Manager install process. the IBMCDB database will be created and tables are added to the database. Once the installation of agent manager completes the Summary of Installation and Configuration Results panel (see Figure 4-30) opens. Click Next to continue.
108
29.The next panel (Figure 4-31) informs you when the Agent Manager service started successfully. Click Finish.
109
30.The next panel (Figure 4-32) indicates the installation of prerequisite software is complete. Click Finish to exit the prerequisite installer.
1. Insert the IBM TotalStorage Productivity Center Suite Installer CD into the CD-ROM drive. If Windows autorun is enabled, the installation program should start automatically. If it does not, open Windows Explorer and go to the IBM TotalStorage Productivity Center CD-ROM drive. Double-click setup.exe. Note: It may take a few moments for the installer program to initialize. Be patient. Eventually, you see the language selection panel (Figure 4-33). 2. The Installer language window (see Figure 4-33) opens. From the list, select a language. This is the language used to install this product. Click OK.
3. You see the Welcome to the InstallShield Wizard for The IBM TotalStorage Productivity Center panel (see Figure 4-34). Click Next.
4. The Software License Agreement panel (Figure 4-35 on page 112) opens. Read the terms of the license agreement. If you agree with the terms of the license agreement, select the I accept the terms of the license agreement radio button. Then click Next. If you do not accept the terms of the license agreement, the installation program ends without installing IBM TotalStorage Productivity Center components.
111
5. The next panel enables you to select the type of installation (Figure 4-36). Select Manager installations of Data, Disk, Fabric, and Replication and then click Next.
112
6. In the next panel (see Figure 4-37), select the components that you want to install. Click Next to continue.
113
7. The suite installer installs the IBM Director first (see Figure 4-38). Click Next.
8. The IBM Director installation is now ready to begin (see Figure 4-39). Click Next.
114
9. The package location for IBM Director panel (see Figure 4-40) opens. Enter the appropriate information and click Next. Note: Make sure the Windows Management Instrumentation service is disabled (see Figure 3.7 on page 70 for detailed instructions). If it is enabled, a window appears prompting you to disable the service after you click Next to continue.
10.The next panel (see Figure 4-41) provides information about the IBM Director post installation reboot option. When prompted, choose the option to reboot later. Click Next.
115
11.The IBM Director Server - InstallShield Wizard panel (Figure 4-42) opens. It indicates that the IBM Director installation wizard will launch. Click Next.
12.The License Agreement window opens (Figure 4-43). Read the license agreement. Click I accept the terms in the license agreement radio button and then click Next.
116
13.The next window (Figure 4-44) displays an advertisement for Enhance IBM Director with the new Server Plus Pack window. Click Next.
14.The Feature and installation directory window (Figure 4-45) opens. Accept the default settings and click Next.
117
15.The IBM Director service account information window (see Figure 4-46) opens. a. Type the domain for the IBM Director system administrator. Alternatively, if there is no domain, then type the local host name (the recommended setup). b. Type a user name and password for IBM Director. The IBM Director will run under this user name and you will log on to the IBM Director console using this user name. In our installation we used the user ID we created to install the TotalStorage Productivity Center. This user must be part of the Administrator group. c. Click Next to continue.
16.The Encryption settings window (Figure 4-47) opens. Accept the default settings in the Encryption settings window. Click Next.
118
17.In the Software Distribution settings window (Figure 4-48), accept the default values and click Next. Note: The TotalStorage Productivity Center components do not use the software-distribution packages function of IBM Director.
18.The Ready to Install the Program window (Figure 4-49) opens. Click Install.
119
19.The Installing IBM Director server window (Figure 4-50) reports the status of the installation.
20.The Network driver configuration window (Figure 4-51) opens. Accept the default settings and click OK.
The secondary window closes and the installation wizard performs additional actions which are tracked in the status window.
120
21.The Select the database to be configured window (Figure 4-52) opens. Select IBM DB2 Universal Database and click Next.
22.The IBM Director DB2 Universal Database configuration window (Figure 4-53) opens. It may be behind the status window. You must click this window to bring it to the foreground. a. In the Database name field, type a new database name for the IBM Director database table or type an existing database name. b. In the User ID and Password fields, type the DB2 user ID and password that you created during the DB2 installation. c. Click Next to continue.
121
23.In the IBM Director DB2 Universal Database configuration secondary window (Figure 4-54), accept the default DB2 node name LOCAL - DB2. Click OK.
24.The Database configuration in progress window is displayed at the bottom of the IBM Director DB2 Universal Database configuration window. Wait for the configuration to complete and the secondary window to close. 25.When the InstallShield Wizard Completed window (Figure 4-55) opens, click Finish.
Important: Do not reboot the machine at the end of the IBM Director installation. The Suite Installer reboots the machine.
122
26.When you see IBM Director Server Installer Information window (Figure 4-56), click No.
Important: Are you installing IBM TotalStorage Productivity Center for Data? If so, have you created the database for IBM TotalStorage Productivity Center for Data or are you using a existing database? If you are installing Tivoli Disk manager, you must have created the administrative superuser ID and group and set the privileges. 27.The Install Status panel (see Figure 4-57) opens after a successful installation. Click Next.
123
28.In the machine reboot window (see Figure 4-58), click Next to reboot the machine. Important: If the server does not reboot at this point, cancel the installer and reboot the server.
124
4.3.3 IBM TotalStorage Productivity Center for Disk and Replication Base
There are three separate installations to perform: Install the IBM TotalStorage Productivity Center for Disk and Replication Base code. Install the IBM TotalStorage Productivity Center for Disk. Install the IBM TotalStorage Productivity Center for Replication. IBM TotalStorage Productivity Center for Disk and Replication Base must be installed by a user who is logged on as a local administrator (for example, as the administrator user) on the system where the IBM TotalStorage Productivity Center for Disk and Replication Base will be installed. If you intend to install IBM TotalStorage Productivity Center for Disk and Replication Base as a server, you need the following required system privileges, called user rights, to successfully complete the installation as described in 3.5.1, User IDs on page 65. Act as part of the operating system Create a token object Increase quotas Replace a process level token Debug programs After rebooting the machine the installer initializes to continue the suite install. A window opens prompting you to select the installation language to be used for this wizard (Figure 4-59). Select the language and click OK.
Figure 4-59 Selecting the language for the IBM TotalStorage Productivity Center installation wizard
125
1. The next panel enables you to select the type of installation (Figure 4-60). Select Manager installations of Data, Disk, Fabric, and Replication and click Next.
2. The next window (Figure 4-61) opens allowing you to select which components to install. Select the components you wish to install (all components in this case) and click Next.
126
3. The installer checks that all prerequisite software is installed on your system (see Figure 4-62). Click Next.
4. Figure 4-63 shows the Installer window about to begin installation of Productivity Center for Disk and Replication Base. The window also displays the products that are yet to be installed. Click Next to begin the installation.
127
5. The Package Location for Disk and Replication Manager window (Figure 4-64) opens. Enter the appropriate information and click Next.
Figure 4-64 Package location for Productivity Center Disk and Replication
6. The Information for Disk and Replication Base Manager panel (see Figure 4-65) opens. Click Next.
128
7. The Welcome panel (see Figure 4-66) opens. It indicates that the Disk and Replication Base Manager installation wizard will be launched. Click Next.
Figure 4-66 IBM TotalStorage Productivity Center for Disk and Replication Base welcome information
129
8. In the Destination Directory panel (Figure 4-67), you confirm the target directories. Enter the directory path or accept the default directory and click Next.
Figure 4-67 IBM TotalStorage Productivity Center for Disk and Replication Base Installation directory
9. In the IBM WebSphere Instance Selection panel (see Figure 4-68), click Next.
130
10.If the installation user ID privileges were not set, you see an information panel stating that you need to set the privileges (see Figure 4-69). Click Yes.
11.The required user privileges are set and an informational window opens (see Figure 4-70). Click OK.
12.At this point, the installation terminates. You must close the installer. Log off of Windows, log back on again, and then restart the installer.
131
13.In the Installation Type panel (Figure 4-71), select Typical and click Next.
Figure 4-71 IBM TotalStorage Productivity Center for Disk and Replication Base type of installation
14.If the IBM Director Support Program and IBM Director Server service is still running, the Servers Check panel (see Figure 4-72) opens and prompts you to stop the services. Click Next to stop the services.
132
15.In the User Name Input 1 of 2 panel (Figure 4-73), enter the name and password for the IBM TotalStorage Productivity Center for Disk and Replication Base super user ID. This user name must be defined to the operating system. In our environment we used tpccimom as our super user. After entering the required information click Next to continue.
Figure 4-73 IBM TotalStorage Productivity Center for Disk and Replication Base superuser information
16.If the specified super user ID is not defined to the operating system a window asking if you would like to create it appears (see Figure 4-74). Click Yes to continue.
133
17.In the User Name Input 2 of 2 panel (Figure 4-75), enter the user name and password for the IBM DB2 Universal Database Server. This is the user ID that was specified when DB2 was installed (see Figure 4-8 on page 90). Click Next to continue.
Figure 4-75 IBM TotalStorage Productivity Center for Disk and Replication Base DB2 user information
134
18.The SSL Configuration panel (Figure 4-76) opens. If you selected IBM TotalStorage Productivity Center for Disk and Replication Base Server, then you must enter the fully qualified name of the two server key files that were generated previously or that must be generated during or after the IBM TotalStorage Productivity Center for Disk and Replication Base installation. The information that you enter will be used later. a. Choose either of the following options: Generate a self-signed certificate: Select this option if you want the installer to automatically generate these certificate files. We generate the certificates in our installation. Defer the generation of the certificate as a manual post-installation task: Select this option if you want to manually generate these certificate files after the installation, using WebSphere Application Server ikeyman utility.
b. Enter the Key file and Trust file passwords. The passwords must be a minimum of six characters in length and cannot contain spaces. You should record the passwords in the worksheets provided in Appendix A, Worksheets on page 991. c. Click Next.
135
The Generate Self-Signed Certificate window opens (see Figure 4-77). Complete all the required fields and click Next to continue.
Figure 4-77 IBM TotalStorage Productivity Center for Disk and Replication Base Certificate information
136
19.Next you see the Create Local Database window (Figure 4-78). Accept the default database name of DMCOSERV, or optionally enter the database name. Click Next to continue. Note: The database name must be unique to IBM TotalStorage Productivity Center for Disk and Replication Base. You cannot share the IBM TotalStorage Productivity Center for Disk and Replication Base database with any other applications.
Figure 4-78 IBM TotalStorage Productivity Center for Disk and Replication Base database name
137
20.The Preview window (Figure 4-79) displays a summary of all of the choices that you made during the customizing phase of the installation. Click Install to complete the installation.
Figure 4-79 IBM TotalStorage Productivity Center for Disk and Replication Base Installer information
138
21.The DB2 database is created, the keys are generated, and the Productivity Center for Disk and Replication base is installed. The Finish window opens. You can view the log file for any possible error messages. The log file is located in installeddirectory\logs\dmlog.txt. The dmlog.txt file contains a trace of the installation actions. Click Finish to complete the installation.
Figure 4-80 Productivity Center for Disk and Replication Base Installer - Finish
Notepad opens and displays the post-installation tasks information. Read the information and complete any required tasks. 22.The Install Status window (Figure 4-81) opens after the successful Productivity Center for Disk and Replication Base installation. Click Next.
Figure 4-81 Install Status for Productivity Center for Disk and Replication Base successful
139
1. A window (Figure 4-83) opens that prompts you for the package location for CD-ROM labeled IBM TotalStorage Productivity Center for Disk. Enter the appropriate information and click Next.
140
2. The next window that opens indicates that the IBM TotalStorage Productivity Center for Disk installer wizard will be launched (see Figure 4-84). Click Next.
3. The Productivity Center for Disk Installer - Welcome panel (see Figure 4-85) opens. Click Next.
Figure 4-85 IBM TotalStorage Productivity Center for Disk Installer Welcome
141
4. The Destination Directory panel (Figure 4-86) opens. Enter the directory path or accept the default directory and click Next.
5. The Installation Type panel (Figure 4-87) opens. Select Typical and click Next.
142
6. The Create Local Database panel (Figure 4-88) opens. Accept the default database name of PMDATA or re-enter a new database name. Then click Next.
Figure 4-88 IBM TotalStorage Productivity Center for Disk - Create Local Database
143
7. Review the information on the IBM TotalStorage Productivity Center for Disk Preview panel (Figure 4-89) and click Install.
Figure 4-89 IBM TotalStorage Productivity Center for Disk Installer - Preview
8. The installer creates the required database (see Figure 4-90) and installs the product. You see a progress bar for the Productivity Center for Disk installation status.
144
9. When the installation is complete, you see the Finish panel (Figure 4-91). Review the post installation tasks. Click Finish.
10.The Install Status window (Figure 4-92) opens after the successful Productivity Center for Disk installation. Click Next.
Figure 4-92 Install Status for Productivity Center for Disk successful
145
1. The Package Location for Replication Manager panel (Figure 4-94) opens. Enter the appropriate information and click Next.
146
2. The next window that opens indicates that the IBM TotalStorage Productivity Center for Replication installer wizard will be launched (see Figure 4-95). Click Next.
3. The Welcome window (Figure 4-96) opens. It suggests documentation that you can review prior to the installation. Click Next to continue or click Cancel to exit the installation.
Figure 4-96 IBM TotalStorage Productivity Center for Replication Installer Welcome
147
4. The Destination Directory panel (Figure 4-97) opens. Enter the directory path or accept the default directory. Click Next to continue.
Figure 4-97 IBM TotalStorage Productivity Center for Replication Installer Destination Directory
148
5. The next panel (see Figure 4-98) asks you to select the installation type. Select the Typical radio button and click Next.
149
6. In the Create Local Database for Hardware Subcomponent window (see Figure 4-99), in the Database name field, enter a value for the new Hardware subcomponent database or accept the default. We recommend that you accept the default. Click Next. Note: The database name must be unique to the Replication Manager subcomponent. You cannot share the Replication Manager subcomponent database with any other applications or with other Replication Manager subcomponents.
Figure 4-99 IBM TotalStorage Productivity Center for Replication: Hardware subcomponent
150
7. In the Create Local Database for ElementCatalog Subcomponent window (see Figure 4-100), in the Database name field, enter for the new Element Catalog subcomponent database or accept the default. Click Next. Note: The database name must be unique to the Replication Manager subcomponent. You cannot share the Replication Manager subcomponent database with any other applications or with other Replication Manager subcomponents.
Figure 4-100 IBM TotalStorage Productivity Center for Replication: Element Catalog subcomponent
151
8. In the Create Local Database for ReplicationManager Subcomponent window (see Figure 4-101), in the Database name field, enter the new Replication Manager subcomponent database or accept the default. Click Next. Note: The database name must be unique to the Replication Manager subcomponent. You cannot share the Replication Manager subcomponent database with any other applications or with other Replication Manager subcomponents.
Figure 4-101 TotalStorage Productivity Center for Replication: Replication Manager subcomponent
152
9. In the Create Local Database for ReplicationManager Subcomponent window (see Figure 4-102), in the Database name field, enter the new SVC hardware subcomponent database or accept the default. Click Next. Note: The database name must be unique to the Replication Manager subcomponent. You cannot share the Replication Manager subcomponent database with any other applications or with other Replication Manager subcomponents.
Figure 4-102 IBM TotalStorage Productivity Center for Replication: SVC Hardware subcomponent
153
10.The Setting Tuning Cycle Parameter window (Figure 4-103) opens. Accept the default value of tuning every 24 hours or change the value. You can change this value later in the ElementCatalog.properties file. Click Next.
Figure 4-103 IBM TotalStorage Productivity Center for Replication: Database tuning cycle
154
11.Review the information in the TotalStorage Productivity Center for Replication Installer Preview panel (Figure 4-104). Click Install.
Figure 4-104 IBM TotalStorage Productivity Center for Replication Installer Preview
155
12.You see the Productivity Center for Replication Installer - Finish panel (see Figure 4-105) upon successful installation. Read the post installation tasks. Click Finish to complete the installation.
13.The Install Status window (Figure 4-106) opens after the successful Productivity Center for Disk and Replication Base installation. Click Next.
Figure 4-106 Install Status for Productivity Center for Replication successful
156
157
1. The panel that opens prompts you to specify the location of the install package for Productivity Center for Fabric Manager (see Figure 4-108). Enter the appropriate path and click Next. Important: If you used the demonstration certificates, point to the CD-ROM drive. If you generated new certificates, point to the manager CD image with the new agentTrust.jks file.
158
2. The next window that opens indicates that the IBM TotalStorage Productivity Center for Fabric installer wizard will be launched (see Figure 4-109). Click Next.
3. A window opens in which you select the language to use for the wizard (see Figure 4-110). Select the required language and click OK.
Figure 4-110 IBM TotalStorage Productivity Center for Fabric installer: Selecting the language
159
4. A panel opens asking you to select the type of installation you wish to perform (Figure 4-111). In this case, we install the IBM TotalStorage Productivity Center for Fabric code. You can also use the Suite Installer to perform remote deployment of the Fabric Agent. You can perform this operation only if you installed the common agent previously on machines. For example, you may have installed the Data Agent on the machines and want to add the Fabric Agent to the same machines. You must have the Fabric Manager installed before you can deploy the Fabric Agent. You cannot select both Fabric Manager Installation and Remote Fabric Agent Deployment at the same time. You can only select one option. Click Next.
160
Figure 4-112 IBM TotalStorage Productivity Center for Fabric: Welcome information
6. The next window that opens prompts you to confirm the target directory (see Figure 4-113). Enter the directory path or accept the default directory. Click Next.
Figure 4-113 IBM TotalStorage Productivity Center for Fabric installation directory
161
7. In the next panel (see Figure 4-114), you specify the port number. This is a range of 25 port numbers for use by IBM TotalStorage Productivity Center for Fabric. The first port number that you specify is considered the primary port number. You only need to enter the primary port number. The primary port number and the next 24 numbers are reserved for use by IBM TotalStorage Productivity Center for Fabric. For example, if you specify port number 9550, IBM TotalStorage Productivity Center for Fabric uses port numbers 9550 through 9574. Ensure that the port numbers you use are not used by other applications at the same time. To determine which port numbers are in use on a particular computer, type either of the following commands from a command prompt.
netstat -a netstat -an
We recommend that you use the first of these two commands. The port numbers in use on the system are listed in the Local Address column of the output. This field has the format host:port. Enter the primary port number and click Next.
Figure 4-114 IBM TotalStorage Productivity Center for Fabric port number
162
8. As shown in Figure 4-115, select the database repository, either DB2 or Cloudscape. If you select DB2, you must have previously installed DB2 on the server. DB2 is the recommended installation option. Click Next.
Figure 4-115 IBM TotalStorage Productivity Center for Fabric database selection type
9. In the next panel (see Figure 4-117 on page 164), select the WebSphere Application Server to use in the installation. WebSphere Application Server was installed as part of the prerequisite software so we chose the Non Embedded (Full) WebSphere Application Server option. If the Fabric manager is to be installed standalone on a server choose the Embedded WebSphere Application Server - Express option. Click Next.
Figure 4-116 Productivity Center for Fabric WebSphere Application Server type selection
163
10.The Single/Multiple User ID/Password Choice panel (see Figure 4-117), using DB2, opens. If you select DB2 as your database, you see this panel. This panel allows you to use the DB2 administrative user ID and password for the DB2, WebSphere, Host Authentication, and NetView. If you select all the boxes, you are only prompted for the DB2 user ID and password which is used for all instances. In our install we only selected DB2 and NetView. A different user ID and password will be used for WebSphere and Host Authentication. Note: If you selected IBM Cloudscape as your database, this panel is not displayed. Click Next.
Figure 4-117 IBM TotalStorage Productivity Center for Fabric user and password options
164
11.The DB2 Administrator user ID and password panel (Figure 4-118), using DB2, opens. If you selected DB2 as your database, you see this panel. This panel allows you to use the DB2 administrative user ID and password for the DB2. The user ID and password specified during the DB2 installation in Figure 4-8 on page 90 was used in this example. Enter the required user ID and password. Click Next. The installer will verify that the user ID entered exists.
Figure 4-118 IBM TotalStorage Productivity Center for Fabric database user information
165
12.In the next window (see Figure 4-119) that opens, type the name of the new database in the Type database name: field or accept the default. In our install we accepted the default database name. Click Next. Note: The database name must be unique. You cannot share the IBM TotalStorage Productivity Center for Fabric database with any other applications.
Figure 4-119 IBM TotalStorage Productivity Center for Fabric database name
13.Since we did not check the box for WebSphere in Figure 4-117 on page 164, the panel in Figure 4-120 on page 167 opens prompting for a WebSphere user ID and password. We used the tpcadmin user ID, which is what we used for the IBM Director service account (refer to Figure 4-46 on page 118). Enter the required information and click Next.
166
14.Since we also did not check the box for Host Authentication (Figure 4-117 on page 164), the following panel (Figure 4-121) opens. Enter the password for Host Authentication. This password is used by the Fabric agents. Click Next.
15.In the window (Figure 4-122 on page 168) that opens, enter the parameters for the Tivoli NetView drive name. Click Next.
167
Figure 4-122 IBM TotalStorage Productivity Center for Fabric database drive information
16.The Agent Manager Information panel (Figure 4-123 on page 169) opens. You must complete the following fields: Agent manager name or IP address: This is the host name or IP address of your Agent Manager. Agent manager registration port: This is the port number of your Agent Manager. The default value is 9511. Agent Manager public port: This is a public port. The default value is 9513. Agent registration password (twice): This is the password used to register the common agent with the Agent Manager as shown in Figure 4-27 on page 105. If the password is not set and the default is accepted, the password is changeMe. This password is case sensitive. The agent registration password resides in the AgentManager.properties file where the Agent Manager is installed. It is located in the following directory: %WSAS_INSTALL_ROOT%\InstalledApps\<cell>\AgentManager.ear\AgentManag er.war\WEB-INF\classes\resource Resource manager registration user ID: This is the user ID used to register the resource manager with the Agent Manager. The default is manager. The Resource Manager registration user ID and password reside in the Authorization.xml file where the Agent Manager is installed. It is located in the following directory: <Agent_Manager_install_dir>\config Resource manager registration password (twice): This is the password used to register the resource manager with the Agent Manager. The default is password. Fill in the information and click Next.
168
Figure 4-123 IBM TotalStorage Productivity Center for Fabric Agent Manager information
17.The next panel (Figure 4-124) that opens provides information about the location and size of IBM TotalStorage Productivity Center for Fabric - Manager. Click Next.
Figure 4-124 IBM TotalStorage Productivity Center for Fabric installation information
18.You see the Status panel. The installation can take about 15 to 20 minutes to complete. 19.When the installation has completed, you see a panel indicating that the wizard successfully installed the Fabric Manager (see Figure 4-125 on page 170). Click Next.
169
Figure 4-125 IBM TotalStorage Productivity Center for Fabric installation status
20.In the next panel (see Figure 4-126), you are prompted to restart your computer. Select No, I will restart my computer later because you do not want to restart your computer now. Click Finish to complete the installation.
Figure 4-126 IBM TotalStorage Productivity Center for Fabric restart options
21.The Install Status panel (see Figure 4-127 on page 171) opens. It indicates that the Productivity Center for Fabric installation was successful. Click Next.
170
171
Figure 4-128 IBM TotalStorage Productivity Center for Data installation information
1. In the window that opens, you are prompted to enter the install package location for IBM TotalStorage Productivity Center for Data (see Figure 4-129). Enter the appropriate information and click Next.
2. The next window that opens indicates that the IBM TotalStorage Productivity Center for Data installer wizard will be launched (see Figure 4-130 on page 173). Click Next.
172
3. In the next panel (see Figure 4-131), select Install Productivity Center for Data and click Next.
Figure 4-131 IBM TotalStorage Productivity Center for Data install window
173
4. Read the License Agreement shown in Figure 4-132. Indicate your acceptance of the agreement by selecting the I have read and AGREE to abide by the license agreement above check box. Then click Next.
Figure 4-132 IBM TotalStorage Productivity Center for Data license agreement
5. The next panel asks you to confirm that you read the license agreement (see Figure 4-133). Click Yes to indicate that you have read and accepted the license agreement.
Figure 4-133 Confirmation the Productivity Center for Data license agreement has been read
174
6. The next window shown in Figure 4-134 allows you to choose the type of installation that you are performing. Select The Productivity Center for Data Server and an Agent on this machine. This installs the server, agent, and user interface components on the machine where the installation program is running. You must install the server on at least one machine within your environment. Click Next.
Figure 4-134 IBM TotalStorage Productivity Center for Data selection options
175
7. Review and enter the license key for the appropriate functions if required. See Figure 4-135. Click Next.
Figure 4-135 IBM TotalStorage Productivity Center for Data license key information
176
8. The installation program validates the license key and you are asked to select the relational database management system (RDBMS) that you want to host the Data Manager repository. See Figure 4-136. The repository is a set of relational database tables where Data Manager builds a database of statistics to keep track of your environment. For our installation, we select IBM DB2 UDB. Click Next.
Figure 4-136 IBM TotalStorage Productivity Center for Data database selection
9. The Create Service Account panel opens to create the TSRMsrv1 local account. Click Yes.
177
10.In the next window (see Figure 4-138), complete these tasks: a. Select the database that was created as a prerequisite. Refer to 3.12.4, Creating the DB2 database on page 81. b. Fill in the required user ID and password. This is the DB2 user ID and password defined previously. c. Click Next.
Figure 4-138 IBM TotalStorage Productivity Center for Data database selection option
11.The Repository Creation Parameters panel (see Figure 4-139 on page 179) for UDB opens. On this panel you can specify the database schema and tablespace name. If you are using DB2 as the repository, you can also choose how you will manage the database space: System Managed (SMS): This option indicates that the space is managed by the OS. In this case you specify the Container Directory, which is then managed by the system, and can grow as large as the free space on the file system. Tip: If you do not have in house database skills, the System Managed approach is recommended. Database Managed (DMS): This option means that the space is managed by the database. In this case you need to specify the Container Directory, Container File, and Size fields. The Container File specifies a filename for the repository, and Size is the predefined space for that file. You can later change this by using the ALTER TABLESPACE command. We accepted the defaults.
178
Tip: We recommend that you use meaningful names for Container Directory and Container File at installation. This can help you in case you need to find the Container File. Enter the necessary information and click Next.
Figure 4-139 IBM TotalStorage Productivity Center for Data repository information
179
12.The Productivity Center for Data Parameters panel (Figure 4-140) opens. Use the Agent Manager Parameters window (Figure 4-141 on page 182) to provide information about the Agent Manager installed in your environment. Click Next.
Figure 4-140 IBM TotalStorage Productivity Center for Data installation parameters
180
13.The Agent Manager Parameters panel (Figure 4-141 on page 182) provides information about the Agent Manager installed in your environment. Table 4-1 provides a description of the fields in the panel.
Table 4-1 Agent Manager Parameters descriptions Field Hostname Registration Port Public Port Resource Manager Username Description Enter the fully qualified network name or IP address of the Agent Manager server as seen by the agents. Enter the port number of the Agent Manager. The default is 9511. Enter the public port for Agent Manager. The default is 9513. Enter the Agent Manager user ID. This is the user ID used to register the common agent with the Agent Manager. The default is manager. Enter the password used to register the common agent with the Agent Manager. This is the password that was set during the Tivoli Agent Manager installation Figure 4-27 on page 105. The default password is changeMe, the password is stored in the AgentManager.properties file, in the %install dir%\AgentManager\image directory.
Click Next. Note: If an error is displayed during this part of the installation, verify that the Agenttrust.jks file was copied across and verify the Agent Registration password.
181
Figure 4-141 IBM TotalStorage Productivity Center for Data Agent Manager install information
182
14.Use the NAS Discovery Parameters panel in Figure 4-142 to configure Data Manager for use with any network-attached storage (NAS) devices in your environment. Click Next. You can leave the fields blank if you do not have any NAS devices.
Figure 4-142 IBM TotalStorage Productivity Center for Data NAS options
183
15.The Space Requirements panel for the Productivity Center for Data Server (Figure 4-143) opens. Enter the directory path or accept the default directory. If the current disk or device does not have enough space for the installation, then you can enter a different location for the installation in the Choose the installation directory field. Or you can click Browse to browse your system for an available and appropriate space. The default installation directory is C:\Program Files\IBM\TPC\Data. Click Next.
Figure 4-143 IBM TotalStorage Productivity Center for Data installation destination options
16.Confirm the path for installing the Productivity Center for Data Server as shown in Figure 4-144. At this point, the installation process has gathered all of the information that is needed to perform the installation. Click OK.
Figure 4-144 IBM TotalStorage Productivity Center for Data Server destination path confirmation
184
17.Review and change the Productivity Center for Data Agent Parameters (see Figure 4-145) as required. We recommend that you accept the defaults. Click Next.
Figure 4-145 IBM TotalStorage Productivity Center for Data agent parameters
185
18.The Windows Service Account panel shown in Figure 4-146 opens. Choose Create a local account for the agent to run under and click Next.
186
19.The Space Requirements panel (see Figure 4-147) opens for the Productivity Center for Data Agent. Enter the directory path or accept the default directory. If the current disk or device does not have enough space for the installation, then you can enter a different location for the installation in the Choose the Common Agent installation directory field. Or you can click Browse to browse your system for an available and appropriate space. The default installation directory is C:\Program Files\Tivoli\ep. Click Next.
Figure 4-147 IBM TotalStorage Productivity Center for Datacommon agent installation information
20.When you see a message similar to the one in Figure 4-148, confirm the path where Productivity Center for Data Agent is to be installed. At this point, the installation process has gathered all of the information necessary to perform the installation. Click OK.
Figure 4-148 IBM TotalStorage Productivity Center for Data Agent destination path confirmation
187
21.When you see a window similar to the example in Figure 4-149, review the choices that you have made. Then click Next.
Figure 4-149 IBM TotalStorage Productivity Center for Data preview options
188
22.A window opens that tracks the progress of the installation (see Figure 4-150).
Figure 4-150 IBM TotalStorage Productivity Center for Data installation information
189
23.When the installation is done, the progress window shows a message indicating that the installation completed successfully (see Figure 4-151). Review this panel and click Done.
Figure 4-151 IBM TotalStorage Productivity Center for Data success information
24.The Install Status panel opens showing the message The Productivity Center for Data installation was successful. Click Next to complete the installation.
Figure 4-152 Install Status for Productivity Center for Data successful
190
Chapter 5.
191
5.1 Introduction
After you have completed the installation of TotalStorage Productivity Center for Disk, TotalStorage Productivity Center for Replication, TotalStorage Productivity Center for Fabric, or TotalStorage Productivity Center for Data, you will need to install and configure the Common Information Model Object Manager (CIMOM) and Service Location Protocol (SLP) agents. Note: For the remainder of this chapter, we refer to the TotalStorage Productivity Center for Disk, TotalStorage Productivity Center for Replication, TotalStorage Productivity Center for Fabric, and TotalStorage Productivity Center for Data simply as the TotalStorage Productivity Center. The TotalStorage Productivity Center uses SLP as the method for CIM clients to locate managed objects. The CIM clients may have built in or external CIM agents. When a CIM agent implementation is available for a supported device, the device may be accessed and configured by management applications using industry-standard XML-over-HTTP transactions. In this chapter we describe the steps for: Planning considerations for Service Location Protocol (SLP) SLP configuration recommendation General performance guidelines Planning considerations for CIMOM Installing and configuring CIM agent for Enterprise Storage Server and DS6000/DS8000 Verifying connection to ESS Verify connection to DS6000/DS8000 Setting up Service Location Protocol Directory Agent (SLP DA) Installing and configuring CIM agent for DS 4000 Family Configuring CIM agent for SAN Volume Controller
192
By deploying one or more DAs, UAs must unicast to DAs for service and SAs must register with DAs using unicast. The only SLP-registered multicast in a network with DAs is for active and passive DA discovery. SAs register automatically with any DAs they discover within a set of common scopes. Consequently, DAs within the UAs scopes reduce multicast. By eliminating multicast for normal UA request, delays and time-outs are eliminated. DAs act as a focal point for SA and UA activity. Deploying one or several DAs for a collection of scopes provides a centralized point for monitoring SLP activity. You may consider to use DAs in your enterprise if any of the following conditions are true: Multicast SLP traffic exceeds 1% of the bandwidth on your network, as measured by snoop. UA clients experience long delays or time-outs during multicast service request. You would like to centralize monitoring of SLP service advertisements for particular scopes on one or several hosts. You can deploy any number of DAs for a particular scope or scopes, depending on the need to balance the load. Your network does not have multicast enabled and consists of multiple subnets that must share services. The configuration of an SLP DA is particularly recommended when there are more than 60 SAs that need to respond to any given multicast service request.
Router configuration
Configure the routers in the network to enable general multicasting or to allow multicasting for the SLP multicast address and port, 239.255.255.253, port 427. The routers of interest are those that are associated with subnets that contain one or more storage devices that are to be discovered and managed by TotalStorage Productivity Center. To configure your router hardware and software, refer to your router reference and configuration documentation. Attention: Routers are sometimes configured to prevent passing of multicast packets between subnets. Routers configured this way prevent discovery of systems between subnets using multicasting. Routers can also be configured to restrict the minimum multicast TTL (time-to-live) for packets it passes between subnets, which can result in the need to set the Multicast TTL higher to discover systems on the other subnets of the router. The Multicast TTL controls the time-to-live for the multicast discovery packets. This value typically corresponds to the number of times a packet is forwarded between subnets, allowing control of the scope of subnets discovered. Multicast discovery does not discover Director V1.x systems or systems using TCP/IP protocol stacks that do not support multicasting (for example, some older Windows 3.x and Novell 3.x TCP/IP implementations).
193
194
You may plan to install CIM agent code on the same server which also has device management interface or you may install it on a separate server. Attention: At this time only few devices come with an integrated CIM Agent, most devices need a external CIMOM for CIM enable management applications (CIM Clients) to be able to communicate with device. For the ease of the installation IBM provides an ICAT (short for Integrated Configuration Agent Technology) which is a bundle that mainly includes the CIMOM, the device provider and an SLP SA.
You must have the CIMOM supported firmware level on the storage devices. It you have an incorrect version of firmware, you may not be able to discover and manage any the storage devices. The data traffic between CIMOM agent and device can be very high, especially during performance data collection. Hence it is recommended to have a dedicated server for the CIMOM agent. Although, you may configure the same CIMOM agent for multiple devices of same type. You may also plan to locate this server within same data center where storage devices are located. This is in consideration to firewall port requirements. Typically, it is best practice to minimize firewall port openings between data center and external network. If you consolidate the CIMOM servers within the data center then you may be able to minimize the need to open the firewall ports only for TotalStorage Productivity Center communication with CIMOM. Co-location of CIM agent instances of the differing type on the same server is not recommended because of resource contention. It is strongly recommended to have a separate and dedicated servers for CIMOM agents and TotalStorage Productivity Center. This is due to resource contention, TCP/IP port requirements and system services co-existence.
195
This section provides an overview of the installation and configuration of the ESS CIM Agent on a Windows 2000 Advanced Server operating system.
Perform the following steps to install the ESS CLI for Windows: 1. Insert the CD for the ESS CLI in the CD-ROM drive, run the setup and follow the instructions as shown in Figure 5-3 on page 197 through Figure 5-11 on page 201. Note: The ESS CLI installation wizard detects if you have an earlier level of the ESS CLI software installed on your system and uninstalls the earlier level. After you uninstall the previous version, you must restart the ESS CLI installation program to install the current level of the ESS CLI.
196
2. Select I accept the terms of the license agreement and click Next.
3. Click Next.
197
4. Click Next.
5. Click Next.
198
6. Click Next.
199
8. Reboot your system before proceeding with the ESS CIM Agent installation. You must do this because the ESS CLI is dependent on environmental variable settings which will not be in effect for the ESS CIM Agent. This is because the CIM Agent runs as a service unless you reboot your system.
200
9. Verify that the ESS CLI is installed: Click Start Settings Control Panel. Double-click the Add/Remove Programs icon. Verify that there is an IBM ESS CLI entry. 10.Verify that the ESS CLI is operational and can connect to the ESS. For example, from a command prompt window, issue the following command: esscli -u userid -p password -s 9.1.11.111 list server Where: 9.1.11.111 represents the IP address of the Enterprise Storage Server usedid represents the Enterprise Storage Server Specialist user name password represents the Enterprise Storage Server Specialist password for the user name Figure 5-12 shows the response from the esscli command.
201
202
4. The Welcome window opens suggesting what documentation you should review prior to installation. Click Next to continue (see Figure 5-14).
203
5. The License Agreement window opens. Read the license agreement information. Select I accept the terms of the license agreement, then click Next to accept the license agreement (see Figure 5-15).
204
The window shown in Figure 5-16 only appears if no valid ESS CLI installed. If you do not plan to manage an ESS from this CIM agent, then click Next. Important: If you plan to manage an ESS from this CIM agent, then click Cancel. Install the ESS CLI following the instructions in 5.5.1, ESS CLI Install on page 196.
205
6. The Destination Directory window opens. Accept the default directory and click Next (see Figure 5-17).
206
7. The Updating CIMOM Port window opens (see Figure 5-18). You Click Next to accept the default port if it available and free in your environment. For our ITSO setup we used default port 5989. Note: If the default port is the same as another port already in use, modify the default port and click Next. Use the following command to check which ports are in use: netstat -a
207
8. The Installation Confirmation window opens (see Figure 5-19). Click Install to confirm the installation location and file size.
208
9. The Installation Progress window opens (see Figure 5-20) indicating how much of the installation has completed.
10.When the Installation Progress window closes, the Finish window opens (see Figure 5-21 on page 210). Check the View post installation tasks check box if you want to view the post installation tasks readme when the wizard closes. We recommend you review the post installation tasks. Click Finish to exit the installation wizard (Figure 5-21 on page 210). Note: Before proceeding, you might want to review the log file for any error messages. The log file is located in xxx\logs\install.log, where xxx is the destination directory where the DS CIM Agent for Windows is installed.
209
11.If you checked the view post installation tasks box, then the window shown in Figure 5-22 appears. Close the window when you have finished reviewing the post installation tasks.
The launchpad window (Figure 5-13 on page 202) appears. Click Exit.
210
If SLP is not started, right-click on the SLP and select Start from the pop-up menu. Wait for the Status column to be changed to Started.
211
If the IBM CIM Object Manager is not started, right-click on the IBM CIM Object Manager ESS and select Start from the pop-up menu. Wait for the Status column to change to Started. If you are able to perform all of the verification tasks successfully, the ESS CIM Agent has been successfully installed on your Windows system. Next, perform the configuration tasks.
Repeat the previous step for each additional DS device that you want to configure. Note: CIMOM collects and caches the information from the defined DS servers at startup time; the starting of the CIMOM might take a longer period of time the next time you start it.
Attention: If the username and password entered is incorrect or the DS CIM agent does not connect to the DS this will cause a error and the DS CIM Agent will not start and stop correctly, use following command to remove the ESS entry that is causing the problem and reboot the server. rmessserver <ip> Whenever you add or remove DS from CIMOM registration, you must re-start the CIMOM to pick up the updated DS device list.
213
Attention: If the username and password entered is incorrect or the ESS CIM agent does not connect to the ESS this will cause a error and the ESS CIM Agent will not start and stop correctly, use following command to remove the ESS entry that is causing the problem and reboot the server. rmessserver <ip> Whenever you add or remove an ESS from CIMOM registration, you must re-start the CIMOM to pick up updated ESS device list.
214
Restart the CIMOM by selecting Start Programs CIM Agent for the IBM TotalStorage DS Open API Start CIMOM service. A Command Prompt window opens to track the progress of the starting of the CIMOM. If the CIMOM has started successfully, the message shown in Figure 5-28 is displayed.
Note: The restarting of the CIMOM may take a while because it is connecting to the defined ESS servers and is caching that information for future use.
215
Note: The users which you configure to have authority to use the CIMOM are uniquely defined to the CIMOM software and have no required relationship to operating system user names, the ESS Specialist user names, or the ESS Copy Services user names. Here is the procedure: Open a Command Prompt window and change directory to the ESS CIM Agent directory, for example C:\Program Files\IBM\cimagent. Type the command setuser -u superuser -p passw0rd at the command prompt to start the setuser interactive session to identify users to the CIMOM. Type the command adduser cimuser cimpass in the setuser interactive session to define new users. where cimuser represents the new user name to access the ESS CIM Agent CIMOM
cimpass represents the password for the new user name to access the ESS CIM
Agent CIMOM
Close the setdevice interactive session by typing exit. For our ITSO Lab setup we used TPCCIMOM as superuser and TPCCIMOM as the password.
216
Verify that SLP has dependency on CIMOM, this is automatically configured when you installed the CIM agent software. Verify this by selecting Start Settings Control Panel. Double-click the Administrative Tools icon. Double-Click the Services icon and subsequently select properties on Service Location Protocol as shown in Figure Figure 5-30.
Click Properties and select the Dependencies tab as shown in Figure 5-31 on page 218. You must ensure that IBM CIM Object Manager has a dependency on Service Location Protocol (this should be the case by default).
217
Verify CIMOM registration with SLP by selecting Start Programs CIM Agent for the IBM TotalStorage DS Open API Check CIMOM Registration. A window opens displaying the WBEM services as shown in Figure 5-32. These services have either registered themselves with SLP or you explicitly registered them with SLP using slptool. If you changed the default ports for a CIMOM during installation, the port number should be correctly listed here. It may take some time for a CIM Agent to register with SLP.
Note: If the verification of the CIMOM registration is not successful, stop and restart the SLP and CIMOM services. Note that the ESS CIMOM will attempt to contact each ESS registered to it. Therefore, the startup may take some time, especially if it is not able to connect and authenticate to any of the registered ESSs. Use the verifyconfig -u superuser -p passw0rd command, where superuser is the user name and passw0rd is the password for the user name that you configured to manage the CIMOM, to locate all WBEM services in the local network. You need to define the TotalStorage Productivity Center for Disk superuser name and passw0rd in order for TotalStorage Productivity Center for Disk to have the authority to manage the CIMOM. The verifyconfig command checks the registration for the ESS CIM Agent and checks that it can
218
connect to the ESSs. At ITSO Lab we had configured two ESSs (as shown in Figure 5-33 on page 219).
219
If you still have problems, Refer to the DS Open Application Programming Interface Reference for an explanation and resolution of the error messages. You can find this Guide in the doc directory at the root of the CIM Agent CD.
Another method to verify that DS CIMOM is up and running is to use the CIM Browser interface. For Windows machines change the working directory to c:\Program Files\ibm\cimagent and run startcimbrowser. The WBEM browser in Figure 5-36 will appear. The default user name is superuser and the default password is passw0rd. If you have already changed it, using the setuser command, the new userid and password must be provided. This should be set to the TotalStorage Productivity Center for Disk userid and password.
220
When login is successful, you should see a panel like the one in Figure 5-37.
221
Open the slp.conf file and scroll down until you find (or search for) the line ;net.slp.isDA = true Remove the semi-colon (;) at the beginning of the line. Ensure that this property is set to true (= true) rather than false. Save the file. Copy this file (or replace it if the file already exists) to the main windows subdirectory for Windows machines (for example c:\winnt), or in the /etc directory for UNIX machines. 4. It is recommended to reboot the SLP server at this stage. Otherwise, alternatively, you may choose to restart the SLP and CIMOM services. You can do this from your Windows desktop Start Menu Settings Control Panel Administrative tools Services. Launch the Services GUI Locate the Service Location Protocol, right click and select stop. It will pop-up another panel which will request to stop IBM CIM Object Manager service. You may click Yes. You may start the SLP daemon again after it has stopped successfully. Alternatively, you may choose to re-start the CIMOM using command line as described in Restart the CIMOM on page 215
service:wbem:https://9.43.226.237:5989,en,65535 # use default scopes: scopes=test1,test2 description=SVC CIMOM Open Systems Lab, Cottle Road authors=Aliprandi,Andrews,Cooper,Eggli,Lovelace,Zerbini creation_date=04/02/20 service:wbem:https://9.11.209.188:5989,en,65535 # use default scopes: scopes=test1,test2
222
description=SVC CIMOM Tucson L2 Lab authors=Aliprandi,Andrews,Cooper,Eggli,Lovelace,Zerbini creation_date=04/02/20 #service:wbem:https://9.42.164.175:5989,en,65535 # use default scopes: scopes=test1,test2 #description=SVC CIMOM Raleigh SAN Central #authors=Aliprandi,Andrews,Cooper,Eggli,Lovelace,Zerbini #creation_date=04/02/20 #---------------------------------------------------------------------------# Register Service - SANFS CIMOMS #---------------------------------------------------------------------------#service:wbem:https://9.82.24.66:5989,en,65535 #Additional parameters for setting the appropriate namespace values #CIM_InteropSchemaNamespace=root/cimv2 #Namespace=root/cimv2 # use default scopes: scopes=test1,test2 #description=SANFS CIMOM Gaithersburg ATS Lab #authors=Aliprandi,Andrews,Cooper,Eggli,Lovelace,Zerbini #creation_date=04/02/20 #service:wbem:https://9.11.209.148:5989,en,65535 #Additional parameters for setting the appropriate namespace values #CIM_InteropSchemaNamespace=root/cimv2 #Namespace=root/cimv2 # use default scopes: scopes=test1,test2 #description=SANFS CIMOM Tucson L2 Lab #authors=Aliprandi,Andrews,Cooper,Eggli,Lovelace,Zerbini #creation_date=04/02/20 #---------------------------------------------------------------------------# Register Service - FAStT CIMOM #---------------------------------------------------------------------------#service:wbem:https://9.1.39.65:5989,en,65535 #CIM_InteropSchemaNamespace=root/lsissi #ProtocolVersion=0 #Namespace=root/lsissi # use default scopes: scopes=test1,test2 #description=FAStT700 CIMOM ITSO Lab, Almaden #authors=Aliprandi,Andrews,Cooper,Eggli,Lovelace,Zerbini #creation_date=04/02/20
223
You have now converted the SLP SA of the CIM Agent to run as an SLP DA. The CIMOM is not affected and will register itself with the DA instead of the SA. However, the DA will automatically discover all other services registered with other SLP SAs in that subnet. Attention: You will need to register the IP address of the server running the SLP DA daemon with the IBM Director to facilitate MDM SLP discovery. You can do this using the IBM Director console interface of TotalStorage Productivity Center for Disk. The procedure to register the IP address is described in 6.2, SLP DA definition on page 248.
224
You may to have traverse through multiple links to get to the download files. At the time of writing this book, we go to the Web page shown in Figure 5-38.
225
By scrolling down the same Web page, we go to the following link for DS 4000 CIMOM code as shown in Figure 5-39. This link leads to the Engenio Provider site. The current supported code level is 1.0.59, as indicated in the Web page.
From the Web site, select the operating system used for the server on which the IBM DS family CIM Agent will be installed. You will download a setup.exe file. Save it to a directory on the server on which you will be installing the DS 4000 CIM Agent (see Figure 5-40 on page 227).
226
Launch the setup.exe file to begin the DS 4000 family CIM agent installation. The InstallShield Wizard for LSI SMI-S Provider window opens (see Figure 5-41). Click Next to continue.
227
The LSI License Agreement window opens next. If you agree with the terms of the license agreement, click Yes to accept the terms and continue the installation (see Figure 5-42).
The LSI System Info window opens. The minimum requirements are listed along with the install system disk free space and memory attributes as shown in Figure 5-43. If the install system fails the minimum requirements evaluation, then a notification window will appear and the installation will fail. Click Next to continue.
228
The Choose Destination Location window appears. Click Browse to choose another location or click Next to begin the installation of the FAStT CIM agent (see Figure 5-44).
The InstallShield Wizard will now prepare and copy the files into the destination directory. See Figure 5-45.
229
The README appears after the files have been installed. Read through it to become familiar with the most current information (see Figure 5-46). Click Next when ready to continue.
In the Enter IPs and/or Hostnames window, enter the IP addresses and hostnames of the FAStT devices that this FAStT CIM agent will manage as shown in Figure 5-47.
230
Use the Add New Entry button to add the IP addresses or hostnames of the FAStT devices that this FAStT CIM agent will communicate with. Enter one IP address or hostname at a time until all the FAStT devices have been entered and click Next (see Figure 5-48).
Do not enter the IP address of a FAStT device in multiple FAStT CIM Agents within the same subnet. This may cause unpredictable results on the TotalStorage Productivity Center for Disk server and could cause a loss of communication with the FAStT devices. If the list of hostnames or IP addresses has been previously written to a file, use the Add File Contents button, which will open the Windows Explorer. Locate and select the file and then click Open to import the file contents. When all the FAStT device hostnames and IP addresses have been entered, click Next to start the SMI-S Provider Service (see Figure 5-49).
231
When the Service has started, the installation of the FAStT CIM agent is complete (see Figure 5-50).
Arrayhosts file
The installer will create a file called %installroot%\SMI-SProvider\wbemservices\cimom\bin\arrayhosts.txt The arrayhosts file is shown in Figure 5-51. In this file the IP addresses of installed DS 4000 units can be reviewed, added, or edited.
232
233
Figure 5-53 TotalStorage Productivity Center for Disk and SVC communication
For additional details on how to configure the SAN Volume Controller Console, refer to the redbook IBM TotalStorage Introducing the SAN Volume Controller and SAN Integration Server, SG24-6423. To discover and manage the SAN Volume Controller, we need to ensure that our TotalStorage Productivity Center for Disk superuser name and password (the account we specify in the TotalStorage Productivity Center for Disk configuration panel, as shown in 5.9.1, Adding the SVC TotalStorage Productivity Center for Disk user account on page 235) matches an account defined on the SAN Volume Controller console. In our case we implemented username TPCSUID and password ITSOSJ. You may want to adapt a similar nomenclature and set up the username and password on each SAN Volume Controller CIMOM to be monitored with TotalStorage Productivity Center for Disk.
234
5.9.1 Adding the SVC TotalStorage Productivity Center for Disk user account
As stated previously, you should implement a unique userid to manage the SAN Volume Controller devices in TotalStorage Productivity Center for Disk. This can be achieved at the SAN Volume Controller console using the following steps: 1. Login to the SAN Volume Controller console with a superuser account 2. Click Users under My Work on the left side of the panel (see Figure 5-54).
235
3. Select Add a user in the drop down under Users panel and click Go (see Figure 5-55).
236
237
5. Enter the User Name and Password and click Next (see Figure 5-57).
6. Select your candidate cluster and move it to the right under Administrator Clusters (see Figure 5-58). Click Next to continue.
238
7. Click Next after you Assign service roles (see Figure 5-59).
239
8. Click Finish after you verify user roles (see Figure 5-60).
9. After you click Finish, the Viewing users panel opens (see Figure 5-61).
240
5.10 Configuring CIMOM for TotalStorage Productivity Center for Disk summary
The TotalStorage Productivity Center discovers both IBM storage devices that comply with the Storage Management Initiative Specification (SMI-S) and SAN devices such as switches, ports, and hosts. SMIS-compliant storage devices are discovered using the Service Location Protocol (SLP).
241
The TotalStorage Productivity Center server software performs SLP discovery on the network. The User Agent looks for all registered services with a service type of service:wbem. The TotalStorage Productivity Center performs the following discovery tasks: Locates individual storage devices Retrieves vital characteristics for those storage devices Populates The TotalStorage Productivity Center internal databases with the discovered information The TotalStorage Productivity Center can also access storage devices through the CIM Agent software. Each CIM Agent can control one or more storage devices. After the CIMOM services have been discovered through SLP, the TotalStorage Productivity Center contacts each of the CIMOMs directly to retrieve the list of storage devices controlled by each CIMOM. TotalStorage Productivity Center gathers the vital characteristics of each of these devices. For the TotalStorage Productivity Center to successfully communicate with the CIMOMs, the following conditions must be met: A common user name and password must be configured for all the CIM Agent instances that are associated with storage devices that are discoverable by TotalStorage Productivity Center (use adduser as described in 5.6.5, CIMOM user authentication on page 215). That same user name and password must also be configured for TotalStorage Productivity Center using the Configure MDM task in the TotalStorage Productivity Center interface. If a CIMOM is not configured with the matching user name and password, it will be impossible to determine which devices the CIMOM supports. As a result, no devices for that CIMOM will appear in the IBM Director Group Content pane. The CIMOM service must be accessible through the IP network. The TCP/IP network configuration on the host where TotalStorage Productivity Center is installed must include in its list of domain names all the domains that contain storage devices that are discoverable by the TotalStorage Productivity Center. It is important to verify that CIMOM is up and running. To do that, use the following command from TotalStorage Productivity Center server: telnet CIMip port Where: CIMip is the ip address where CIM Agent run and port is the port value used for the communication (5989 for secure connection, 5988 for unsecure connection).
242
#---------------------------------------------------------------------------# Register Service - SVC CIMOMS #---------------------------------------------------------------------------service:wbem:https://9.43.226.237:5989,en,65535 # use default scopes: scopes=test1,test2 description=SVC CIMOM Open Systems Lab, Cottle Road authors=Aliprandi,Andrews,Cooper,Eggli,Lovelace,Zerbini creation_date=04/02/20 service:wbem:https://9.11.209.188:5989,en,65535 # use default scopes: scopes=test1,test2 description=SVC CIMOM Tucson L2 Lab authors=Aliprandi,Andrews,Cooper,Eggli,Lovelace,Zerbini creation_date=04/02/20 #service:wbem:https://9.42.164.175:5989,en,65535 # use default scopes: scopes=test1,test2 #description=SVC CIMOM Raleigh SAN Central #authors=Aliprandi,Andrews,Cooper,Eggli,Lovelace,Zerbini #creation_date=04/02/20 Chapter 5. CIMOM install and configuration
243
#---------------------------------------------------------------------------# Register Service - SANFS CIMOMS #---------------------------------------------------------------------------#service:wbem:https://9.82.24.66:5989,en,65535 #Additional parameters for setting the appropriate namespace values #CIM_InteropSchemaNamespace=root/cimv2 #Namespace=root/cimv2 # use default scopes: scopes=test1,test2 #description=SANFS CIMOM Gaithersburg ATS Lab #authors=Aliprandi,Andrews,Cooper,Eggli,Lovelace,Zerbini #creation_date=04/02/20 #service:wbem:https://9.11.209.148:5989,en,65535 #Additional parameters for setting the appropriate namespace values #CIM_InteropSchemaNamespace=root/cimv2 #Namespace=root/cimv2 # use default scopes: scopes=test1,test2 #description=SANFS CIMOM Tucson L2 Lab #authors=Aliprandi,Andrews,Cooper,Eggli,Lovelace,Zerbini #creation_date=04/02/20 #---------------------------------------------------------------------------# Register Service - FAStT CIMOM #---------------------------------------------------------------------------#service:wbem:https://9.1.39.65:5989,en,65535 #CIM_InteropSchemaNamespace=root/lsissi #ProtocolVersion=0 #Namespace=root/lsissi # use default scopes: scopes=test1,test2 #description=FAStT700 CIMOM ITSO Lab, Almaden #authors=Aliprandi,Andrews,Cooper,Eggli,Lovelace,Zerbini #creation_date=04/02/20
244
Part 3
Part
245
246
Chapter 6.
247
Here, CIMip is the IP address where the CIM Agent runs, and port is the port value used for the communication (5989 for a secure connection; 5988 for an unsecure connection).
248
We are assuming that you have followed the steps outlined in Chapter 5, CIMOM install and configuration on page 191. You should complete the following tasks in order to discover devices defined to our Productivity Center common base host. Make sure that: All CIM agents are running and are registered with the SLP server. The SLP agent host is defined in the IBM Director options (Figure 6-1) if it resides in a different subnet from that of the TotalStorage Productivity Center server (Options Discovery Preferences MDM SLP Configuration tab). Note: If the Productivity Center common base host server resides in the same subnet as the CIMOM, then it is not a requirement that the SLP DA host IP address be specified in the Discovery Preferences panel as shown in Figure 6-2. Refer to Chapter 2, Key concepts on page 27 for details on SLP discovery. Here we provide a step-by-step procedure: 1. Discovery will happen automatically based on preferences that are defined in the Options Discovery Preferences MDM SLP Configuration tab. The default values for Auto discovery interval and Presence check interval is set to 0 (see Figure 6-1). These values should be set to a more suitable value, for example, to 1 hour for Auto discovery interval and 15 minutes for Presence check interval. The values you specify will have a performance impact on the CIMOMs and Productivity Center common base server, so do not set these values too low.
249
Continue entering IP addresses for all SLP DA servers. Click OK when finished (see Figure 6-2).
2. Turn off automatic inventory on discovery. Important: Because of the time and CIMOM resources needed to perform inventory on storage devices, it is undesirable and unnecessary to perform this operation each time Productivity Center common base performs a device discovery.
250
Turn off automatic inventory by selecting Options Server Preferences as shown in Figure 6-3.
Now uncheck the Collect On Discovery check box as shown in Figure 6-4, all other options can remain unchanged. Select OK when done.
251
3. You can click Discover all Systems in the top left corner of the IBM Director Console to initiate an immediate discovery task (see Figure 6-5).
252
4. You can also use the IBM Director Scheduler to create a scheduled job for new device discovery. Either click the scheduler icon in the IBM Director tool bar or use the menu, Tasks Scheduler (see Figure 6-6).
253
Establish parameters for the new job. Under the Date/Time tab, include date and time to perform the job, and whether the job is to be repeated (see Figure 6-8).
From the Task tab (see Figure 6-9), select Discover MDM storage devices/SAN Elements, then click Select.
254
Click File Save as, or use the Save as icon. Provide a descriptive job name in the Save Job panel (see Figure 6-10) and click OK.
Now run the discovery process by selecting Tasks Discover Systems All Systems and Devices (Figure 6-11).
255
Double-click the Manage CIMOM task to see the status of the discovery (Figure 6-12).
256
It should not be necessary to change any information if you followed the recommendation to use the same superuser id and password for all CIMOMs. Select the CIMOM to be configured and click Properties to configure a CIMOM (Figure 6-14).
1. In order to verify and re-confirm the connection, you may select the respective connection status and click Properties. Figure 6-16 on page 258 shows the properties panel. You may verify the information and update if necessary. The namespace, username and password are picked up automatically, hence they are not normally required to be entered manually. This username is used by CIMOM to logon to TotalStorage Productivity Center for Disk. If you have difficulty getting a successful connection, then you may manually enter namespace, username, and password. Update the properties panel and test the connection to the CIMOM: a. Enter the Namespace value. It is \root\ibm for the ESS, DS6000 and DS8000 It is \interop for the DS4000. b. Select the protocol. It is typically https for ESS, DS6000 and DS8000. It is http for DS4000. c. Enter the User name and password. The default is the superuser password entered earlier. If you entered a different user name and password with the setuser command for the CIM agent, then enter that user name and password here. d. Click Test Connection to verify correct configuration. e. You should see the panel in Figure 6-15. Click Close on the panel.
257
2. After the connection to the CIMOM is successful, then perform discovery again as shown before in Figure 6-11 on page 255. This will discover the storage devices connected to each CIMOM (Figure 6-17).
3. Click the Test Connection button to see a panel similar to Figure 6-15 on page 257, showing that the connection is successful. Tip: If you move or delete CIMOMs in your environment, the old CIMOM entries are not automatically updated, and entries with a Failure status will be seen as in Figure 6-13 on page 256. These invalid entries can slow down discovery performance, as TotalStorage Productivity Center tries to contact them each time it performs a discovery. You cannot delete CIMOM entries directly from the Productivity Center common base interface. Delete them using the DB2 control center tool as described in 16.6, Manually removing old CIMOM entries on page 911.
258
259
3. In the IBM Director Installation panel (Figure 6-19), select IBM Director Console installation.
4. After a moment, the InstallShield Wizard for IBM Director Console panel (Figure 6-20) opens. Click Next.
260
5. In the License Agreement panel (Figure 6-21), select I accept the terms in the license agreement. Then click Next.
6. The next panel (Figure 6-22) contains information about enhancing IBM Director. Click Next to continue.
261
7. The Feature and installation directory selection panel (Figure 6-23) allows you to change how a program feature is installed. Click Next.
8. In the Ready to Install the Program window (Figure 6-24), accept the default selection. Then click Install to start the installation.
262
9. The installation takes a few minutes. When it is finished, you see the InstallShield Wizard Completed window (Figure 6-25). Click Finish to complete the installation.
263
3. Start the LaunchPad.bat batch file. Coincidently this file has the same name as the TotalStorage Productivity Center Launchpad, although it has nothing to do with it. 4. Click Installation wizard to begin the installation (Figure 6-27 on page 265).
264
5. For a brief moment, you see a DOS box with the installer being unpacked. When this is done, you see the Welcome window shown in Figure 6-28. Click Next.
265
6. The License Agreement window (Figure 6-29) is displayed. Select I accept the term in the license agreement and click Next.
266
7. The Destination Directory window (Figure 6-30) opens. Accept the default path or enter the target directory for the installation. Click Next.
267
8. In the Select Product Type window (Figure 6-31), select Productivity Center for Disk and Replication Base Console for the product type. Click Next.
268
9. The Preview window (Figure 6-32) contains the installation information. Review it and click Install to start the console install.
10.When you reach the Finish window, click Finish to exit the add-on installer (Figure 6-33).
Figure 6-33 Installation finished Chapter 6. Configuring IBM TotalStorage Productivity Center for Disk
269
11.You return to the IBM TotalStorage Productivity Center for Disk and Replication Base installer window shown in Figure 6-27 on page 265. Click Exit to end the installation. The IBM Director remote console is now installed. The add-ons for IBM TotalStorage Productivity Center for Disk and Replication Base have been added. If the TotalStorage Productivity Center Launchpad is installed, it detects that the IBM Director remote console is available the next time the LaunchPad is started. Also the Launchpad can now be used to start IBM Director.
270
Next, you will see Welcome panel shown in Figure 6-35.Click Next.
Figure 6-35 Welcome panel from TotalStorage Productivity Center for Disk installer
The License Agreement panel shown in Figure 6-36 on page 272 appears. Select I accept the terms in the license agreement and click Next to continue.
271
Choose the default destination directory as shown in Figure 6-37 and click Next.
272
In the next panel, choose to install Productivity Center for Disk Client and click Next as shown in Figure 6-38.
273
In the next panel, select both check boxes for products, if you would like to install the console and command line client for the Performance Manager function (see Figure 6-39). Click Next.
274
The Productivity Center for Disk Installer - CoServer Parameters panel opens (see Figure 6-40). Enter the TPC user ID and password and the IP that the remote console will use to validate with the TPC server. This is the IP of the TPC server and IBM Director logon.
The Productivity Center for Disk Installer - Preview panel appears (see Figure 6-41 on page 276). Review the information and click Install to start the process of installing the remote console.
275
When the install is complete you will see the Productivity Center for Disk Installer - Finish panel as shown in Figure 6-42. Click Finish to complete the install process.
276
You may click Manage Disk Performance and Replication as highlighted in the figure. This will launch IBM director remote console. You may logon to director server and start using remote console functions except for Replication Manager. Note: At this point, you have installed the remote console for Performance Manager function only and not for replication manager. You can install remote console for replication manager if you wish.
277
278
Chapter 7.
279
280
5. The software license agreement panel appears (Figure 7-3). Click the radio button next to I accept the terms of the license agreement and click Next to continue.
6. In the TotalStorage Productivity Center install options panel (Figure 7-4), click the radio button next to User interface installations of Data, Disk, Fabric, and Replication and click Next.
281
7. In the Remote GUI/Command LIne Client component window (Figure 7-5), check the box by The Productivity Center for Replication - Command Line Client and click Next.
8. A window opens (Figure 7-6) to begin the replication command line client install.
282
9. In the next window, enter the location of the Replication Manager install package (Figure 7-7).
10.A window opens prompting you to interact with the Replication Manager install wizard (Figure 7-8).
283
11.The window in Figure 7-9 appears until the install wizard is launched.
12.The Productivity Center for Replication Installer - Welcome wizard window (Figure 7-10) opens. Click Next.
284
13.Specify the directory path of the Replication Manager installation files in the window shown in Figure 7-11. Click Next.
14.In the CoServer Parameters window shown in Figure 7-12, enter the following information: Host Name: Host name or IP address of the Replication Manager server Host Port: Port number of the Replication Manager server (default value is 9443) User Name: User name of the CIM Agent managing the storage device(s) User Password: User password of the CIM Agent managing the storage device(s)
285
15.Review the information in the Preview window shown in Figure 7-13 and click Install.
286
16.After successfully installing the remote CLI, the window in Figure 7-14 appears. Click Finish.
17.After clicking Finish, the postinstall.txt file opens.You may read the file or close and view it at a later time. 18.A window opens informing you of a successful installation (see Figure 7-15). Click Next to finish.
287
288
Chapter 8.
289
290
If you want to configure a CIM Agent that cannot be discovered because of the restriction explained in 8.1.1, CIM and SLP interfaces within Data Manager on page 290, then you also need to enter the IP address and select the right protocol.
If you completed the worksheets (see Appendix A, Worksheets on page 991), have them available for the next steps.
291
292
Next steps
After you configure the CIM Agent properties, run discovery on the storage subsystems. During this process, the Data Manager talks to the CIM Agent to gather information about the devices. When this is completed, you see an entry for the subsystem in the Storage Subsystem Administration panel (Figure 8-5 on page 294).
293
2. The panel shown in Figure 8-5 on page 294 opens. a. Highlight the subsystem. b. Place a check mark in the Monitored column. Note: Select the Monitored column if you want Data Manager to probe the subsystem whenever a probe job is run against it. If you deselect the Monitored check box for a storage subsystem, the following actions occur: All the data gathered by the server for the storage subsystem is removed from the enterprise repository. You can no longer run Monitoring, Alerting, or Policy Management jobs against the storage subsystem. c. Click Set disk alias.
3. The Set Disk Alias window (Figure 8-6) opens. a. Enter the Alias/Name. b. Click OK to finish.
4. You may need to refresh the GUI for the changes to become effective. Right-click an old entry in the Navigation Tree, and select Refresh.
Next steps
Now that you have set up the CIM Agent properties and specified to monitor the subsystems, run a probe against it to collect data about the disks and LUNs. After you do this, you can look at the results in different reports.
294
295
Port 2077
3. Locate the line AfpaEnable. Comment out the three Afpa... lines as shown in Example 8-2.
Example 8-2 Afpa #AfpaEnable #AfpaCache on #AfpaLogFile "C:\Program Files\WebSphere\AppServer\HTTPServer/logs/afpalog" V-ECLF
296
4. Locate the line that starts with <Directory. Modify the line to point the directory to C:\Program Files\IBM\TPC\Data\gui as shown in Example 8-3.
Example 8-3 Directory setting # --------------------------------------------------------------------------# This section defines server settings which affect which types of services # are allowed, and in what circumstances. # Each directory to which Apache has access, can be configured with respect # to which services and features are allowed and/or disabled in that # directory (and its subdirectories). # # # # Note: Where filenames are specified, you must use forward slashes instead of backslashes. e.g. "c:/apache" instead of "c:\apache". If the drive letter is omitted, the drive where Apache.exe is located will be assumed
# First, we configure the "default" to be a very restrictive set of # permissions. # # # # Note that from this point forward you must specifically allow particular features to be enabled - so if something's not working as you might expect, make sure that you have specifically enabled it below.
# This should be changed to whatever you set DocumentRoot to. #<Directory "C:\Program Files\WebSphere\AppServer\HTTPServer/htdocs/en_US"> <Directory "C:\Program Files\IBM\TPC\Data\gui">
5. Locate the line that starts with DocumentRoot. Modify the line to point the directory to C:\Program Files\IBM\TPC\Data\gui as shown in Example 8-4.
Example 8-4 DocumentRoot # # # # -------------------------------------------------------------------------------In the following section, you define the name space that users see of your http server. This also defines server settings which affect how requests are serviced, and how results should be formatted.
# See the tutorials at http://www.apache.org/ for # more information. # # # # Note: Where filenames are specified, you must use forward slashes instead of backslashes. e.g. "c:/apache" instead of "c:\apache". If the drive letter is omitted, the drive where Apache.exe is located will be assumed.
# DocumentRoot: The directory out of which you will serve your # documents. By default, all requests are taken from this directory, but # symbolic links and aliases may be used to point to other locations. #DocumentRoot "C:\Program Files\WebSphere\AppServer\HTTPServer/htdocs/en_US" DocumentRoot "C:\Program Files\IBM\TPC\Data\gui"
297
6. Locate the line that starts with DirectoryIndex. Modify the line to use TPCD.html as the index document as shown in Example 8-4 on page 297.
Example 8-5 Directory index # DirectoryIndex: Name of the file or files to use as a pre-written HTML # directory index. Separate multiple entries with spaces. #DirectoryIndex index.html DirectoryIndex tpcd.html
7. Save the file. 8. Start the HTTP server. 9. Open a command prompt. a. Change to the directory C:\Program Files\WebSphere\AppServer\HTTPServer. b. Type apache, and press Enter. 10.This starts the HTTP server as a foreground application. Now when you use a Web browser, simply enter:
http://servername:portumber
You see a Web page, and a Java application is then loaded. (Java is installed if necessary.) Note: Do not omit the http://. Since we do not use the default, you have to tell the browser which protocol to use.
298
299
3. The Web Site Creation Wizard opens, displaying the Welcome panel (see Figure 8-8). Click Next.
4. The Web Site Description panel (Figure 8-9) opens. Enter a description in the panel and click Next.
300
5. The IP Address and Port Settings panel (Figure 8-10) opens. Enter an unused port number and click Next.
6. In the Web Site Home Directory panel (Figure 8-11), enter the home directory of the Web server. This is the directory where the files for the remote Web GUI are stored. The default is C:\Program Files\IBM\TPC\Data\gui. Click Next.
301
7. The Web Site Access Permissions panel (Figure 8-12) opens. Accept the default access permissions, and click Next.
8. When you see the window indicating that you have successfully completed the Web Site Creation Wizard (Figure 8-13), click Finish.
9. In the Internet Information Services window (Figure 8-7 on page 299), right-click the new Web server entry, and select Properties.
302
10.The Data Manager Properties window (Figure 8-14) opens. a. Select the Documents tab.
b. Click Add. c. In the window that opens, enter tpcd.html. Click OK. d. Click OK to close the Properties window. 11.This starts the HTTP Server as a foreground application. Now, when you use a Web browser, simply enter:
http://servername:portumber
You see a Web page and a Java application is loaded. (Java is installed if necessary.) Note: Do not omit the http://. Since we dont use the default, you have to tell the browser which protocol to use.
303
2. The next panel (see Figure 8-16 on page 305) is the Software License Agreement. Click I accept the terms in the license agreement and click Next to continue.
304
3. The next panel allows you to select the components to be installed. For the remote console installation, select the User interface installations of Data, Disk, Fabric, and Replication (see Figure 8-17). Click Next to continue.
305
4. The next panel allows you to select which Remote GUI will be installed. Select the Productivity Center for Data (see Figure 8-18) and click Next to continue.
5. The next panel is informational (see Figure 8-19) and verifies that the Productivity Center for Data GUI will be installed.
306
6. The install package location panel is displayed. Specify the required information (see Figure 8-20) and click Next to continue.
7. Another information panel is displayed (see Figure 8-21) indicating that the product installer will be launched. Click Next to continue.
307
8. .In the window that opens, like the one in Figure 8-22, select Install Productivity Center for Data and click Next.
9. The License Agreement panel (Figure 8-23) opens. Select I have read and AGREE to abide by the license agreement above and click Next.
308
10.A License Agreement Confirmation window (Figure 8-24) opens. Click Yes to confirm.
11.The next window that opens prompts you to specify what you want to install (see Figure 8-25). In this example, we already had the agent installed on our machine. Therefore all options are still available. Select The GUI for reporting and click Next.
309
12.In the Productivity Center for Data Parameters panel (Figure 8-26), enter the Data Manager connection details and a Data Manager server name. Change the port if necessary and click Next.
310
13.In the Space Requirements panel (Figure 8-27), you can change the installation directory or leave the default. Click Next.
14.If the directory does not exist, you see the message shown in Figure 8-28. Click OK to continue or Cancel to change the directory.
311
15.You see the window shown in Figure 8-29 indicating that Productivity Center for Data has verified your entries and is ready to start the installation. Click Next to start the installation.
312
16.During the installation, you see a progress indicator. When the installation is finished, you see the Install Progress panel (Figure 8-30). Click Done to exit the installer.
The IBM TotalStorage Productivity Center for Data remote console is now installed. If the TotalStorage Productivity Center Launchpad is installed, it detects that Productivity Center for Data remote console is available the next time the LaunchPad is started. The LaunchPad can now be used to start Productivity Center for Data.
313
2. From the list of agents, select those you wish to monitor by checking the box under Licensed (Figure 8-32). After checking the desired boxes, click the RDBMS Logins tab.
314
3. To successfully scan a database, you must provide a login name and password for each instance. Click Add New... (Figure 8-33).
4. In the RDBMS Login Editor window, enter the required information: Database - the database type you wish to monitor Agent Host - the host you wish to monitor Instance - the name of the instance User - login ID for the instance Password - password for the instance Port - port where database is listening
315
316
You can specify these parameters: SNMP Community - The name of the SNMP community for sending traps Host - The system (event manager) which will receive the traps Port - The port on which traps will be sent (the standard port is 162) TEC (Tivoli Enterprise Console) TEC Server - for sending traps to; the system (TEC) that will receive the traps TEC Port - to which traps will be sent (the standard port is 5529) E-mail Mail Server - The mail server which will be used for sending the e-mail. Mail Port - The port used for sending the mail to the mail server. Default Domain - Default domain to be used for sending the e-mail. Return To - The return address for undeliverable e-mail. Reply To - The address to use when will replying to an Alert-triggered e-mail.
Alert Log Disposition Delete Alert Log Records - older than how long the Alert Log files will be kept.
317
318
Chapter 9.
319
9.1.1 IBM TotalStorage Productivity Center for Disk and Replication Base
When a supported storage area network (SAN) Manager is installed and configured, IBM TotalStorage Productivity Center for Disk and IBM TotalStorage Productivity Center for Replication leverages the SAN Manager to provide enhanced function. Along with basic device configuration functions, such as logical unit number (LUN) creation, allocation, assignment, and deletion for single and multiple devices, basic SAN management functions such as LUN discovery, allocation, and zoning are provided in one step. In Version 2.1 of TotalStorage Productivity Center, IBM TotalStorage Productivity Center for Fabric is the supported SAN Manager. The set of SAN Manager functions that are exploited are: The ability to retrieve SAN topology information, including switches, hosts, ports, and storage devices. The ability to retrieve and to modify the zoning configuration on the SAN. The ability to register for event notification this ensures that IBM TotalStorage Productivity Center for Disk is aware when the topology or zoning changes as new devices are discovered by the SAN Manager, and when host LUN configurations change.
9.1.2 SNMP
IBM TotalStorage Productivity Center for Fabric acts as a Simple Network Management Protocol (SNMP) manager to receive traps from managed devices in the event of status changes or updates. These traps are used to manage all the devices that the Productivity Center for Fabric is monitoring to provide the status window shown by NetView. These traps should then be passed onto a product, such as Tivoli Event Console (TEC), for central monitoring and management of multiple devices and products within your environment. When using the IBM TotalStorage Productivity Center Suite Installer, the SNMP configuration is performed for you. If you install IBM TotalStorage Productivity Center for Fabric manually, then you need to configure SNMP. The NetView code that is provided when you install IBM TotalStorage Productivity Center for Fabric is to be used only for this product. If you configure this NetView as your SNMP listening device for non-IBM TotalStorage Productivity Center for Fabric purposes, then you need to purchase the relevant NetView license.
320
Refer to the IBM Redbook Exploring Storage Management Efficiencies and Provisioning: Understanding IBM TotalStorage Productivity Center and IBM TotalStorage Productivity Center with Advanced Provisioning, SG24-6373, which presents an overview of the product components and functions. It explains the architecture and shows the use of storage provisioning workflows.
321
2. The next screen (see Figure 9-2) is the Software License Agreement. Click I accept the terms in the license agreement and click Next to continue.
3. In the Suite Installer panel (Figure 9-3), select the Agent installations of Data, Fabric, and CIM Agent option. Then click Next.
322
4. In the next window, select one or more Agent components (see Figure 9-4). In this example, we chose The Productivity Center for Fabric - Agent option. Click Next.
5. In the next panel, confirm the components to install. See Figure 9-5. Click Next.
323
6. As shown in Figure 9-6, enter the install package location. In our case, we installed from the I: drive. You will most likely use the product CD. Click Next.
7. A panel (Figure 9-7) opens indicating that the Productivity Center for Fabric installer will be launched. At this point, the Suite Installer is invoking the Installer process for the individual agent install. If you install the Agent directly from the CD, without using the Suite Installer, you commence the process after this point. The Suite Installer masks a few displays from you when it calls the product installer. Click Next.
324
8. In the window that opens, select the language for installation. We chose English. Click Next.
9. As In Figure 9-9 on page 326, specify the Fabric Manager Name and Fabric Manager Port Number. Type the Fabric Manager Name with the machine name where the Productivity Center for Fabric - Manager is installed. If it is a different domain, you must fully qualify the server name. In our case, colorado is the machine name of the server where Productivity Center for Fabric - Manager is installed. The port number is automatically inserted, but you can change it if you used a different port when you installed the Productivity Center for Fabric - Manager. Click Next.
325
10.The Host Authentication password is entered in the panel shown in Figure 9-10. This was specified during the Agent Manager install and is used for agent installs.
326
11.In the next panel (Figure 9-11), you have the option to change the default installation directory. We clicked Next to accept the default.
12.The Agent Information panel (Figure 9-12) asks you to specify a label which is applied to the Agent on this machine. We used the name of the machine. The port number is the port through which this Agent communicates. Click Next.
327
13.In the panel in Figure 9-13, specify the account that the Fabric Agent is to run under. We used the Administrator account.
14.In the Agent Management Information panel (Figure 9-14 on page 329), enter the location of the Tivoli Agent Manager. In our configuration, colorado is the machine name in our Domain Name Server (DNS) where Tivoli Agent Manager is installed. The Registration Password is the password that you used when you installed Tivoli Agent Manager. Click Next.
328
15.Finally you see a confirmation panel (Figure 9-15) that shows the installation summary. Review the information and click Next.
329
16.You see the installation status bar. Then you see a panel indicating a successful installation (Figure 9-16). Click Finish.
17.The panel in Figure 9-17 indicates the successful install of the Fabric agent.
18.You return to the Suite Installer window where you have the option to install other Agents. Click Cancel to finish.
330
Upon successful installation, you notice that nothing is added to the Start menu. The only evidence that the Agent is installed and running is that a Service is automatically started. Figure 9-18 shows the started Services in our Windows environment.
If you look in the Control Panel, under Add/Remove Programs, there is now an entry for IBM TotalStorage Productivity Center for Fabric - Agent. To remove the Agent, you click this entry.
Pre-installation tasks
Before you begin the installation, make sure that you have met the requirements that are discussed in the following sections.
331
332
2. The next screen (see Figure 9-20) is the Software License Agreement. Click I accept the terms in the license agreement and click Next to continue.
3. The first Suite Installer window (Figure 9-21) opens. Select the User interface Installations of Data, Fabric, and Replication option and then click Next.
333
4. In the next panel (Figure 9-22), select one or more remote GUI or command line components. To install the console, select The Productivity Center for Fabric - Remote GUI Client. Click Next.
334
6. As shown in Figure 9-24, enter the location of the source code for the installation. In most cases, this the product CD drive. In our case, we installed the code from the E: drive. Click Next.
7. The next panel (Figure 9-25) indicates that the Fabric Installer will be launched. If you install the Agent directly from the CD, without using the Suite Installer, you begin the process after this point. The Suite Installer masks a few displays from you when it calls the product installer. Click Next.
335
8. Select the language. We selected English. The Suite Installer launches the Fabric installer (see Figure 9-26).
9. The InstallShield Wizard opens for IBM TotalStorage Productivity Center for Fabric Console (see Figure 9-27, InstallShield Wizard for Console on page 336). Click Next.
336
10.In the next panel, you can specify the location of the directory into which the product will be installed. Figure 9-28 shows the default location. Click Next.
11.Specify the name and port number of the host where the Productivity Center for Fabric Manager is installed. See Figure 9-29. Click Next.
337
12.In the next panel (Figure 9-30), specify a starting port number from which the installer will allocate a series of ports for communication. We used the default. Click Next.
13.Type the password that you will use for all remote consoles or that the managed hosts will use for authentication with the manager (see Figure 9-31). This password must be the same as the one you entered in the Fabric Manager Installation. Click Next.
338
14.Specify the drive where NetView is to be installed or accept the default (see Figure 9-32). Click Next.
15.As shown in Figure 9-33, specify a password which will be used to run the NetView Service. Then click Next.
339
16.A panel opens that displays a summary of the installation (Figure 9-34). Click Next to begin the installation.
17.The installation completes successfully as indicated by the message in the panel shown in Figure 9-35. Click Next.
340
18.You are prompted to restart your machine (Figure 9-36). You may elect to restart immediately or at another time. We chose Yes, restart my computer. Click Finish.
After rebooting your system, you see a new Service is automatically started, as shown in Figure 9-37.
To start the Remote Console, click Start Programs Tivoli NetView NetView Console.
341
Disk array
SAN
Disk array Switch
SNMP
342
NetView listens for SNMP traps on port 162, and the default community is public. When the trap arrives to the Tivoli NetView console, it is logged in the NetView Event browser and then forwarded to Productivity Center for Fabric as shown in Figure 9-39. Tivoli NetView is configured during installation of the Productivity Center for Fabric Server for trap forwarding to the Productivity Center for Fabric Server.
Tivoli NetView
SAN Manager
NetView forwards SNMP traps to the defined TCP/IP port, which is the sixth port derived from the base port defined during installation. We used the base port 9550, so the trap forwarding port is 9556. With this setup, the SNMP trap information appears in the NetView Event browser. Productivity Center for Fabric uses this information for changing the topology map. Note: If the traps are not forwarded to Productivity Center for Fabric, the topology map is updated based on the information coming from Agents at regular polling intervals. The default Productivity Center for Fabric Server installation (including the NetView installation) sets up the trap forwarding correctly.
343
2. The trapfrwd daemon must be running before traps are forwarded. Tivoli NetView does not start this daemon by default. To configure Tivoli NetView to start the trapfrwd daemon, enter these commands at a command prompt:
ovaddobj \usr\ov\lrf\trapfrwd.lrf ovstart trapfrwd
3. To verify that trapfrwd is running, in NetView, select Options Server Setup. In the Server Setup Tivoli NetView window (Figure 9-40 on page 345), you see that trapfrwd is running.
344
After trap forwarding is enabled, configure the SAN components, such as switches, to send their SNMP traps to the NetView console. Note: This type of setup gives you the best results, especially for devices where you cannot change the number of SNMP recipients and the destination ports.
345
Method 3: Traps to the Productivity Center for Fabric and SNMP console
In this example, you set up the SAN devices to send SNMP traps to both the Productivity Center for Fabric Server and to a separate SNMP console, which you installed in your organization. See Figure 9-41.
p
Managed Host (Agent) Disk array Managed Host (Agent)
SAN
Disk array Switch
Disk array Disk array TotalStorage Productivity Center For Fabric Server
The receiving port number for the Productivity Center for Fabric Server is the primary port number plus six ports. The receiving port number for the SNMP console is 162. In this case traps are used to reflect the topology changes and they will display in the SNMP console events. The SNMP console, in this case, can be another Tivoli NetView installation or any other SNMP management application. For such a setup, the devices have to support setting multiple traps receivers and changing the trap destination port. Since this functionality is not supported in all devices, we do not recommend this scenario.
346
2. The SAN Configuration window (Figure 9-42) opens. a. Select the Switches and Other SNMP Agents tab on the left side.
b. You see the outband agents in the right panel. Define all the switches in the SAN that you want to monitor. To define such an Agent, click Add. c. The Enter IP Address window (Figure 9-43) opens. Enter the host name or IP address of the switch and click OK.
d. The agent appears in the agent list as shown in Figure 9-42. The state of the agent must be Contacted if you want Productivity Center for Fabric to get data from it. e. To remove an already defined agent, select it and click Remove.
347
Here is the procedure: 1. In the SAN Configuration window (Figure 9-42 on page 347), select the defined Agent and click Advanced. 2. In the SNMP Agent Configuration window (Figure 9-44), enter the user name and password for the switch login. Click OK to save this information.
You can now see zone information for your switches. Tip: You must enter user ID and password information only for one switch in each SAN to retrieve the zoning information. We recommend that you enter this information for at least two switches for redundancy. Enabling more switches than necessary for API zone discovery may slow performance.
348
After you set the poll interval, click OK to save the changes.
Tip: You do not need to configure the polling interval if all your devices are set to send SNMP traps to the local NetView console or the Productivity Center for Fabric Server. 349
350
10
Chapter 10.
Deployment of agents
Chapter 9, Configuring IBM TotalStorage Productivity Center for Fabric on page 319, covers the installation of the managers that are the central part of IBM TotalStorage Productivity Center. During that installation, the Resource Managers of Productivity Center for Data and Productivity Center for Fabric were installed and registered to a Tivoli Agent Manager, either an existing one, or one that was installed as a prerequisite in the first phase of the installation. This chapter explains how to set up the individual agents (subagents) on a managed host. The agents of Data Manager and Fabric Manager are called subagents, because they reside within the scope of the common agent.
351
352
2. In the next panel (Figure 10-2), select one or more agents to install. The options include the IBM TotalStorage Enterprise Storage Server (ESS) Common Information Model (CIM) Agent. However, this agent does not use any functions of Tivoli Agent Manager.
The next window asks you to enter the location of the installation code. Then the panel that follows tells you that the individual product installer is launched and you are asked to interact with it.
353
Figure 10-3 IBM TotalStorage Productivity Center for Data installation action
354
2. A window opens showing the license agreement (see Figure 10-4). Select the I have read and AGREE to abide by the license agreement above check box and click Next.
3. The License Agreement Confirmation window (Figure 10-5) opens. It asks you to confirm that you have read the license agreement. Click Yes.
355
4. In the next panel (Figure 10-6), choose the option of Productivity Center for Data that you want to install. To install the agent locally on the same machine on which the installer is currently running, select An agent on this machine and click Next.
356
5. In the Productivity Center for Data Parameters panel (Figure 10-7), enter the server name and port of your Productivity Center for Data Manager. In our environment, the name of the server was gallium, and the port was the default, which is 2078. We did not need to use the fully qualified host name, but this may be different in your environment. Click Next.
The installer tries to contact the Data Manager server. If this is successful, you see a message like Server gallium:2078 connection successful - server parameters verified in the Progress Log section of the installation window Figure 10-7. 6. The installer checks whether a common agent is already installed on the machine. Because in our environment no common agent was installed on the machine, the installer issues the message No compatible Common Agents were discovered so one will be installed. See the Progress log in Figure 10-8 on page 358.
357
7. As shown in Figure 10-8, enter the parameters of the agent: a. Use the suggested default port 9510. b. Deselect Agent should perform a SCAN when first brought up because this may take a long time and you want to schedule this during the night. c. Leave Agent may run scripts sent by server as selected. d. The Agent Registration Information is the password that you specified during the installation of the Tivoli Agent Manager. Note: Do not change the common agent port, because this may prevent the deployment of agents later. e. Click Next to continue the installation.
Figure 10-8 Parameter for the common agent and Data Agent
358
8. In the Space Requirements panel (Figure 10-9), accept the default directory for the common agent installation. Click Next to proceed with the installation.
9. If the directory that you specify does not exist, you see the message shown in Figure 10-10. Click OK to acknowledge this message and continue the installation.
359
10.Figure 10-11 shows the last panel before the installation starts. Review the progress log. If you want to review the parameters, click Prev to go to the previous panels. Then click Next.
360
11.The installation starts and displays the progress in the Install Progress window (Figure 10-12). The progress bar is not shown in the picture, but you can see the messages in the progress log. When the installation is complete, click Done to end the installation.
12.The installer closes now, and the Suite Installer is active again. It also reports the successful installation. Click Next to return to the panel shown in Figure 10-1 on page 352 to install another agent (for example a Fabric Agent) or click Cancel to exit the installation.
361
Note: For agent deployment, you do not need to have the certificate files available, because the target machines already have the necessary certificates installed during the common agent installation.
Data Agent
You can perform this installation from any machine. It does not have to be the Data Manager server itself. When you use the Suite Installer, there is no option to deploy agents. However, you can choose to install an agent, to launch the Data Manager installer, and later deploy an agent instead of installing it (see Figure 10-6 on page 356). We did not use the Suite Installer for the agent deployment. 1. Start the installer by running setup.exe from the Data Manager installation CD. 2. After a few seconds, you see the panel shown in Figure 10-13. If you have the Data Manager or the agent already installed on that machine where you started the installer, select Uninstall Productivity Center for Data or Apply maintenance to Productivity Center for Data. Click Next.
3. A window opens displaying the license agreement (see Figure 10-4 on page 355). Follow the steps as explained in steps 2 on page 355 and 3 on page 355.
362
4. In the next window that opens (Figure 10-14), select Agents on other machines. Then click Next.
363
5. In the Productivity Center for Data Parameters panel (Figure 10-15), enter the Productivity Center for Data server name and the port number. Then click Next.
364
6. The installer tries to verify your input by connecting to the Data Manager server. The message Server gallium:2078 connection successful - server parameters verified is displayed in the progress log (see Figure 10-16) if it is successful. Click Next. 7. In our environment, we did not have a Windows domain, so we entered the details of the target machines manually. Click Manually Enter Agents.
Figure 10-16 Select the Remote Agents to install: Manually entering Agents
365
8. In the Manually Enter Agent window (Figure 10-17), enter the IP address or host name of the target computer and a user ID and password of valid Windows users on that machine. You can only enter more than one machine here, if all the machines can be managed with the same user ID and password. Click OK after you enter all computers that can be managed with the same user ID.
366
9. The list with the computers that installs the subagent is updated and now appears as shown in Figure 10-18. If you want to install the subagent onto a second computer, but the computer uses a different user ID than the previous one, click Manually Enter Agents again to enter the information for that second computer. Repeat this step for every computer that uses a different user ID and password. After you enter all target computers, click Next.
Figure 10-18 Selecting the Remote Agents to install: Computers targeted for a remote agent install
10.At this time, the installer tries to contact the common agent on target computers to get information about them. This may take a while, so at first you cannot select anything in the window that is presented next (see Figure 10-19 on page 368). Look at the progress log in the lower section of the window to determine what is currently happening. If the installer cannot contact the target computer, verify that the common agent is running. You can do that by looking at the status of the Windows services of the target machine. Another way is to open a telnet connection from a Command Prompt to that machine on port 9510.
c:\>telnet 9.1.38.104 9510
If the common agent is running, it listens for requests on that port and opens a connection. You simply see a blank screen. If the common agent is not running, you see the message Connecting To 9.1.38.104...Could not open a connection to host on port 9510 : Connect failed. When the installer is done with this step, you see the message Productivity Center for Data subagent an 9.1.38.104 will be installed at C:\Program Files\tivoli\ep\TPC\Data. Deselect Agent should perform a SCAN when first brought up, because this may take a long time and you want to schedule this during the night. Click Install to start the deployment.
367
11.When the deployment is finished, you see the message shown in Figure 10-20. Review the progress log. Click OK to end the installation.
368
Fabric Agent
There are differences between the Data Agent deployment and the Fabric Agent deployment. To remotely deploy one or many fabric manager subagents, you must be logged on to the fabric manager server. This is different than the data subagent deployment where you can start the installation from any machine. At this time, there is no way to use the Suite Installer, so you have to use the native fabric manager installer. The Fabric Manager comes with a separate package for the Fabric Agent. Data Manager comes with only one installation program for all the possible install options (server, agent, remote agent or GUI). To start the deployment, you start the Fabric Manager installer. You do not start the installer for the Fabric Agent. 1. Launch setup.exe from the fabric manager installation media. 2. After a Java Virtual Machine is prepared and you select the language of the installer, a window opens that prompts you to select the type of installation to perform. See Figure 10-21. Select Remote Fabric Agent Deployment and click Next.
369
4. The IBM License Agreement Panel (Figure 10-22) opens. Select I accept the terms in the license agreement and click Next.
5. The installer connects to the Tivoli Agent Manager and presents a list of hosts. Select the hosts to deploy the agents. See Figure 10-23. Click Next to start the deployment.
370
6. The next panel (Figure 10-24) displays the selected hosts. Verify the information. You can click Back to change your selection or click Next to start the installation.
7. When the installation is completed, you see a summary window similar to the example in Figure 10-25. Click Finish.
371
372
Part 4
Part
373
374
11
Chapter 11.
375
376
Alternatively access IBM Director from Windows Start Programs IBM Director IBM Director Console Log on to IBM Director using the superuser id and password defined at installation. Please note that passwords are case sensitive. Login values are:IBM Director Server: Hostname of the machine where IBM Director is installed User ID: The username to logon with. This is the superuser ID. Enter it in the form <hostname>\<username> Password: The case sensitive superuser ID password Figure 11-2 shows the IBM Director Login panel you will see after launching IBM Director.
377
Note: The Manage Performance and Manage Replication tasks that you see in Figure 11-3 become visible when TotalStorage Productivity Center for Disk or TotalStorage Productivity Center for Replication are installed. Although this chapter covers Productivity Center common base, you would have installed either TotalStorage Productivity Center for Disk, TotalStorage Productivity Center for Replication, or both.
Figure 11-3 IBM Director Console with Productivity Center common base
378
When the Productivity Center common base functions are subsequently used to create/remove LUNs the volume inventory is automatically kept up to date and it is therefore not necessary to repeatedly run inventory collection from the storage devices. Version 2.3 of Productivity Center common base does not currently contain the full feature set of all functions for the supported storage devices. This will make it necessary to use the storage devices own management tools for some tasks. For instance you can create new VDisks with Productivity Center common base on a SAN Volume Controller but you can not delete them. You will need to use the SAN Volume Controllers own management tools to do this. For these types of changes to be reflected in Productivity Center common base an inventory collection will be necessary to re-synchronize the storage device and Productivity Center common base inventory. Attention: The use of volume inventory is common to ALL supported storage devices and must be performed before disk management functions are available. To start inventory collection, right-click the chosen device and select Perform Inventory Collection as shown in Figure 11-4.
A new panel will appear (Figure 11-5 on page 380) as a progress indication that the inventory process is running. At this stage Productivity Center common base is talking to the relevant CIMOM to collect volume information from the storage device. After a short while the information panel will indicated that the collection has been successful. You can now close this window.
379
Attention: When the panel in Figure 11-5 indicates that the collection has been done successfully, it does not necessarily mean that the volume information has been fully processed by Productivity Center common base at this point. To track the detailed processing status, launch the Inventory Status task as seen in Figure 11-6.
380
To see the processing status of an inventory collection, launch the Inventory Status task as shown in Figure 11-7.
The example Inventory Status panel seen in Figure 11-7 shows the progress of the processing for a SAN Volume Controller. Use the Refresh button in the bottom left of the panel to update it with the latest progress. You can also launch the Inventory Status panel before starting an inventory collection to watch the process end to end. In our test lab the inventory process time for an SVC took around 2 minutes, end to end.
381
Enter a more meaningful device name as in Figure 11-9 and click OK.
382
383
384
In either case, in the bottom left corner, the status will change from Ready to Starting Task, and it will remain this way until the volume inventory appears. Figure 11-12 shows the Volumes panel for the select ESS device that will appear.
385
When you click OK, TotalStorage Productivity Center for Fabric will be called to assist with zoning this volume to the host. If TotalStorage Productivity Center for Fabric is not installed, you will see a message panel as shown in Figure 11-14. When the volume has been successfully assigned to the selected host port, the Assign host ports panel will disappear and the ESS Volumes panel will be displayed once again, reflecting now the additional host port mapping number in the far right side of the panel, in the Number of host ports column. Note: If TotalStorage Productivity Center for Fabric is installed, refer to Chapter 14, Using TotalStorage Productivity Center for Fabric on page 703, for complete details of its operation. Also note that TotalStorage Productivity Center for Fabric is only invoked for zoning when assigning hosts to ports. It is not invoked to remove zones when hosts are unassigned.
386
Use the drop-down fields to select the Storage type and choose from Available arrays on the ESS. Then enter the number of volumes you want to create and the Volume quantity along with the Requested size. Finally select the host ports you want to have access to the new volumes from the Defined host ports scrolling list. You can select multiple hosts by holding down the control key <Ctrl> while clicking hosts. On clicking OK TotalStorage Productivity Center for Fabric will be called to assist with zoning the new volumes to the host(s). If TotalStorage Productivity Center for Fabric (formally known as TSANM) is not installed you will see a message panel as seen in Figure 11-16. If TotalStorage Productivity Center for Fabric is installed, refer to Chapter 14, Using TotalStorage Productivity Center for Fabric on page 703 for complete details of its operation.
387
388
389
390
In either case, in the bottom left corner, the status will change from Ready to Starting Task, and it will remain this way until the volume inventory appears. Figure 11-22 shows the Volumes panel for the select DS8000 device that will appear.
391
When you click OK, TotalStorage Productivity Center for Fabric will be called to assist with zoning this volume to the host. If TotalStorage Productivity Center for Fabric is not installed, you will see a message panel as seen in Figure 11-24 on page 393. When the volume has been successfully assigned to the selected host port, the Assign host ports panel will disappear and the ESS Volumes panel will be displayed once again, reflecting now the additional host port mapping number in the far right side of the panel, in the Number of host ports column. Note: If TotalStorage Productivity Center for Fabric is installed, refer to Chapter 14, Using TotalStorage Productivity Center for Fabric on page 703 for complete details of its operation. Also note that TotalStorage Productivity Center for Fabric is only invoked for zoning when assigning hosts to ports. It is not invoked to remove zones when hosts are unassigned.
392
Use the drop-down fields to select the Storage type and choose from Available arrays on the DS8000. Then enter the number of volumes you want to create and the Volume quantity along with the Requested size. Finally select the host ports you want to have access to the new volumes from the Defined host ports scrolling list. You can select multiple hosts by holding down the control key <Ctrl> while clicking hosts. On clicking OK, TotalStorage Productivity Center for Fabric will be called to assist with zoning the new volumes to the host(s). If TotalStorage Productivity Center for Fabric is not installed you will see a message panel as shown in Figure 11-26 on page 394.
393
394
Figure 11-28 shows the DS8000 device manager launched by Productivity Center common base.
Figure 11-28 DS8000 device manager launched by Productivity Center common base
395
396
397
Figure 11-30 The MDisk properties panel for SAN Volume Controller
Figure 11-30 shows candidate or unmanaged MDisks, which are available for inclusion into an existing MDisk group. To add one or more unmanaged disks to an existing MDisk group: Select the MDisk group from the pull-down menu. Select one MDisk from the list of candidate MDisks, or use the <Ctrl> key to select multiple disks. Click the OK button at the bottom of the screen and the selected MDisk(s) will be added to the MDisk group (Figure 11-31 on page 399).
398
399
Note: If TotalStorage Productivity Center for Fabric is installed and configured, extra panels will appear to create appropriate zoning for this operation. See Chapter 14, Using TotalStorage Productivity Center for Fabric on page 703 for details. Click OK to start a process that will create a new volume on the selected storage device and then add it to the SAN Volume Controllers MDisk group.
Figure 11-32 Create volumes to be added as MDisks Productivity Center common base will now request the specified storage amount from the specified back-end storage device (see Figure 11-33).
400
The next step is to add the MDisks to an MDisk group (see Figure 11-34).
401
Figure 11-35 shows the result of adding the mdisk4 to the selected MDisk group.
402
Tip: Before SAN Volume Controller virtual disk properties (VDisks) can be displayed, as with other storage devices managed by Productivity Center common base, an initial inventory must be completed. If you try to use the Virtual Disk function on a SAN Volume Controller that has not been inventoried, you will receive a notification that this needs to be done. To perform an inventory collection, see 11.4, Performing volume inventory on page 378.
403
Viewing VDisks
Figure 11-37 shows the VDisk inventory and volume attributes for the selected SAN Volume controller.
404
Creating a VDisk
To create a new VDisk, use the Create button as shown in Figure 11-37 on page 404. You need to provide a suitable VDisk name and select the MDisk group from which you want to create the VDisk. Specify the number of VDisks to be created and the size in megabytes or gigabytes that each VDisk should be. Figure 11-38 shows some example input in these fields.
The Host ports section of the VDisk properties panel allows you to use TotalStorage Productivity Center for Fabric functionality to perform zoning actions to provide VDisk access to specific host WWPNS. If TSANM is not installed, you will receive a warning If TotalStorage Productivity Center for Fabric is installed, refer to Chapter 14, Using TotalStorage Productivity Center for Fabric on page 703 for details on how to configure and use it.
405
Figure 11-39 shows that the creation of the VDisk was successful.
406
407
Figure 11-41 shows the volume inventory for the selected device. From this panel you can Create and Delete volumes or assign and unassign volumes to hosts.
408
Select the desired Storage Type and array from Available arrays using the drop-down menus. Then enter the Volume quantity and Requested volume size of the new volumes. Finally select the host posts you wish to assign to the new volumes from the Defined host ports scroll box, holding the <Crtl> key to select multiple ports. The Defined host ports section of the panel allows you to use TotalStorage Productivity Center for Fabric (formally TSANM) functionality to perform zoning actions to provide volume access to specific host WWPNS. If TSANM is not installed, you will receive the warning shown in Figure 11-43. If TotalStorage Productivity Center for Fabric is installed refer to Chapter 14, Using TotalStorage Productivity Center for Fabric on page 703 for details on how to configure and use it.
If TotalStorage Productivity Center for Fabric is not installed, click OK to continue. You should then see the panels shown below (Figure 11-43 through Figure 11-48 on page 412).
409
410
411
412
The Defined host ports section of the panel allows you to use TotalStorage Productivity Center for Fabric (formally TSANM) functionality to perform zoning actions to provide volume access to specific host WWPNS. If TSANM is not installed, you will receive the warning shown in Figure 11-50. If TotalStorage Productivity Center for Fabric is installed, refer to Chapter 14, Using TotalStorage Productivity Center for Fabric on page 703 for details on how to configure and use it.
413
If TotalStorage Productivity Center for Fabric is not installed, click OK to continue (Figure 11-51).
414
TotalStorage Productivity Center for Fabric is not called to perform zoning cleanup in version 2.1. This functionality is planned in a future release.
415
416
Here are the tasks to create an Event Action Plan. 1. To begin, do one of the following actions: Right-click Event Actions Plan in the Event Action Plans pane to access the context menu, and then select New. Select File New Event Action Plan from the menu bar. Double-click the Event Action Plan folder in the Event Action Plans pane (see Figure 11-57).
417
2. Enter the name you want to assign to the plan and click OK to save the new plan. The new plan entry with the name you assigned is displayed in the Event Action Plans pane. The plan is also added to the Event Action Plans task as a child entry in the Director Console (see Figure 11-58). Now that you have defined an event action plan, you can assign one or more filters and actions to the plan.
Notes: You can create a plan without having defined any filters or actions. The order in which you build a filter, action, and Event Action Plan does not matter. 3. Assign at least one filter to the Event Action Plan using one of the following methods: Drag the event filter from the Event Filters pane to the Event Action Plan in the Event Action Plans pane. Highlight the Event Action Plan, then right-click the event filter to display the context menu and select Add to Event Action Plan. Highlight the event filter, then right-click the Event Action Plan to display the context menu and select Add Event Filter (see Figure 11-59 on page 419).
418
The filter is now displayed as a child entry under the plan (see Figure 11-60).
419
4. Assign at least one action to at least one filter in the Event Action Plan using one of the following methods: Drag the action from the Actions pane to the target event filter under the desired Event Action Plan in the Event Action Plans pane. Highlight the target filter, then right-click the desired action to display the context menu and select Add to Event Action Plan. Highlight the desired action, then right-click the target filter to display the context menu and select Add Action. The action is now displayed as a child entry under the filter (see Figure 11-61).
5. Repeat the previous two steps for as many filter and action pairings as you want to add to the plan. You can assign multiple actions to a single filter and multiple filters to a single plan. Note: The plan you have just created is not active because it has not been applied to a managed system or a group. In the next section we explain how to apply an Event Action Plan to a managed system or group.
420
Repeat this step for all associations you want to make. You can activate the same Event Action Plan for multiple systems (see Figure 11-63).
Once applied, the plan is activated and displayed as a child entry of the managed system or group to which it is applied when the Associations - Event Action Plans item is checked.
421
Message Browser
When an event occurs, the Message Browser (see Figure 11-64) pops up on the server console.
If the message has not yet been viewed, then that Status for that message will be blank. When viewed, a checked envelope icon will appear under the Status column next to the message. To see greater detail on a particular message, select the message in the left pain and click the Event Details button (see Figure 11-65).
422
Export
Event Action Plans can be exported to three types of files: Archive: Backs up the selected action plan to a file that can be imported into any IBM Director Server. HTML: Creates a detailed listing of the selected action plans, including its filters and actions, in an HTML file format. XML: Creates a detailed listing of the selected action plans, including its filters and actions, in an XML file format. To export an Event Action Plan, do the following steps: 1. Open the Event Action Plan Builder. 2. Select an Event Action Plan from those available under the Event Action Plan folder. 3. Select File Export, then click the type of file you want to export to (see Figure 11-66). If this Event Action Plan will be imported by an IBM Director Server, then select Archive.
423
4. Name the archive and set a location to save in the Select Archive File for Export window as shown in Figure 11-67.
Tip: When you export an action plan, regardless of the type, the file is created on a local drive on the IBM Director Server. If an IBM Director Console is used to access the IBM Director Server, then the file could be saved to either the Server or the Console by selecting Server or Local from the Destinations pull-down. It cannot be saved to a network drive. Use the File Transfer task if you want to copy the file elsewhere.
424
Import
Event Action Plans can be imported from a file, which must be an Archive export of an action plan from another IBM Director Server. Follow these steps to import an Event Action Plan: 1. Transfer the archive file to be imported to a drive on the IBM Director Server. 2. Open the Event Action Plan Builder from the main Console window. 3. Click File Import Archive (see Figure 11-68).
4. From the Select File for Import window (see Figure 11-69), select the archive file and location. The file must be located on the IBM Director Server. If using the Console, you must transfer the file to the IBM Director Server before it can be imported.
425
5. Click OK to begin the import process. The Import Action Plan window opens, displaying the action plan to import (see Figure 11-70). If the action plan had been assigned previously to systems or groups, you will be given the option to preserve associations during the import. Select Import to complete the import process.
426
12
Chapter 12.
427
428
429
2. Or, right-click a storage device in the center column, and select the Performance Data Collection Panel menu option as shown in Figure 12-3.
Either operation results in a new window named Create Performance Data Collection Task (Figure 12-4). In this window you will specify: A task name A brief description of the task The sample frequency in minutes The duration of data collection task (in hours)
430
In our example, we are setting up a data collection task on an ESS with Device ID 2105.16603. After we have created a task name Cottle _ESS with sample frequency of 5 minutes and duration is 1 hour. It is possible to add more ESSs to the same data collection task, by clicking the Add button on the right-hand side. You can click individual devices, or select multiples by making use of the Ctrl key. See Figure 12-5 for an example of this panel. In our example, we created a task for the ESS with device ID 2105.22513.
431
Once we have established the scope of our data collection task and have clicked the OK button, we see our new data collection task available in the right-hand task column (see Figure 12-6). We have created task Cottle_ESS in the example. Tip: When providing a description for a new data collection task, you may elect to provide information about the duration and frequency of the task.
432
In order to schedule it, right-click the selected task (see Figure 12-7).
You have the option to use the job scheduling facility of TotalStorage Productivity Center for Disk, or to execute the task immediately.
433
If you select Execute Now, you will see a panel similar to the one in Figure 12-9, providing you with some information about task name and task status, including the time the task was initialized.
If you would rather schedule the task to occur at a future time, or to specify additional parameters for the job schedule, you can walk through the panel in Figure 12-10. You may provide a scheduled job description for the scheduled job. In our example, we created a job, 24March Cottle ESS.
434
Once you are finished customizing the job options, you may save it using File Save as menu. Or, you may do this by clicking the diskette icon panel. in the top left corner of the advanced
435
When you save with advanced job options, you may provide a descriptive name for the job as shown in Figure 12-12.
You should receive a confirmation that your job has been saved as shown in Figure 12-13.
436
437
Upon double-clicking Task Status, it launches the following panel as shown in Figure 12-15.
To review the task status, you can click the task shown under the Task name column. For example, we selected the task FCA18P, which was aborted, as shown in Figure 12-16 on page 439. Subsequently, it will show the details with Device ID, Device status, and Error Message ID in the Device status box. You can click the entry in the device status box. It will further show the Error message in the Error message box.
438
439
The performance database panel can be launched on clicking Performance Database as shown in Figure 12-17. It will display the Performance Database Properties panel as shown in Figure 12-18 on page 441.
440
You can use the performance database panel to specify properties for a performance database purge task. The sizing function on this panel shows used space and free space in the database. You can choose to purge performance data based on age of the data, the type of the data, and the storage devices associated with the data (Figure 12-18).
The Performance database properties panel shows the following data: Database name The name of the database Database location The file system on which the database resides. Total file system capacity The total capacity available to the file system, in gigabytes. Space currently used on file system Space is shown in gigabytes and also by percentage. Performance manager database full The amount of space used by Performance Manager. The percentage shown is the percentage of available space (total space - currently used space) used by the Performance Manager database.
441
The following formula is used to derive the percentage of disk space full in the Performance Manager database: a = the total capacity of the file system b = the total allocated space for Performance Manager database on the file system c = the portion of the allocated space that is used by the Performance Manager database For any decimal amount over a particular number, the percentage is rounded up to the next largest integer. For example, 5.1% is rounded to and displayed as 6%. Space status advisor The Space status advisor monitors the amount of space used by the Performance Manager database and advises you as to whether you should purge data. The advisor levels are: Low: You do not need to purge data now High: You should purge data soon Critical: You need to purge data now Disk space thresholds for status categories: low if utilization <0.8, high if 0.8 <= utilization <0.9 and critical otherwise. The delimiters between low/high/critical are 80% and 90% full. Purge database options Groups the database purge information. Name Type A name for the performance database purge task. The maximum length for a name can be from 1 to 250 characters. Description (optional) Type a description for the performance database purge task. The maximum length for a description can be from 1 to 250 characters. Device type Select one or more storage device types for the performance database purge. Options are SVC, ESS, or All. (Default is All.) Purge performance data older than Select the maximum age for data to be retained when the purge task is run. You can specify this value in days (1-365) or years (1-10). For example, if you select the Days button and a value of 10, the database purge task will purge all data older than 10 days when it is run. Therefore, if it has been more than 10 days since the task was run, all performance data would be purged. Defaults are 365 days or 10 years. Purge data containing threshold exception information Deselecting this option will preserve performance data that contains information about threshold exceptions. This information is required to display exception gauges. This option is selected by default. Save as task button When you click Save as task, the information you specified is saved and the panel closes. The newly created task is saved to the IBM Director Task pane under the Performance Manager Database. Once it is saved, the task can be scheduled using the IBM Director scheduler function.
442
Creating a gauge
Open the IBM Director and do one of the following tasks: Right-click the storage device in the center pane and select Gauges (see Figure 12-19).
You can click Gauges on the panel shown and it will produce the Job Status window as shown in Figure 12-21 on page 444. It is also possible to launch Gauge creation by expanding Multiple Device Manager - Manage Performance in the rightmost column. Drag the Gauges item to the storage device desired and drop to open the gauges for that device (see Figure 12-20 on page 444).
443
This will produce the Job status window (see Figure 12-21) while the Performance gauges window opens. You will see the Job status window while other selected windows are opening.
The Performance gauges window will be empty until a gauge is created for use. We have created three gauges.(see Figure 12-22).
Clicking the Create button to the left brings up the Job status window while the Create performance gauge window opens.
444
The Create performance gauge window changes values depending on whether the cluster, array, or volume items are selected in the left pane. Clicking the cluster item in the left pane produces a window as seen in Figure 12-23.
Performance
Cluster Performance gauges provide details on the average cache holding time in seconds as well as the percent of I/O requests that were delayed due to NVS memory shortages. Two Cluster Performance gauges are required per ESS to view the available historical data for each cluster. Additional gauges can be created to view live performance data. Device: Select the storage device and time period from which to build the performance gauge. The time period can be changed for this device within the gauge window, thus allowing an overall or detailed view of the data. Name: Enter a name that is both descriptive of the type of gauge as well as the detail provided by the gauge. The name must not contain white space, special characters, or exceed 100 characters in length. Also, the name must be unique on the TotalStorage Productivity Center for Disk Performance Manager Server. If test were used as a gauge name, then it could not be used for another gauge - even if another storage device were selected as it would not be unique in the database. Example names: 28019P_C1H would represent the ESS serial number (28019), the performance gauge type (P), the cluster (C1), and historical (H), while 28019E would represent the exception (E) gauge for the same ESS. Gauges for the clusters and arrays would build on that nomenclature to group the gauges by ESS on the Gauges window.
Chapter 12. Using TotalStorage Productivity Center Performance Manager
445
Description: Use this space to enter a detailed description of the gauge that will appear on the gauge and in the Gauges window. Metric(s): Click the metric(s) that will be display by default when the gauge is opened for viewing. Those metrics with the same value under the Units column in the Metrics table can be selected together using either Shift mouse-click or Ctrl mouse-click. The metrics in this field can be changed on the historical gauge after the gauge has been opened for viewing. In other words, a historical gauge for each metric or group of metrics is not necessary. However, these metrics cannot be changed for live gauges. A new gauge is required for each metric or group of metrics desired. Component: Select a single device from the Component table. This field cannot be changed when the gauge is opened for viewing. Data points: Selecting this radio button enables the gauge to display most recent data being obtained from currently running performance collectors against the storage device. One most recent performance data gauge is required per cluster and per metric to view live collection data. The Device pull-down menu displays text informing the user whether or not a performance collection task is running against this Device. You can select number of datapoints for your requirements to display the last x data points from the date of the last collection. The data collection could be currently running or the most recent one. Date Range: Selecting this radio button presents data over a range of dates/times. Enter the range of dates this gauge will use as a default for the gauge. The date and time values may be adjusted within the gauge to any value before or after the default values and the gauge will display any relevant data for the updated time period. Display gauge: Checking this box will display the newly created gauge after clicking the OK button. Otherwise, if left blank, the gauge will be saved without displaying.
446
Click the OK button when ready to save the performance gauge (see Figure 12-24). In this example, we have created a gauge with the name 22513C1H and the description, average cache holding time. We selected a starting and ending date as 11-March-2005. This corresponds with our data collection task schedule.
The gauge appears after clicking the OK button with the Display gauge box checked or when the Display button is clicked after selecting the appropriate gauge on the Performance gauges window (see Figure 12-26 on page 448). If you decide not to display gauge and save only, then you will see a panel as shown here in Figure 12-25.
447
The top of the gauge contains the following labels: Graph Name Description Device Component level Component ID Threshold The name of the gauge The description of the gauge The storage device selected for the gauge Cluster, array, volume The ID # of the component (cluster, array, volume) The thresholds that were applied to the metrics
Time of last data collection Date and time of the last data collection The center of the gauge contains the only fields that may be altered in the Display Properties section. The Metrics may be selected either individually or in groups as long as the data types are the same (for example, seconds with seconds, milliseconds with milliseconds, or percent with percent). Click the Apply button to force a Performance Gauge section update with the new y-axis data. The Start Date:, End Date:, Start Time:, and End Time: fields may be varied to either expand the scope of the gauge or narrow it for a more granular view of the data. Click the Apply button to force a Performance Gauge section update with the new x-axis data. For example, we applied Total I/O Rate metric to the saved gauge, and the resultant graph is shown in Figure 12-27 on page 449. Here, the Performance Gauge section of the gauge displays graphically, the information over time selected by the gauge and the options in the Display Properties section.
448
Figure 12-27 Cluster performance gauge with applied I/O rate metric
Click the Refresh button in the Performance Gauge section to update the graph with the original metrics and date/time criteria. The date and time of the last refresh appear to the right of the Refresh button. The date and time displayed are updated first, followed by the contents of the graph, which can take up to several minutes to update. Finally, the data used to generate the graph is displayed at the bottom of the window (see Figure 12-28 on page 450). Each of the columns in the data section can be sorted up or down by clicking the column heading (see Figure 12-32 on page 453). The sort reads the data from left to right, so the results may not be as expected. The gauges for the array and volume components function in the same manner as the cluster gauge created above.
449
Exception
Exception gauges display data only for those active thresholds that were crossed during the reporting period. One Exception gauge displays threshold exceptions for the entire storage device based on the thresholds active at the time of collection.
450
To create an exception gauge, select Exception from the Type pull-down menu (see Figure 12-29).
By default, the Cluster will be highlighted in the left pane and the metrics and component sections will not be available. Device: Select the storage device and time period from which to build the performance gauge. The time period can be changed for this device within the gauge window thus allowing an overall or detailed view of the data. Name: Enter a name that is both descriptive of the type of gauge as well as the detail provided by the gauge. The name must not contain white space, special characters, or exceed 100 characters in length. Also, the name must be unique on the TotalStorage Productivity Center for Disk Performance Manager Server. Description: Use this space to enter a detailed description of the gauge that will appear on the gauge and in the Gauges window Date Range: Selecting this radio button presents data over a range of dates/times. Enter the range of dates this gauge will use as a default for the gauge. The date and time values may be adjusted within the gauge to any value before or after the default values and the gauge will display any relevant data for the updated time period. Display gauge: Checking this box will display the newly created gauge after clicking the OK button. Otherwise, if left blank, the gauge will be saved without displaying.
451
Click the OK button when ready to save the performance gauge. We created an exception gauge as shown in Figure 12-30.
The top of the gauge contains the following labels: Graph Name Description Device Threshold The name of the gauge The description of the gauge The storage device selected for the gauge The thresholds that were applied to the metrics
Time of last data collection Date and time of the last data collection The center of the gauge contains the only fields that may be altered in the Display Properties section. The Start Date: and End Date: fields may be varied to either expand the scope of the gauge or narrow it for a more granular view of the data. Click the Apply button to force an Exceptions Gauge section update with the new x-axis data. The Exceptions Gauge section of the gauge displays graphically, the information over time selected by the gauge, and the options in the Display Properties section (see Figure 12-31 on page 453).
452
Click the Refresh button in the Exceptions Gauge section to update the graph with the original date criteria. The date and time of the last refresh appear to the right of the Refresh button. The date and time displayed are updated first, followed by the contents of the graph, which can take up to several minutes to update. Finally, the data used to generate the graph are displayed at the bottom of the window. Each of the columns in the data section can be sorted up or down by clicking the column heading (see Figure 12-32).
453
Display Gauges
To display previously created gauges, either right-click the storage device and select gauges (see Figure 12-19 on page 443) or drag and drop the Gauges item on the storage device (see Figure 12-20 on page 444) to open the Performance gauges window, shown here in Figure 12-33).
Gauge Properties
The Properties button allows the following fields or choices to be modified.
Performance
These are the performance related possibilities: Description Metrics Component Data points Date range date and time ranges
454
You can change the data displayed in the gauge from Data points with an active data collection to Date range (see Figure 12-34). Selecting Date range allows you to choose the Start date and End Date using the performance data stored in the DB2 database.
455
Exception
You can change the Type property of the gauge definition from Performance to Exception. For a gauge type of Exception, you can only choose to view data for a Date range (see Figure 12-35).
Delete a gauge
To delete a previously created gauge, either right-click the storage device and select gauges (see Figure 12-19 on page 443) or drag and drop the Gauges item on the storage device (see Figure 12-20 on page 444) to open the Performance gauges window shown in Figure 12-33 on page 454. Select the gauge to remove and click Delete. A pop-up window will prompt for confirmation to remove the gauge (see Figure 12-36).
456
To confirm, click Yes and the gauge will be removed. The gauge name may now be reused, if desired.
457
Upon opening the thresholds submenu, you will see the following display, which shows the default thresholds in place for ESS, as shown in Figure 12-38.
On the right-hand side, there are buttons for Enable, Disable, Copy Threshold Properties, Filters, and Properties. If the selected task is already enabled, then the Enable button will appear greyed out, as in our case. If we attempt to disable a threshold that is currently enabled, by clicking the Disable button, a message will be displayed as shown in Figure 12-39.
You may elect to continue, and disable the selected threshold, or to cancel the operation by clicking Dont disable threshold.
458
The copy threshold properties button will allow you to copy existing thresholds to other devices of similar type (ESS, in our case). The window in Figure 12-40 is displayed.
Note: As shown in Figure 12-40, the copying threshold panel is aware that we have registered on our ESS CIM agent host both clusters of our model 800 ESS, as indicated by the semicolon delimited IP address field for the device ID 2105.22219. The Filters window is another available thresholds option. From this panel, you can enable, disable, and modify existing filter values against selected thresholds as shown in Figure 12-41.
459
Finally, you can open the properties panel for a selected threshold, and are shown the panel in Figure 12-42. You have options to acknowledge the values at their current settings, or modify the warning or error levels, or select the alert level (none, warning only, and warning or error are the available options).
460
As long as at least one data collection task has been completed, you are able to proceed with the steps to create a gauge to view your performance data.
SVC has following thresholds with their default properties: VDisk I/Os rate Total number of virtual disk I/Os for each I/O group. SAN Volume Controller defaults: Status: Disabled Warning: None Error: None VDisk bytes per second Virtual disk bytes per second for each I/O group. SAN Volume Controller defaults: Status: Disabled Warning: None Error: None
461
MDisk I/O rate Total number of managed disk I/Os for each managed disk group. SAN Volume Controller defaults: Status: Disabled Warning: None Error: None MDisk bytes per second Managed disk bytes per second for each managed disk group. SAN Volume Controller defaults: Status: Disabled Warning: None Error: None You may only enable a particular threshold once minimum values for warning and error levels have been defined. If you attempt to select a threshold and enable it without first modifying these values, you will see a notification like the one in Figure 12-45.
Tip: In TotalStorage Productivity Center for Disk, default threshold warning or error values of -1.0 are indicators that there is no recommended minimum value for the threshold and are therefore entirely user defined. You may elect to provide any reasonable value for these thresholds, keeping in mind the workload in your environment. To modify the warning and error values for a given threshold, you may select the threshold, and click the Properties button. The panel in Figure 12-46 will be shown. You can modify the threshold as appropriate, and accept the new values by selecting the OK button.
462
As long as at least one data collection task has been completed, you are able to proceed with the steps to create a gauge to view your performance data (Figure 12-49 on page 464 through Figure 12-51 on page 466).
463
464
465
466
You may only enable a particular threshold once minimum values for warning and error levels have been defined. If you attempt to select a threshold and enable it without first modifying these values, you will see a notification like the one in Figure 12-45 on page 462.
Tip: In TotalStorage Productivity Center for Disk, default threshold warning or error values of -1.0 are indicators that there is no recommended minimum value for the threshold and are therefore entirely user defined. You may elect to provide any reasonable value for these thresholds, keeping in mind the workload in your environment. To modify the warning and error values for a given threshold, you may select the threshold, and click the Properties button. The panel in Figure 12-46 on page 462 will be shown. You can modify the threshold as appropriate, and accept the new values by clicking the OK button.
467
468
Click the Create button to create a new gauge. You will see a panel similar to Figure 12-56.
We selected Cluster in the top left corner, Total I/O Rate metric in the metrics box, and Cluster 1 in the component box. Also, we entered the following parameters: Name: 22219P_drilldown_analysis Description: Eiderdown analysis for 22219 ESS
469
For the Date range, we selected our historical data collection sampling period and clicked Display gauge. Upon clicking OK, we got the next panel as shown in Figure 12-57.
470
Figure 12-58 Zooming on specific time period for Total IO rate metric
471
The subsequent panel is shown in Figure 12-60. We selected Array level metric for Cluster 1, Device Adapter 1, Loop A, and Disk Group 2 for Avg. Response time as circled in Figure 12-60.
472
473
The chart Writes and Total I/O are shown as overlapping and Reads are shown as zero. Tip: If you select multiple metrics that do not have the same units for the y-axis, then the error is displayed as shown in Figure 12-63.
474
startesscollection/startsvccollection: These commands are used to build and run data collection against the ESS or SAN Volume Controller, respectively. lscollection: This command is used to list the running, aborted, or finished data collection tasks on the Performance Management server. stopcollection: This command may be used to stop data collection against a specified task name. lsgauge: You can use the lsgauge command to display a list of existing gauge names, types, device types, device IDs, modified dates, and description information. rmgauge: Use this command to remove existing gauges. showgauge: This command is used to display performance data output using an existing defined gauge. setessthresh/setsvcthresh: These two commands are respectively used to set ESS and SAN Volume Controller performance thresholds. cpthresh: You can use the cpthresh command to copy threshold properties from one selected device to one or more other devices. setfilter: You can use setfilter to set or change the existing threshold filters. lsfilter: This command may be used to display the threshold filter settings for all devices specified.
475
setoutput: This command may be used to view or modify the existing data collection output formats, including settings for paging, row printing, format (default, XML, or character delimited), header printing, and output verbosity. lsdev: This command can be used list the storage devices that are used by TotalStorage Productivity Center for Disk. lslun: This command can be used list the LUNs or Performance Manager volumes associated with storage devices. lsthreshold: This command can be used to list the threshold status associated with storage devices. lsgauge: This command can be used list the existing gauge names, gauge type, device name, device ID, date modified, and optionally device information. showgauge: Use this command to display performance output by triggering an existing gauge. showcapacity: This command displays managed capacity, the sum of managed capacity by device type, and the total of all ESS and SAN Volume Controller managed storage. showdbinfo: This command displays percent full, used space, and free space of the Performance Manager database. lsprofile: Use this command to display Volume Performance Advisor profiles. cpprofile: Use this command to copy Volume Performance Advisor profiles. mkprofile: Use this command to create a workload profile that you can use later with mkrecom command to create a performance recommendation for ESS volume allocation. mkreom: Use this command to generate and, optionally, apply a performance LUN advisor recommendation for ESS volumes. lsdbpurge: This command can be used to display the status of database purge tasks running in TotalStorage Productivity Center for Disk. tracklun: This command can be used to obtain historical performance statistics used to create a profile. startdbpurge: Use this command to start a database purge task. showdev: Use this command to display device properties. setoutput: This command sets output format for the administrative command line interface. cpthresh: This command can be used to copy threshold properties from one device to other devices that are of the same type. rmprofile: Use this command to remove delete performance LUN advisor profiles.
476
Figure 12-65 Sample perfcli command from Windows command line interface
Figure 12-66 and Figure 12-67 show perfcli sample commands within the perfcli tool.
477
478
Storage administrators and application developers need tools that pull together all the components of the decision process used for provisioning storage. They need tools to characterize and manage workload profiles. They need tools to monitor existing performance, and tools to help them understand the impact of future workloads on current performance. What they need is a tool that automates this entire process, which is what VPA for ESS does.
479
The VPA will tend to utilize resources that can best accommodate a particular type of workload. For example, high write content will make Raid 5 arrays busier than RAID 10 and VPA will therefore bias to RAID 10. Faster devices will be less busy, so VPA biases allocations to the faster devices. VPA also analyzes the historical data to determine how busy the internal ESS components (arrays, disk adapters, clusters) are due to other workloads. In this way, VPA tries to avoid allocating on already busy ESS components. If VPA has a choice among several places to allocate volumes, and they appear to be about equal, it is designed to apply a randomizing factor. This keeps the advisor from always giving the same advice, which might cause certain resources to be overloaded if everyone followed that advice. This also means that several usages of VPA by the same user may not necessarily get the same advice, even if the workload profiles are identical. Note: VPA tries to allocate the fewest possible volumes, as long as it can allocate on low utilization components. If the components look too busy, it will allocate more (smaller) volumes as a way of spreading the workload.It will not recommend more volumes than the maximum specified by the user. VPA may however be required to recommend allocation on very busy components. A utilization indicator in the user panels will indicate whether allocations would cause components to become heavily utilized. The I/O demand specified in the workload profile for the new storage being allocated is not a Service Level Agreement (SLA). In other words, there is no guarantee that the new storage, once allocated, will perform at or above the specified access density. The VPA will make recommendations unless the available space on the target devices is exhausted. An invocation of VPA can be used for multiple recommendations. To handle a situation when multiple sets of volumes are to be allocated with different workload profiles, it is important that the same VPA wizard be used for all sets of recommendations. Select Make additional recommendations on the View Recommendations page, as opposed to starting a completely new sequence for each separate set of volumes to be allocated. VPA is designed to remember each additional (hypothetical) workload when making additional recommendations. There are, of course, limitations to the use of an expert advisor such as VPA. There may well be other constraints (like source and target Flashcopy requirements), which must be considered. Sometimes these constraints can be accommodated with careful use of the tool, and sometimes they may be so severe that the tool must be used very carefully. That is why VPA is designed as an advisor. In summary, the Volume Performance Advisor (VPA) provides you a tool to help automate complex decisions involved in data placement and provisioning. It short, it represents a future direction of storage management software! Computers should monitor their resources and make autonomic adjustments based on the information. The VPA is an expert advisor which provides you a step in that direction.
481
On the Windows platform, follow these steps: 1. Start Run regedit.exe 2. Open the HKEY_LOCAL_MACHINE SOFTWARE Tivoli Director CurrentVersion file. 3. Modify the LogOutput. Set the value to be equal to 1. 4. Reboot the server The output log location from the instructions above is X:/program files/ibm/director/log (where X is the drive where the Director application was installed). The log file for the Director is com.tivoli.console.ConsoleLauncher.stderr. On the Linux platform, TWGRas.properties sets the output logging on. You need to remove the comment from the last line in the file (twg.sysout=1) and ensure that you have set TWG_DEBUG_CONSOLE as an environment variable. For example in bash: $export TWG_DEBUG_CONSOLE=true
Workload profiles
The basic VPA concept, and the storage administrators goal, is to balance the workload across all device components. This requires detailed ESS configuration information including all components (clusters, device adapters, logical subsystems, ranks, and volumes)
482
To express the workload represented by the new volumes, they are assigned a workload profile. A workload profile contains various performance attributes: I/O demand, in I/O operations per second per GB of volume size Average transfer size, in KBs per second Percentage mix of I/O - sequential or random, and read or write Cache utilization - percent of: cache hits for random reads, cache misses for random writes Peak activity time - the time period when the workload is most active You can create your own workload profile definitions in two ways By copying existing profiles, and editing their attributes By performing an analysis of existing volumes in the environment This second option is known as a Workload Analysis. You may select one or more existing volumes, and the historical performance data for these volumes retrieved, to determine their (average) performance behavior over time.
483
4. You can choose two methods to launch VPA: a. Drag and Drop the VPA icon to the storage device to be examined (see Figure 12-69).
Figure 12-69 Drag and Drop the VPA icon to the storage device
484
b. Select storage device right-click the device select Volume Performance Advisor (see Figure 12-70).
If a storage device is selected for the drag and drop step, that is not in the scope of the VPA, the following message will open (see Figure 12-71). Devices such as a CIMOM or an SNMP device will generate this error. Only ESS is supported at this time.
485
In the ESS User Validation panel, specify the user name, password, and port for each of the IBM TotalStorage Enterprise Storage Servers (ESSs) that you want to examine. During the initial setup of the VPA, on the ESS User Validation window, you need to first select the ESS (as shown in Figure 12-73 on page 487) and then input correct username, correct password and password verification.
You must click Set after you have input the correct username, password, and password verification in the appropriate fields (see highlighted portion with circle in Figure 12-74 on page 487). When you click Set, the application will populate the data you input (masked) into the correct Device Information fields in the Device Information box (see Figure 12-75 on page 487).
If you do not click Set, before selecting OK, the following error(s) will appear depending on what data needs to be entered. BWN005921E (ESS Specialist username has not been entered correctly or applied) BWN005922E (ESS Specialist password has not been entered correctly or applied) If you encounter these errors, ensure you have correctly input the values in the input fields in the lower part of the ESS user validation window and then retry by clicking OK. The ESS user validation window contains the following fields: Devices table - Select an ESS from this table. It includes device IDs and device IP addresses of the ESS devices on which this task was dropped. ESS Specialist username - Type a valid ESS Specialist user name and password for the selected ESS. Subsequent displays of the same information for this ESS show the user name and password that was entered. You can change the user name by entering a new user name in this field. ESS Specialist password - Type a valid ESS Specialist password for the selected ESS. Any existing password entries are removed when you change the ESS user name. Confirm password - Type the valid ESS Specialist password again exactly as you typed it in the password field. ESS Specialist port - Type a valid ESS port number. The default is 80. Set button - Click to set names, passwords, and ports without closing the panel. Remove button - Click to remove the selected information. Add button - Click to invoke the Add devices panel.
486
487
Click the OK button to save the changes and close the panel. The application will attempt to access the ESS storage device. The error message in Figure 12-76 can be indicative of use of an incorrect username or password for authentication. Additionally, If you have a firewall and are not adequately authenticating to the storage device, the error may appear. If this does occur, check to ensure you are using the correct username and password for the authentication and have firewall access and are properly authenticating to establish storage device connectivity.
488
You use the Volume performance advisor - Settings window to identify your requirements for host attachment and the total amount of space that you need. You can also use this panel to specify volume number and size constraints, if any. We will begin with our example as shown in Figure 12-78.
489
Here we describe the fields in this window: Total space required (GB) - Type the total space required in gigabytes. The smallest allowed value is 0.1 GB. We requested 3 GB for our example. Note: You cannot exceed the volume space available for examination on the server(s) you select. To show the error, in this example we selected host Zombie and Total required space as 400 GB. We got the error shown in Figure 12-79. Action: Retry with different values and look at the server log for details. Solution(s): Select a smaller maximum Total (volume) Space required GB and retry this step. Select more hosts which will include adequate volume space for this task. You may want to select the box entitled Consider volumes that have already been allocated but not assigned in the performance recommendation. Director log file enabling will generate logs for troubleshooting Director GUI components, including the PM coswearer. In this example, the file we reference is: com.tivoli.console.ConsoleLauncher.stderr. (com.tivoli.console.ConsoleLauncher.stdout is also useful) The sample log is shown in Figure 12-80.
Specify a volume size range button - Click the button to activate the field, then use the Minimum size (GB) spinner and the Maximum size (GB) spinner to specify the range. In this example, we selected 1 GB as minimum and 3 GB as maximum.
490
Specify a volume quantity range button - Click the button to activate the field, then use the Minimum number spinner and the Maximum number spinner to specify the range. Consider volumes that have already been allocated but not assigned to hosts in the performance recommendation. If you check this box, VPA will use these types of volumes in the volume performance examination process. When this box (Consider volumes...) is checked and you click Next, the VPA wizard will open the following warning window (see Figure 12-81).
Note: The BWN005996W message is a warning (W). You have selected to reuse unassigned existing volumes which could potentially cause data loss. Go Back to the VPA Settings window by clicking OK if you do not want to consider unassigned volumes. Click the Help button for more information. Explanation: The Volume Performance Advisor will assume that all currently unassigned volumes are not in use, and may recommend the reuse of these volumes. If any of these unassigned volumes are in use for example, as replication targets or other data replication purposes and these volumes are recommended for reuse, the result could be potential data loss. Action: Go back to the Settings window and unselect Consider volumes that have already been allocated but not assigned to hosts in the performance recommendation if you do not want to consider volumes that may potentially be used for other purposes. If you want to continue to consider unassigned volumes in your recommendations, then continue. Host Attachments table - Select one or more hosts from this table. This table lists all hosts (by device ID) known to the ESS that you selected for this task. It is important to only choose hosts for volume consideration that are the same server type. It is also important to note that the VPA takes into consideration the maximum volume limitations of server type such as (Windows 256 volumes maximum) and AIX (approximately 4000 volumes). If you select a volume range above the server limit, VPA will display an error. In our example we used the host Zombie. Next button - Click to invoke the Choose workload profile window. You use this window to select a workload profile from a list of existing profile templates.
491
5. Click Next, after inputting your preferred parameters, and the Choose workload profile window will display (see Figure 12-82).
492
Note: You cannot modify the properties of the workload profile from this panel. The panel options are greyed out (inactive). You can make changes to a workload profile from Manage Profile Create like panel. Next button - Click to invoke the Choose candidate locations window. You can use this panel to select volume locations for the VPA to consider.
6. After reviewing the properties for predefined workload profiles, you may select a workload profile from the table which closely resemble your workload profile requirements. For our scenario, we have selected the OLTP Standard workload name from the Choose workload profile window. We are going to use this workload profile for the LUN placement recommendations. Name - Shows the default profile name. The following restrictions apply to the profile name. The workload profile name must be between 1 to 64 characters. Legal characters are A-Z, a-z, 0-9, -, _, ., and : The first character cannot be - or _.
Chapter 12. Using TotalStorage Productivity Center Performance Manager
493
Description - Shows the description of workload profile. Total I/O per second per GB - Shows the values for the selected workload profile Total I/O per second rate. Average transfer size (KB) - Shows the values for the selected workload profile. Caching information box - Shows the cache hits and destage percentages: Random read cache hits Range from 1 - 100%. The default is 40%. Random write destage Range from 1 - 100%. The default is 33%. Read/Write information box - Shows the read and write values. The percentages for the four fields must equal 100% Sequential reads - The default is 14%. Sequential writes - The default is 23%. Random reads - The default is 36%. Random writes - The default is 32%.
Peak activity information box Since currently we are only viewing properties of an existing profile, the parameters for this box are not selectable. But as reference for subsequent usage, you may review this box. After you review properties for this box, you may click the Close button. While creating new profile, this box will allow you to input following parameters: Use all available performance data radio button. You can select this option if you want to include all available performance data previously collected in consideration for this workload profile. Use the specified peak activity period radio button. You can select this button as an alternate option (instead of using the Use all available performance data option) for consideration in this workload profile definition. Time setting drop down menu. Select from the following options for the time setting you want to use for this workload profile. - Device time - Client time - Server time - GMT Past days to analyze spinner. Use this (or manually enter the number) to select the number of days of historical information you want to consider for this workload profile analysis Time Range drop down lists. Select the Start time and End time to consider using the appropriate fields.
Close button - Click to close the panel. You will be returned to the Choose workload profile window.
494
You can use the Choose candidate locations page to select volume locations for the performance advisor to consider. You can choose to either include or exclude the selected locations for the advisor's consideration. The VPA uses historical performance information to advise you about where volumes should be created. The Choose candidate locations page is one of the panels the performance advisor uses to collect and evaluate the information. Device list - Displays device IDs or names for each ESS on which the task was activated (each ESS on which you dropped the Volume advisor icon). Component Type tree - When you select a device from the Device list, the selection tree opens on the left side of the panel. The ESS component levels are shown in the tree. The following objects might be included: ESS cluster device adapter
Chapter 12. Using TotalStorage Productivity Center Performance Manager
495
The component level names are followed by information about the capacity and the disk utilization of the component level. For example, we used System component level. It shows Component ID - 2105-F20-16603,Type- System, Description 2105-F20-16603-IBM, Available capacity - 311GB, Utilization - Low.(see Figure 12-84 on page 495). Tip: You can select the different ESS component types and the VPA will reconsider the volume placement advise based on that particular select. To familiarize yourself with the options, select each component in turn to determine which component type centric advise you prefer before proceeding to the next step. Select a component type from the tree to display a list of the available volumes for that component in the Candidates table (see Figure 12-84 on page 495). We chose system for this example. It represents entire ESS system in this case. Click Add button to add the component selected in the Candidates table to the Selected candidates table. See Figure 12-85. It shows Selected candidate as 2105-F20-16603.
Figure 12-85 VPA Chose candidate locations Component Type tree example (system)
496
You can use the Verify settings panel to verify the volume settings that you specified in the previous panels of the VPA.
497
Approve recommendations
After you have successfully completed the Verify Settings step, click the Next button, and the Approve Recommendations window opens (see Figure 12-87).
You use the Recommendations window to first view the recommendations from the VPA and then to create new volumes based on the recommendations. In this example, VPA also recommends the location of volume as 16603:2:4:1:1700 in the Component ID column. This means recommended volume location is at ESS with ID 16603, Cluster 2, Device Adapter 4, Array 1 and volume ID1700. With this information, it is also possible to create volume manually via ESS specialist Browser interface or use VPA to create the same. In the Recommendations window of the wizard, you can choose whether the recommendations are to be implemented, and whether to loop around for another set of recommendations. At this time, you have two options (other than to cancel the operation). Make your final selection to Finish or return to the VPA for further recommendations. a. If you do not want to assign the volumes using the current VPA advice, or want the VPA to make another recommendation, check only the Make Additional Recommendations box. 498
IBM TotalStorage Productivity Center V2.3: Getting Started
b. If you want to use the current VPA recommendation and make additional volume assistants at this time, select both the Implement Recommendations and Make Additional Recommendations check boxes. If you choose both options, you must first wait until the current set of volume recommendations are created, or created and assigned, before continuing. If you make this type of selection, a secondary window will appear which runs synchronously within the VPA. Tip: Stay in the same VPA session if you are going to implement volumes and add new volumes. This will enable VPA to provide advice for your current selections, checking for previous assignments, and verifying that no other VPA is processing the same volumes.
499
2. Click the Continue button to proceed with VPA advice (see Figure 12-88 on page 499).
3. In Figure 12-89, we can see that the volumes are being created on the server we selected previously. This process takes a little time, so be patient. 4. Figure 12-90 indicates that the volume creation and assigning to ESS has completed. Be patient and momentarily, the VPA loopback sequence will continue.
500
5. After the volume creation step has successfully completed, the following Settings window will again open so that you may add more volumes (see Figure 12-91).
501
For the additional recommendations, we decided to use same server. But, we specified the Volume quantity range instead of Volume size range for the requested space of 2GB. See Figure 12-92.
502
After clicking Next, the Choose Profile Panels opens. We selected the same profile as before: OLTP Standard. See Figure 12-93.
503
After clicking Next, the Choose candidate locations panel opens. We selected Cluster from the Component Type drop-down list. See Figure 12-94.
The Component Type Cluster shows Component ID as 2105-F20-16603:2, Types as Cluster, Descriptor as 2, Available capacity as 308GB and Utilization as Low. This indicates that VPA plans to provision additional capacity on this Cluster 2 of ESS.
504
After clicking the Add button, Cluster 2 is a selected candidate for new volume. See Figure 12-95.
505
Upon clicking Next, the Verify settings panel opens as shown in Figure 12-96.
506
After verifying settings and clicking Next, VPA recommendations window opens. See Figure 12-97.
507
Since the purpose of this example is to show our readers the VPA looping only, we decided to un-check both check boxes for Implement Recommendations and Make additional recommendations. Clicking Finish completed the VPA example (Figure 12-98).
508
M A N A G I N G P R O F I L E S
C h o o s e C reate p ro file
R es u lts ac c ep ted
S a ve P ro file
Before using VPA for any additional disk space requirement for an application, you will need to: determine typical I/O workload type of that application and; have performance data collected which covers peak load time periods You will need to determine the broad category in selected I/O workload fits in, e.g. whether it is OLTP high, OLTP standard, Data Warehouse, Batch sequential or Document archival. This is shown as highlighted box in the diagram. The TotalStorage Productivity Center for Disk provides pre-defined profiles for these workload types and it allows you to create additional similar profiles by choosing Create like profiles. If do not find any match with pre-defined profiles, then you may prefer to Create a new profile. While choosing Create like or Create profiles, you will also need to specify historical performance data samples covering peak load activity time period. Optionally, you may specify additional I/O parameters. Upon submitting the Create or Create like profile, the performance analysis will performed and results will be displayed. Depending upon the outcome of the results, you may need to re-validate the parameters for data collection task and ensure that peak load samples are taken correctly. If the results are acceptable, you may Save the profile. This profile can be referenced for future usage by VPA. In Choosing workload profiles on page 510, we cover step-by-step tasks using an example.
509
Note: Using a predefined profile does not require pre-existing performance data, but the other two methods require historical performance data from the target storage device. You can launch the workload profiles management tool using the drag and drop method from the IBM Director console GUI interface. Drag the Manage Workload Profile task to the target storage device as shown in Figure 12-101.
510
If you are using Manage Workload Profile or VPA tool for first time of the selected ESS device, then you will need to authorize ESS user validation. This has been described in detail in ESS User Validation on page 485. The ESS User Validation is the same for VPA and Manage Workload Profile tools. After the successful ESS User validation, the Manage Workload Profile panel will be opened as shown in Figure 12-102.
You can create or manage a workload profile using the following three methods: 1. Selecting a predefined workload profile Several predefined workloads are shipped with Performance Manager. You can use the Choose workload profile panel to select the predefined workload profile that most closely matches your storage allocation needs. The default profiles shipped with Performance Manager are shown in Figure 12-103.
You can select the properties panel of the respective pre-defined profile to verify the profile details. A sample profile for OLTP Standard is shown in Figure 12-83 on page 493.
511
2. Creating a workload profile similar to another profile You can use the Create like panel to modify the details of a selected workload profile.You can then save the changes and assign a new name to create a new workload profile from the existing profile. To Create like a particular profile, these are the tasks involved: a. Create a performance data collection task for target storage device: You may need to include multiple storage devices based on your profile requirements for the application. b. Schedule data collection task: You may need to ensure that a data collection task runs over a sufficient period of time, which truly represents a typical I/O load of the respective application. The key is to have sufficient historical data. Tip: Best practice is to schedule frequency of a performance data collection task in such a way that it covers peak load periods of I/O activity and it has at least a few samples of peak loads. The number of samples depends on I/O characteristics of the application. c. Determine the closest workload profile match: Determine whether new workload profile matches w.r.t existing or pre-defined profiles. Note that it may not be the exact fit, but should be of somewhat similar type. d. Create the new similar profile: Using the Manage Workload Profile task, create new profile. You will need to select appropriate time period for historical data, which you have collected earlier. In our example, we created similar profile using Batch Sequential pre-defined profile. First, we select Batch Sequential profile and click Create like button as shown in Figure 12-104.
512
The Properties panel for Batch Sequential is opened, as shown in Figure 12-105.
We changed the following values for our new profile: Name: ITSO_Batch_Daily Description: For ITSO batch applications Average transfer size: 20KB Sequential reads: 65% Random reads: 10% Peak Activity information: We used time period as past 24 days from 12AM to 11PM.
513
514
This new profile - ITSO_Daily_batch is now available in Manage workload profile panel as shown in Figure 12-107. This profile can now be used for VPA analysis. This completes our example.
3. Creating a new workload profile from historical data You can use the Manage workload profile panel to create a workload profile based on historical data about existing volumes. You can select one or more volumes as the base for the new workload profile. You can then assign a name to the workload profile, optionally provide a description, and finally create the new profile. To create a new workload profile, click the Create button as shown in Figure 12-108.
515
This will launch a new panel for Creating workload profile as shown in Figure 12-109. At this stage, you will need to specify the volumes for performance data analysis. In our example, we selected all volumes. For selecting multiple volumes but not all, click the first volume, hold the Shift key and click the last volume in the list. After all the required volumes are selected (shown as dark blue), click the Add button. See Figure 12-109. Note: The ESS volumes you specify should be representative of I/O behavior of the application, for which you are planning to allocate space using the VPA tool.
516
Upon clicking the Add button, all the selected volumes will be moved to the selected volumes box as shown in Figure 12-110.
Figure 12-110 Selected volumes and performance period for new workload profile
In the Peak activity information box, you will need to specify an activity sample period for Volume performance analysis. You can select the option Use all available performance data or select Use the specified peak activity period. Based on your application peak I/O behavior, you may specify the sample period with Start date, Duration in days, and Start / End time. For the time setting, you can choose the drop-down box: Device time, or Client time, or Server time, or GMT
517
After you have entered all the fields, click Next. You will see the Create workload profile Review panel as shown in Figure 12-111.
You can specify a Name for the new workload profile and a Description. You may put in detailed description that covers: The application name for which the profile is being created What application I/O activity is represented by the peak activity sample When it was created Who created it (optional) Any other relevant information your organization requires In our example, we created profile named New_ITSO_app1_profile. At this point you may click Finish. At this point, the TotalStorage Productivity Center for Disk will begin Volume performance analysis based on the parameters you have provided. This process may take some time depending upon number of volumes and sampling time period. Hence, be patient. Finally, it will show the outcome of the analysis.
518
In our example, we got the results notification message as shown in Figure 12-112. Analysis yielded that results are not statistically significant, as shown message: BWN005965E: Analysis results are not significant. This may indicate that: There is not enough I/O activity on selected volumes, OR The time period chosen for sampling is not correct, OR Correct volumes were not chosen You have an option to Save or Discard the profile. We decided to save the profile (Figure 12-113).
Upon saving the profile, it is now listed in the Manage workload profile panel as shown in Figure 12-113.
The new profile can now be referenced by VPA for future usage.
519
520
13
Chapter 13.
521
WWW Server
The Web Server is optional, and handles communications to allow remote Web access to the Server. The WWW Server can run on the same physical server as the Data Manager Server.
522
Novell NetWare and NAS devices do not currently support locally installed Agents - they are managed through an Agent installed on a machine that uses (accesses) the NetWare or NAS device. The Agent will discover information on the volumes or filesystems that are accessible to the Agents host. The Agents are quite lightweight. Agents listen for commands from the host, and then perform a Probe (against the operating system), and/or a Scan (against selected filesystems). Normal operations might see one scheduled Scan per day or week, plus various ad hoc Scans. Scans and Probes are discussed later in this chapter.
523
524
Figure 13-2 shows the TotalStorage Productivity Center for Data dashboard. This is the default right-hand pane display when you start TotalStorage Productivity Center for Data and shows a quick summary of the overall health of the storage environment. It can quickly show you potential problem areas for further investigation.
The dashboard contains four viewable areas, which cycle among seven pre-defined sets of panels. To cycle, use the Cycle Panels button. Use the Refresh button to update the display.
Enterprise-wide summary
The Enterprise-wide Summary panel shows statistics accumulated from all the Agents. The statistics are: Total filesystem capacity available Total filesystem capacity used Total filesystem free capacity Total allocated and unallocated disk space Total disk space unallocated to filesystems Total LUN capacity Total usable LUN capacity Total number of monitored servers Total number of unmonitored servers Total number of storage subsystems Total number of users Total number of disks Total number of LUNs Total number of filesystems Total number of directories Total number of files
525
Alerts Pending
This panel shows active Alerts that have been triggered but are still pending.
Pings
A Ping is a standard ICMP Ping which checks registered Agents for availability. If an Agent does not respond to a Ping (or a pre-defined number of Pings) you can set up an Alert to take some action. The actions could be one, any, or all of: SNMP trap TEC Event Notification at login Entry in the Windows event log Run a script Send e-mail to a specified user(s)
526
Pings are used to generate Availability Reports, which list the percentage of times a computer has responded to the Ping. An example of an Availability Report for Ping is shown in Figure 13-3. Availability Reports are discussed in detail in 13.11.3, Availability Reporting on page 604.
Probes
Probes are used to gather information about the assets and system resources of monitored servers, such as processor count and speed, memory size, disk count and size, filesystems, etc. The data collected by the Probe process is used in the Asset Reports described in 13.11.1, Asset Reporting on page 595. Figure 13-4 shows an Asset report for detected disks.
527
Scans
The Scan process is used to gather statistics about usage and trends of the server storage. Data collected by the Scan jobs are tailored by Profiles. Results of Scan jobs are stored in the enterprise repository. This data supplies the data for the Capacity, Usage, Usage Violations, and Backup Reporting functions. These reports can be scheduled to run regularly, or they can be run ad hoc by the administrator.
Profiles limit the scanning according to the parameters specified in the Profile. Profiles are
used in Scan jobs to specify what file patterns will be scanned, what attributes will be gathered, what summary view will be available in reports and the retention period for the statistics. TotalStorage Productivity Center for Data supplies a number of default Profiles which can be used, or additional Profiles can be defined. Table 13-1 on page 547 shows the default Profiles provided. Some of these include: Largest files - Gathers statistics on the largest files Largest directories - Gathers statistics on the largest directories Most at risk - Gathers statistics on the files that have been modified the longest time ago and have not been backed up since modified (Windows Agents only)
528
Figure 13-6 shows a sample of a report produced from data collected in Scans.
Figure 13-6 Summary View - by filesystem, disk space used and disk space free
This report shows a list of the filesystems on each Agent, the amount of space used in each, expressed in bytes and as a percentage, the amount of free space, and the total capacity available in the filesystem.
13.2.3 Reporting
Reporting in TotalStorage Productivity Center for Data is very powerful, with over 300 pre-defined views, and the capability to customize those standard views, save the custom report, and add it to your menu for scheduled or ad hoc reports. You can also create your own individual reports according to particular needs and set them to run as needed, or in batch (regularly). Reports can be produced in table format for a variety of charting (graph) views. You can export reports to CSV or HTML formats for external usage. Reports are generated against data already in the repository. A common practice is to schedule Scans and Probes just before running reports. Reporting can be done at almost any level in the system, from the enterprise down to a specific entity and any level in between. Figure 13-6 shows a high-level summary report. Or, you can drill down to something very specific. Figure 13-7 is an example of a lower-level report, where the administrator has focused on a particular Agent, KANAGA, to look at a particular disk on a particular controller.
529
Reports can be produced either system-wide or grouped into views, such as by computer, or OS type. Restriction: Currently, there is a maximum of 32,767 (216 -1) rows per report. Therefore, you cannot produce a report to list all the .HTM files in a directory containing a million files. However, you can (and it would be more productive to do so) produce a report of the 20 largest files in the directory, or the 20 oldest files, for example.
TotalStorage Productivity Center for Data allows you to group information about similar entities (disk, filesystems, etc.) from different servers or business units into a summary report, so that business and technology administrators can manage an enterprise infrastructure. Or, you can summarize information from a specific server - the flexibility and choice of configuration is entirely up to the administrator. You can report as at a point in time, or produce a historical report, showing storage growth trends over time. Reporting lets you track actual demand for disk over time, and then use this information to forecast future demand for the next quarter, two quarters, year, etc. Figure 13-8 is an example of a historical report, showing a graph of the number of files on the C drive on the Agent KANAGA.
530
TotalStorage Productivity Center for Data has three basic types of reports: Computers and filesystems Databases Chargeback
Reporting categories
Major reporting categories for filesystems and databases are: Asset Reporting uses the data collected Probes to build a hardware inventory of the storage assets. You can then navigate through a hierarchical view of the assets by drilling down through computers, controllers, disks, filesystems, directories, and exports. For database reporting, information on instances, databases, tables, and data files is presented for reporting. Storage Subsystems Reporting provides information showing storage capacity at a computer, filesystem, storage subsystem, LUN, and disk level. These reports also enable you to view the relationships among the components of a storage subsystem. For a list of supported devices, see <table>. Availability Reporting shows responses to Ping jobs, as well as computer uptime. Capacity Reporting shows how much storage capacity is installed, how much of the installed capacity is being used, and how much is available for future growth. Reporting is done by disk and filesystem, and for databases, by database. Usage Reporting shows the usage and growth of storage consumption, grouped by filesystem, and computers, individual users, or enterprise-wide. Usage Violation Reporting shows violations to the corporate storage usage policies, as defined through TotalStorage Productivity Center for Data. Violations are either of Quota (defining how much storage a user or group of users is allowed) or Constraint (defining which file types, owners and file sizes are allowed on a computer or storage entity). You can define what action should be taken when a violation is detected - for example, SNMP trap, e-mail, or running a user-written script. Backup Reporting identifies files which are at risk because they have not been backed up.
Chapter 13. Using TotalStorage Productivity Center for Data
531
Figure 13-9 TotalStorage Productivity Center for Data Reports on the Web
13.2.4 Alerts
An Alert defines an action to be performed if a particular event occurs or condition is found. Alerts can be set on physical objects (computers and disks) or a logical objects (filesystems, directories, users, databases, and OS user groups). Alerts can tell you, for instance, if a disk has a lot of recent defects, or if a filesystem or database is approaching capacity. Alerts on computers and disks come from the output of Probe jobs and are generated for each object that meets the triggering condition. If you have specified a triggered action (running a script, sending an e-mail, etc.) then that action will be performed if the condition is met. Alerts on filesystems, directories, users, and OS user groups come from the combined output of a Probe and a Scan. Again, if you have specified an action, that action will be performed if the condition is met. An Alert will register in the Alert log, plus you can also define one, some, or all of the following actions to be performed in addition: Send an e-mail indicating the nature of the Alert. Run a specific script with relevant parameters supplied from the content of the Alert. Make an entry into the Windows event log. Pop up next time the user logs in to TotalStorage Productivity Center for Data. Send an SNMP trap. Log a TEC event Refer to 13.4, OS Alerts on page 555 for details on alerts. 532
IBM TotalStorage Productivity Center V2.3: Getting Started
Chargeback is a very powerful tool for raising the awareness within the organization of the cost of storage, and the need to have the appropriate tools and processes in place to manage storage effectively and efficiently. Refer to 13.17, Charging for storage usage on page 700 for more details on Chargebacks.
13.3 OS Monitoring
The Monitoring features of TotalStorage Productivity Center for Data enable you to run regularly scheduled or ad hoc data collection jobs. These jobs gather statistics about the storage assets and their availability and their usage within your enterprise, and make the collected data available for reporting. This section gives a quick overview of the monitoring jobs, and explains how they work through practical examples. Reporting on the collected data is explained in Data Manager reporting capabilities on page 592.
533
Except for Discovery, you can create multiple definitions for each of those monitoring features of TotalStorage Productivity Center for Data. To create a new definition, right-click the feature and select Create <feature>. Figure 13-11 shows how to create a new Scan job.
534
Once saved, any definition within TotalStorage Productivity Center for Data can be updated by clicking the object. This will put you in Edit mode. Save your changes by clicking the floppy disk icon in the top menu bar. Discovery, Pings, Probes, and Scan menus contain jobs that can run on a scheduled basis or ad hoc. To execute a job immediately, right-click the job then select Run Now (see Figure 13-12). Each execution of a job creates a time-stamped output that can be displayed by expanding the tree under the job (you may need to right-click the job and select Refresh Job List).
The color of the job output represents the job status: Green - Successful run Brown - Warnings occurred during the run Red - Errors occurred during the run Blue - Running jobs To view the output of a job, double click the job. Groups and Profiles are definitions that may be used by other jobs - they do not produce an output in themselves. As shown in Figure 13-12, all objects created within Data Manager are prefixed with the user ID of the creator. Default definitions, created during product installation, are prefixed with TPCUser.Default. Groups, Discovery, Probes, Scans, and Profiles are explained in the following sections.
13.3.2 Groups
Before defining monitoring and management jobs, it may be useful to group your resources so you can limit the scope of monitoring or data collection.
535
Computer Groups
Computer Groups allow you to target management jobs on specific computers based on your own criteria. Some criteria you might consider for grouping computers are platform type, application type, database type, and environment type (for example, test or production). Our lab environment contains Windows 2000 servers. In order to target specific servers for monitoring based on OS and/or database type, we will defined the following groups: Windows Systems Windows DB Systems To create the first group, expand Data Manager Monitoring Groups Computer, right-click Computer and select Create Computer Group. Our first group will contain all Windows systems as shown in Figure 13-13. To add or remove a host from the group, highlight it in either the Available or Current Selections panel and use the arrow buttons. You can also enter a meaningful description in the field.
To save the new Group, click the floppy disk icon in the menu bar, and enter the Group name in the confirmation box shown in Figure 13-14.
536
We created the other group using the same process, and named it Windows DB Systems. Important: To avoid redundant data collection, a computer can belong to only one Group at a time. If you add a system which is already in a Group, to a second Group, it will automatically be removed from the first Group.
Figure 13-15 shows the final Group configuration, with the members of the Windows Systems group.
Note: The default group TPCUser.Default Computer Group contains all servers that have been discovered, but not yet assigned to a Group.
Filesystem Groups
Filesystem Groups are used to associate together filesystems from different computers that
have some commonality. You can then use this group definition to focus the Scan and the Alert processes to those filesystems. To create a Filesystem Group, you have to select explicitly each filesystem for each computer you want to include in the group. There is no way to do a grouped selection, e.g. / (root) filesystem for all UNIX servers or C:\ for all Windows platforms. Note: As for computers, a filesystem can belong to only one Group.
537
Directory Groups
Use Directory Groups to group together directories to which you want to apply the same storage management rules. Figure 13-16 shows the Directory Group definition screen by going to Data Manager Monitoring Groups Directory and right-clicking Directory and selecting Create Directory Group.
The Directory Group definition has two views for directory selection: Use directories by computer to specify several directories for one computer. Use computers by directory to specify one directory for several computers. The button on the bottom of the screen toggles between New computer and New directory depending on the view you select.
538
We will define one Directory Group with a DB2 directory for a specific computer (Colorado). To define the Group: 1. Select directories by computer. 2. Click New computer. 3. Select colorado from the pull-down Computer field. 4. Enter C:\DB2\NODE0000 in the Directories field and click Add (see Figure 13-17).
5. Click OK. 6. Save the group as DB2 Node. Figure 13-18 shows our final Groups configuration and details of the OracleArchive Group.
539
User Groups
You can define Groups made up of selected user IDs. These groupings will enable you to easily define and focus storage management rules such as scanning and Constraints on the defined IDs. Note: You can include in a User Group only user IDs defined on the discovered hosts, which have files belonging to them.
Note: As with users, an OS User Group can belong to only one Group at a time.
13.3.3 Discovery
The Discovery process is used to discover new computers within your enterprise that have not yet been monitored by Data Manager. The discovery process will: Request a list of Windows systems from the Windows Domain Controller Contact, through SNMP, all NAS filers and check if they are registered in the nas.config file Discover all NetWare servers in the NetWare trees reported by Agents Search UNIX Agents mount tables, looking for remote filesystems and discover NAS filers More details of NAS and NetWare discovery are given in the manual IBM Tivoli Storage Resource Manager: A Practical Introduction, SG24-6886. Use the path Data Manager Monitoring Discovery to change the settings of the Discovery job. The following options are available.
540
Alert tab
The second tab, Alert, enables you to be notified when a new computer is discovered. See 13.4, OS Alerts on page 555 for more details on the Alerting process.
Options tab
The third tab, Options (Figure 13-20) sets the discovery runtime properties.
Figure 13-20 Discovery job options Chapter 13. Using TotalStorage Productivity Center for Data
541
Uncheck the Skip Workstations field if you want to discover the Windows workstations reported by the Windows Domain Controller.
13.3.4 Pings
The Ping process will: launch TCP/IP pings against monitored computers generate statistics on computer Availability in the central repository generate an Alert if the process fails because of an unavailable host summarizes the Ping process. Pings gather statistics about the availability of monitored servers. The scheduled job will Ping your servers and consider them active if it gets an answer. This is purely ICMP-protocol based - there is no measurement of individual application availability. When you create a new Ping job, you can set the following options.
Computers tab
Figure 13-21 shows the Computers tab, which is used to limit the scope of the computers that are to be Pinged.
542
Options tab
On the Options tab, you specify how often the Ping statistics are saved in the database repository. By default, TotalStorage Productivity Center for Data keeps its Ping statistics in memory for eight Pings before flushing them to the database and calculating an average availability. You can change the flushing interval to another time amount, or a number of Pings (for example, to calculate availability after every 10 Pings). The system availability is calculated as:
(Count of successful pings) / (Count of pings)
A lower interval can increase database size, but gives you more accuracy on the availability history. We selected to save to the database at each Ping (), which means we will have an availability of 100% or of 0%, but we have a more granular view of the availability of our servers (Figure 13-23).
543
Alert tab
The Alert tab (shown in Figure 13-24) is used to generate Alerts for each host that is unavailable. Alert mechanisms are explained in more detail in 13.4, OS Alerts on page 555. You can choose any Alert type from the following: SNMP trap to send a trap to the Event manager defined in Administrative services Configuration General Alert Disposition TEC Event to send an event to a Tivoli Enterprise Console Login Notification to direct the Alert to the specified user in the Alert Log (see 13.4, OS Alerts on page 555) Windows Event Log to generate an event to the Windows event log Run Script to run a script on the specified server Email to send a mail to the specified user through the Mail server defined in Administrative services Configuration General Alert Disposition
544
We selected to run a script that will send popup messages to selected administrators. The script is listed in Example 13-1. Optimally, you would send an event to a central console such as the Tivoli Enterprise Console. Note that certain parameters are passed to the script - more information is given in Alerts tab on page 560.
Example 13-1 Script PINGFAILED.BAT net send /DOMAIN:Colorado Computer %1 did not respond to last %2 ping(s). Please check it
We then saved the Ping job as PingHosts, and tested it by right-clicking and selecting Run now. As the hosts did not respond, we received notifications as shown in Figure 13-25.
More details about the related reporting features of TotalStorage Productivity Center for DataTotalStorage Productivity Center for Data are in 13.11.3, Availability Reporting on page 604.
13.3.5 Probes
The Probe process will: Gather Assets data on monitored computers Store data in the central repository Generate an Alert if the process fails
545
The Probe process gathers data about the assets and system resources of Agents such as: Memory size Processor count and speed Hard disks Filesystems The data collected by the Probe process is used by the Asset Reports described in 13.11.1, Asset Reporting on page 595.
Computers tab
Figure 13-26 shows that we included the TPCUser.Default Computer Group in the Probe so that all computers, including those not yet assigned to an existing Group, will be Probed. We saved the Probe as ProbeHosts.
Important: Only the filesystems that have been returned by a Probe job will be available for further use by Scan, Alerts, and policy management within TotalStorage Productivity Center for Data.
Alert tab
As this is not a business-critical process, we asked to be alerted by mail for any failed Probe. Figure 13-27 shows the default mail text configuration for a Probe failure.
546
13.3.6 Profiles
Profiles are used: In Scan jobs To limit files to be scanned To specify files tabulates to be scanned To select the summary view Directories and filesystems User ids OS user groups To set statistics retention period TotalStorage Productivity Center for Disk provides default profiles that provide data for all the default reports. Profiles are used in Scan jobs to specify: The pattern of files to be scanned The attributes of files to be gathered The summary view that will be available in reports The statistics retention period Specifying correct profiles avoids gathering unnecessary information that may lead to space problems within the repository. However, you will not be able to report on or check Quotas on files that are not used by the Profile. Data Manager comes with several default profiles, (shown in Table 13-1) prefixed with TPCUser, which can be reused in any Scan jobs you define.
Table 13-1 Default profile Default profile name BY_ACCESS BY_CREATION Description Gathers statistics by length of time since last access of files Gathers statistics by length of time since creation of files
547
Default profile name BY_MOD_NOT_BACKED_UP BY_MODIFICATION FILE_SIZE_DISTRIBUTION LARGEST_DIRECTORIES LARGEST_FILES LARGEST_ORPHANS MOST_AT_RISK
Description Gathers statistics by length of time since last modification (only for files not backed up since modification). Windows only Gathers statistics by length of time since last modification of files Gathers file size distribution statistics Gathers statistics on the n largest directories. (20 is the default amount.) Gathers statistics on the n largest files. (20 is the default amount.) Gathers statistics on the n largest orphan files. (20 is the default amount.) Gathers statistics on the n files that have been modified the longest time ago and have not yet been backed up since they were modified. Windows only. (20 is the default amount.) Gathers statistics on the n oldest orphan files. (20 is the default amount.) Gathers statistics on the n most obsolete files (i.e., files that have not been accessed or modified for the longest period of time). (20 is the default amount.) Summarizes space usage by file extension Summarizes space usage by Filesystem or Directory Summarizes space usage by OS Group Summarizes space usage by Owner Gathers statistics on network-wide space consumed by temporary files Gathers statistics on non-OS files not accessed in the last year and orphaned files
OLDEST_ORPHANS MOST_OBSOLETE_FILES
Those default profiles, when set in a Scan job, gather data needed for all the default Data Manager reports. As an example, we will define an additional Profile to limit a Scan job to the 500 largest Postscript or PDF files unused in the last six months. We also want to keep weekly statistics at a filesystem and directory level for two weeks.
Statistics tab
On the Statistics tab (shown in Figure 13-28), we specified: Retain filesystem summary for two weeks Gather data based on creation data Select the 500 largest files The Statistics tab is used to specify the type of data that is gathered, and has a direct impact on the type of reports that will be available. In our specific case, the Scan associated with this profile will not create data for reports based on user IDs and users groups. Neither will it create data for reports on directory size.
548
The Summarize space usage by section of the Statistics tab specifies how the space usage data must be summarized. If no summary level is checked, the data will not be summarized, and therefore will not be available for reporting in the corresponding level of Usage Reporting section of TotalStorage Productivity Center for Data. In our particular case, because we select to summarize by filesystem and directory, we will see space used by PDF and Postscript files at those levels, providing we set up the Scan profile correctly. See 13.3.7, Scans on page 552 for information on this. We will not see which users or groups have allocated those PDF and Postscript files. Restriction: For Windows servers, users and groups statistics will not be created for FAT filesystems. The Accumulate history section sets the retention period of the collected data. In this case, we will see a weekly summary for the last two weeks. The Gather statistics by length of time since section sets the base date used to calculate the file load. It determines if data will be gathered and summarized for the Data Manager Reporting Usage Files reporting view. The Gather information on the section sets the amount of files to retrieve for each of the report views available under Data Manager Reporting Usage Access Load.
549
With the New Condition menu, you can create a single filter on the files while the New Group enables you to combine several conditions with: All of Any of None of Not all of The file is selected if all conditions are met (AND) The file is selected if at least one condition is met (OR) The file is NOT selected if at least one condition is met (NOT OR) The file is selected if none of the conditions are met (NOT AND)
The Condition Group can contain individual conditions or other condition groups. Each individual condition will filter files based on one of the listed items: Name Last access time Last modified Creation time Owner user ID Owner group Windows files attributes Size Type Length We want to select files that meet our conditions: (name is *.ps or name is *.pdf) and unused since six months. The AND between our two conditions will be translated to All of, while the OR within our first condition will be translated to Any of. On the screen shown in Figure 13-29, we selected New Group. From the popup screen, Figure 13-30, we selected All of and clicked OK.
550
Now, within our All of group we will create one dependant Any of group using the same sequence. The result is shown in Figure 13-31.
Now, we create individual conditions within each group by right-clicking New Condition on the group where the conditions must be created. Figure 13-32 shows the creation of our first condition for the Any of group. We enter in our file specifications (*.ps and *.pdf) here.
551
We repeated the operation for the second condition (All of). The final result is shown in Figure 13-33.
The bottom of the right pane shows the textual form of the created condition. You can see that it corresponds to our initial condition. We saved the profile as PS_PDF_FILES (Figure 13-34).
13.3.7 Scans
The Scan process is used to gather data about files and to summarize Usage statistics as specified in the associated profiles. It is mandatory for Quotas and Constraints management. The Scan process gathers statistics about the usage and trends of the server storage. Scan job results are stored in the repository and supply the data necessary for the Capacity, Usage, Usage Violations, and Backup Reporting facilities. To create a new Scan job, Data Manager Monitoring Scans, right-click and select Create Scan. The scope of each Scan job is set by five different tabs on the right pane.
Filesystems tab
You can specify a specific filesystem for one computer, a filesystem Group (see Filesystem Groups on page 537) or all filesystems for a specific computer. Only the filesystems you have selected will be scanned. Figure 13-35 shows how to configure the Scan to gather data on all our servers. Note: Only filesystems found by the Probe process will be available for Scan. 552
IBM TotalStorage Productivity Center V2.3: Getting Started
Profiles tab
As explained in 13.3.6, Profiles on page 547, the Profiles are used to select the files that are scanned for information gathering. A Scan job scans and gathers data only for files that are scoped by selected Profiles. You can specify Profiles at two levels: Filesystems: All selected filesystems will be scanned and data summarized for each filesystem. Directory: All selected directories (if included in the filesystem) will be scanned and data summarized for each directory.
553
Figure 13-36 shows how to configure a Scan to have data summarized at both the filesystem and directory level. .
Alert tab
You can be alerted through mail, script, Windows Event Log, SNMP trap, TEC event, or Login notification if the Scan job fails. The Scan job may fail if an Agent is unreachable. Click the floppy icon to save your new Scan job, shown in Figure 13-37.
554
x x
x -
x x
13.4 OS Alerts
TotalStorage Productivity Center for Data enables you to define Alerts on computers, filesystems, and directories. Once the Alerts are defined, it will monitor the results of the Probe and Scan jobs, and will trigger an Alert when the threshold or the condition is met. TotalStorage Productivity Center for Data provides a number options for Alert mechanisms from which you can choose depending on the severity you assign to the Alert. Depending on the severity of the triggered event or the functions available in your environment, you may want to be alerted with:
555
An SNMP trap to an event manager. Figure 13-38 shows a Filesystem space low Alert as displayed in our SNMP application, IBM Tivoli NetView. Defining the event manager is explained in 8.5, Alert Disposition on page 316.
A TEC (Tivoli Enterprise Console) event. An entry in the Alert Log (see Figure 13-39). You can configure Data Manager, so that the Alert Log will be automatically displayed when you log on to the GUI by using Preferences Edit General (see Figure 13-40).
556
An entry in the Windows Event log, as shown in Figure 13-41. This is useful for lower severity alerts or when you are monitoring your Windows event logs with an automated tool such as IBM Tivoli Distributed Monitoring.
Running a specified script - The script runs on the specified computer with the authority of the Agent (root or Administrator). See 13.5.5, Scheduled Actions on page 582 for special considerations with scripts execution. An e-mail - TotalStorage Productivity Center for Data must be configured with a valid SMTP server and port as explained in 8.5, Alert Disposition on page 316.
557
558
Except for the Alert Log, you can create multiple definitions for each of those Alert features of TotalStorage Productivity Center for Data. To create a new definition, right-click the feature and select Create <feature>. Figure 13-43 shows how to create a new Filesystem Alert.
559
Alerts tab
The Alerts tab contains two parts:
Triggering condition to specify the computer component you want to be monitored. You can monitor a computer for:
RAM increased RAM decreased Virtual Memory increased Virtual Memory decreased New disk detected Disk not found New disk defect found Total disk defects exceed. You will have to specify a threshold. Disk failure predicted New filesystem detected
Information about disk failures is gathered through commands against disks with the following exceptions: IDE disks do support only Disk failure predicted queries AIX SCSI disks do not support failures and predicted failures queries
Triggered action where you specify the action that must be executed. If you choose to run
a script, it will receive several positional parameters that depends on the triggering condition. The parameters display on the Specify Script panel, which is accessed by checking Run Script and clicking the Define button. 560
IBM TotalStorage Productivity Center V2.3: Getting Started
Figure 13-45 shows the parameters passed to the script for a RAM decreased condition.
Figure 13-46 shows the parameters passed to the script for a Disk not found condition.
Computers tab
This limits the Alert process to specific computers or computer Groups (Figure 13-47).
Figure 13-47 Computer alerts - Computers tab Chapter 13. Using TotalStorage Productivity Center for Data
561
Alerts tab
As for Computer Alerts, the Alerts tab contains two parts. In the Triggering condition section you can specify to be alerted if a: Filesystem is not found, which means the filesystem was not mounted during the most recent Probe or Scan. Filesystem is reconfigured. Filesystem free space is less than a threshold specified in percent, KB, MB, or GB. Free UNIX filesystem inode count is less than a threshold (either percent or inodes count).
562
You can choose to run a script (click the Define button next to Run Script), or you can also change the content of the default generated mail by clicking Edit Email. You will see a popup with the default mail skeleton which is editable. Figure 13-49 shows the default e-mail message.
Alerts tab
Directory Alerts configuration is similar to Filesystem alerts. The supported triggers are: Directory not found Directory consumes more than the specified threshold set in percent, KB, MB or GB.
Directories tab
Since Probe jobs do not report on directories and Scan jobs report only on directories. if a directory Profile has been assigned (See Putting it all together on page 555) you can only choose to be alerted for any directory that has already been included in a Scan and actually scanned.
563
There are nine different views. Each of them will show only the Alerts related to the selected view except: All view - Shows all Alerts Alerts Directed to <logged user> - Shows all Alerts where the current logged user has been specified in the Login notification field When you click the icon on the left of a listed Alert, you will see detailed information on the selected Alert as shown in Figure 13-51.
564
13.5.1 Quotas
Quotas can be set at either a user or at an OS User Group level. For the OS User Group level, this could be either an OS User Group, (see OS User Group Groups on page 540), or a standard OS group (such as system on UNIX, or Administrators on Windows). The User Quotas trigger an action when one of the monitored users has reached the limit while the OS User group Quotas trigger the action when the sum of space used by all users of monitored groups has reached the limit. The Quotas definition mechanism is the same for both except for the following differences: The menu tree to use: Data Manager Policy Management Quotas User Data Manager Policy Management Quotas OS User group
565
The monitored elements you can specify: User and user groups for User Quotas OS User Group and OS User Group Groups for OS User Group Quota We will show how to configure User Quotas. User Group Quotas are configured similarly. Note that the Quota enforcement is soft - that is, users are not automatically prevented from exceeding their defined Quota, but the defined actions will trigger if that happens. There are three sub-entries for Quotas: Network Quotas, Computer Quotas, and Filesystem Quotas
Network Quotas
A Network Quota defines the maximum cumulated space a user can occupy on all the scanned servers. An Alert will be triggered for each user that exceeds the limit specified in the Quota definition. Use Data Manager Policy Management Quotas User Network, right-click and select Create Quota to create a new Quota. The right pane displays the Quota configuration screen with four tabs.
Users tab
Figure 13-52 shows the Users tab for Network Quotas.
From the Available column, select any user ID or OS User Group you want to monitor for space usage. The Profile pull-down menu is used to specify the file types that will be subject to the Quota. The list will display all Profiles that create summaries by user (by file owner). Select the Profile you want to use from the pull-down. The default Profile Summary by Owner collects information about all files and summarizes them on the user level. The ALLGIFFILES profile collects information about GIF files and creates a summary at a user level as displayed in Figure 13-53. This (non-default) profile was created using the process shown in 13.3.6, Profiles on page 547.
566
Using this profile option, we can define general Quotas for all files and more restrictive Quotas for some multimedia files such as GIF and MP3.
Filesystem tab
On the Filesystem tab shown Figure 13-54, select the filesystems or computers you want to be included in the space usage for Quota management.
In this configuration, for each user, his cumulated space usage on all servers will be calculated and checked against the Quota limit.
Chapter 13. Using TotalStorage Productivity Center for Data
567
When to check
The Quota management is based on the output of the Scan jobs. Therefore, each Quota definition must be scheduled to run after the Scan jobs that collect the adequate information. The When to CHECK tab is standard, and allows you to define a one off or a recurring job.
Alert tab
On the Alert tab, specify the Quota limit in: KB, MB or GB, and the action to run when the Quota is exceeded.
You can choose from the standard Alerts type available with TotalStorage Productivity Center for Data. Each Alert will be fired once for each user exceeding their Quota. We have selected to run a script that we wrote, QUOTAUSERNET.BAT, listed in Example 13-2.
Example 13-2 QUOTAUSERNET.BAT script echo NETWORK quota exceeded - %1 %2 uses %3 - Limit set to %4 >>quotausernet.txt
The Alert has fired for user root and Administrators. This clearly shows that administrative users such as root and Administrators should not normally be included in standard Quotas monitoring.
568
Computer Quotas
Computer Quotas enable you to fire Alerts when a user exceeds their space Quota on a specific computer as shown in Figure 13-56. Multiple Alerts are generated if a user violates the Quota on separate computers.
Filesystem Quotas
A Filesystem Quota defines a space usage limit at the filesystem level. An Alert will be fired for each filesystem where a user exceeds the limit specified in the Quota definition. Use Data Manager Policy Management Quota User Filesystem, right-click, and select Create Quota to create a new Quota. After setting up and running a Quota for selected filesystems, we received the following entries in the Alert History, shown in Figure 13-57.
569
13.5.3 Constraints
The main features of Constraints are listed in Figure 13-58.
Constraints
Reports and triggers actions based on specific files which use too much space on monitored servers Files can be selected based on
server and filesystem name pattern (eg: *.mp3, *.avi) owner age size attributes
Actions triggered through standard Alerting mechanism when total space used by files exceeds a threshold
ibm.com/redbooks
Constraints are used to generate Alerts when files matching specified criteria are consuming too much space on the monitored servers.
570
Constraints provide a deeper level of Data Management. Quotas will allow reporting on users who have exceeded their space limitations. With Constraints, we can get more detailed information to specify limits on particular file types or other attributes, such as owner, age, and so on. The output of a Constraint when applied to a Scan will return a list of the files that are consuming too much space. Note: Unlike Quotas, Constraints are automatically checked during Scan jobs and do not need to be scheduled. Also, the Scan does not need to be associated with Profiles that will cause data to be stored for reporting.
Filesystems tab
The Filesystems tab helps you select the computers and filesystems you want to check for the current Constraint. The selection method for computers and filesystems is the same as for Scan jobs (see 13.3.7, Scans on page 552).
Use the buttons on the top of the screen, to allow or forbid files depending on their name. The left column shows some default file patterns, or you can use the bottom field to create your own pattern. Click >> to add your pattern to the allowed/forbidden files.
571
Users tab
The Users tab (shown in Figure 13-60) is used to allow or restrict the selected users in the Constraint.
Important: The file condition is logically ORed with the User condition. A file will be selected for Constraint processing if it meets at least one of the conditions.
Options tab
The Options tab provides additional conditions for file selection, and limits the number of selected files to store in the central repository. Once again, the conditions added in the tab will be logically ORed with the previous set in the File Types and Users tab.
572
The bottom part of the tab, shown in Figure 13-61, contains the textual form of the Condition, taking into account all the entries made in the Filesystems, File Types, Users and Options tabs.
You can change this condition or add additional conditions, by using the Edit Filter button. It displays the file filter popup (Figure 13-62) to change, add, and remove conditions or conditions groups as previously explained in 13.3.6, Profiles on page 547.
573
We changed the file filter to a more appropriate one by changing the OR operator to AND (Figure 13-63).
Alert tab
After selecting the files, you may want to generate an Alert only if the total used space meeting the Constraint conditions exceeds a predefined limit. Use the Alert tab to specify the triggering condition and action (Figure 13-64).
574
In our Constraint definition, a script is triggered for each filesystem where the selected files exceed one Gigabyte. We select the script by checking the Run Script option and selecting Change... as shown in Figure 13-65. The script will be passed several parameters including a path to a file that contains the list of files meeting the Constraint. You can use this list to execute any action including delete or archive commands.
Our example uses a sample script (tsm_arch_del.vbs) which is shipped with TotalStorage Productivity Center for Data, which archives all the files in the produced list to a Tivoli Storage Manager server, and then deletes them from local storage. This script is installed with TotalStorage Productivity Center for Data server, and stored in the scripts subdirectory of the server installation. It can be edited or customized if required - we recommend that you save the original files first. Versions for Windows (tsm_arch_del.vbs) and UNIX (tsm_arch_del) are provided. If you will run this Constraint on a UNIX agent, then PERL is required to be installed on the agent. A Tivoli Storage Manager server must be available and configured for this script to work. For more information on the sample scripts, see Appendix A of the IBM Tivoli Storage Resource Manager Users Guide, SC32-9069.
575
Filesystem Extension
Automates filesystem extension Supported platforms
AIX using JFS SUN using VxFS
Support for automatic LUN provisioning with IBM ESS Storage Subsystem Actions triggered through standard Alerting mechanism when a filesystem is performed
ibm.com/redbooks
576
We use filesystem extension policy to automatically extend filesystems when utilization reaches a specified threshold. We can also enable LUN provisioning to extend filesystems within an ESS. To set up filesystem extension policy, select Data Manager Policy Management Filesystem Extension. Right-click Filesystem Extension and select Create Filesystem Extension Rules as seen in Figure 13-67.
577
In the Filesystems tab, select the filesystems which will use filesystem extension policy by moving them to the Current Selections panel. Note the Enabled checkbox - the default is to check it, meaning the rule will be active. If you uncheck the box, it will toggle to Disabled - you can still save the rule, but the job will not run. To specify the extension parameters, select the Extension tab (Figure 13-68).
This tab specifies how a filesystem will be extended. An explanation of the fields is provided below.
Amount to Extend
We have the following options: Add - the amount of space used for extension in MB or GB, or as a percentage of filesystem capacity. Make Freespace - the amount of freespace that will be maintained in the filesystems by this policy. If freespace falls below the amount that is specified, the difference will be added. Freespace can be specified in MB or GB increments, or by a percentage of filesystem capacity. Make Capacity - the total capacity that will be maintained in the selected filesystems. If the capacity falls below the amount specified, the difference will be added.
578
579
LUN Provisioning is an optional feature for filesystem extension. When the Enable Automatic LUN Provisioning is selected, LUN provisioning is enabled. In the Create LUNs that are at least field, you can specify a minimum size for new LUNs. If you select this option, LUNs of at least the size specified will be created. If no size is specified, then the Amount to Extend option specified for the filesystem (in Amount to Extend on page 578) will be used. For more information on LUN provisioning, see IBM Tivoli Storage Resource Manager 1.2 Users Guide. The Model for New LUNs feature means that new LUNs will be created similar to existing LUNs in your setup. At least one ESS LUN must be currently assigned to the TotalStorage Productivity Center for Data Agent associated with the filesystem you want to extend. There are two options for LUN modeling: Model new LUNs on others in the volume group of the filesystem being extended provisioned LUNS are modeled on existing LUNs in the extended filesystems volume group. Model new LUNs on others on the same host as the filesystem being extended provisioned LUNS are modeled on existing LUNs in the extended filesystems volume group. If the corresponding LUN model cannot satisfy the requirements. it will look for other LUNs on the same host. The LUN Source option defines the location of the new LUN in the ESS, and has two options: Same Storage Pool - provisioned LUNs will be created using space in an existing Storage Pool. In ESS terminology this is called the Logical Sub System or LSS. Same Storage Subsystem - provisioned LUNs can be created in any Storage Pool or ESS LSS. The When to Enforce Policy tab (Figure 13-70) specifies when to apply the filesystem extension policy to the selected filesystems.
580
The options are: Enforce Policy after every Probe or Scan automatically enforces the policy after every Probe or Scan job. The policy will stay in effect until you either change this setting or disable the policy. Enforce Policy Now enforces the policy immediately for a single instance. Enforce Policy Once at enforces the policy once at the specified time, specifying the month, day, year, hour, minute, and AM/PM The Alert tab (Figure 13-71) can define an Alert that will be triggered by the filesystem extension job.
581
Currently the only available condition is A filesystem extension action started automatically. Refer to Alert tab on page 544 for an explanation of the definitions. Important: After making configuration changes to any of the above filesystem extension options, you must save the policy, as shown in Figure 13-72. If you selected Enforce Policy Now, the policy will be executed after saving.
For more information on filesystem extension and LUN provisioning, see IBM Tivoli Storage Resource Manager: A Practical Introduction.
Computers tab
On the Computers tab, select the computers or computer groups to execute the script.
582
The Script Name pull-down field lists all files (including non-script files) in the servers script directory. Attention: For Windows Agents, the script must have an extension that has an associated script engine on the computer running the script (for example: .BAT, .CMD, or .VBS). For UNIX Agents: The extension is removed from the specified script name The path to the shell (for example, /bin/bsh, /bin/ksh) must be specified in the first line of the script If the script is located in a Windows TotalStorage Productivity Center for Data Server scripts directory, the script must have been created on a UNIX platform, and then transferred in binary mode to the Server or you can use UNIX OS tools such as dos2unix to convert the scripts. This will ensure that the CR/LF characters are correctly inserted for execution under UNIX.
Alerts tab
With the Alert tab you can choose to be notified when a script fails due to an unreachable Agent or a script not found condition. The standard Alert Mechanism described in 13.4, OS Alerts on page 555 is used.
583
We will now review the Groups, Probes, Scans, and Profiles definitions for Data Manager for Databases, and show the main differences compared to the core Data Manager monitoring functions. Figure 13-74 shows the navigation tree for Data Manager for Databases.
13.6.1 Groups
To get targeted monitoring of your database assets, you can create Groups consisting of: Computers Databases-Tablespaces Tables Users
Computer Groups
All databases residing on the selected computers will be probed, scanned, and managed for Quotas. The groups you have created using TotalStorage Productivity Center for Data remain available for TotalStorage Productivity Center for Data for Databases. If you create a new Group, the computers you put in it will be removed from the Group they currently belong to. To create a Computer Group, use Data Manager - Databases Monitoring Groups Computer, right-click, and select Create Computer Group. Computer Groups on page 536 gives more information on creating Computer Groups.
Databases-Tablespaces Groups
Creating Groups with specific databases and tablespaces may be useful for applying identical management rules for databases with the same functional role within your enterprise. 584
IBM TotalStorage Productivity Center V2.3: Getting Started
An example could be to create a group with all the Oracle-Server system databases, as you will probably apply the same rules for space and alerting on those databases. This is shown in Figure 13-75.
Table Groups
You can use Table Groups to create Groups of the same set of tables for selected or all database instances. You can use two different views to create a table group:
Tables by instance selects several tables for one instance. Instances by table selects several instances for one table.
You can combine both views as each entry you add will be added to the group.
User Groups
As for core TotalStorage Productivity Center for Data, you can put user IDs in groups. The user groups you create will be available for the whole TotalStorage Productivity Center for Data product set. Tip: The Oracle and MS SQL-Server user IDs (SYSTEM, sa, ...) are also included in the available users list after the first database Probe.
13.6.2 Probes
The Probe process is used to gather data about the files, instances, logs, and objects that make up monitored databases. The results of Probe jobs are stored in the repository and are used to supply the data necessary for Asset Reporting. Use Data Manager - Databases Monitoring Probe, right-click, and select Create Probe to define a new Probe job. In the Instance tab of the Probe configuration, you can select specific instances, computers, and computer groups (Figure 13-76).
585
The Computers list contains only computers that have been defined for Data Manager for Databases. The definition procedure is described in Configuring Data Manager for Databases on page 313.
13.6.3 Profiles
As for TotalStorage Productivity Center for Data, Profiles in Data Manager for Databases are used to determine the databases attributes that are to be scanned. They also determine the summary level and retention time to keep in the repository. Use Data Manager - Databases Monitoring Profiles, right-click, and select Create Profile to define a new profile. Figure 13-77 shows the Profile definition screen.
586
You can choose to gather data on tables size, database extents, or database free space and summarize the results at the database or user level.
13.6.4 Scans
Scan jobs in Data Manager for Databases collect statistics about the storage usage and trends within your databases. The gathered data is used as input to the usage reporting and Quota analysis. Defining a Scan job requires defining: The database, computer, and instances to Scan The tables to monitor for detailed information such as size, used space, indexes, rows count The profile that will determine the data that is gathered and the report views that will be made available by the Scan The job scheduling frequency Oracle-only additional options to gather information about pages allocated to a segment that has enough free space for additional rows The alerting mechanism to use should the Scan fail All this information is set through the Scan definition screen that contains one tab for each previously listed item. To define a new Scan, select Data Manager - Databases Monitoring Scans, right-click and select Create Scan as in Figure 13-78.
Note: If you request detailed scanning of tables, the tables will only be scanned if their respective databases have also been selected for scanning.
587
An interesting Alert is the Archive Log Directory Contains More Than for Oracle, since the Oracle application can hang if there is no more space available for its archive log. This Alert can be used to monitor the space used in this specific directory and trigger a script that will archive the files to an external manager such as Tivoli Storage Manager once the predefined threshold is reached. For a detailed example, refer to IBM Tivoli Storage Resource Manager: A Practical Introduction, SG24-6886.
588
Alert type Database Log freespace lower than Last dump time previous to n days
Oracle
Sybase x x
MSSQL x
x
x
x
x x x x
589
Users tab
On the Users tab, specify the database users you want to be monitored for Quotas. You can also select a profile in the Profile pull-down field on the top right of the tab. In this field, you can select any Profile that stores summary data on a user level. The Quota will only be fired for databases that have been scanned using this Profile (Figure 13-79).
Database-Tablespace tab
Use this tab to restrict Quota checking to certain databases. You can choose several databases or computers. If you choose a computer, all the databases running on it will be included for Quota management.
Alert tab
On the Alert tab you can specify the space limit allowed for each user and the action to run. If no action is selected, the Quota violation will only be logged in the Alert log.
590
13.9.1 Database up
Data Manager for Databases can be used to test for database availability using Probe and Scan jobs since they will fail and trigger an Alert if either the database or the listener is not available. Since those jobs use system resources to execute, you may instead choose scheduled scripts to test for database availability. Due to limited scheduling options and the need for user-written scripts, we recommend using dedicated monitoring products such as Tivoli Monitoring for Databases.
591
Freelist count
You cannot monitor the count of freelists in an Oracle table using Data Manager for Databases.
My Reports will be covered in more detail in 13.14, Creating customized reports on page 683, and 13.15, Setting up a schedule for daily reports on page 697.
The additional feature, TotalStorage Productivity Center for Data for Chargeback produces storage usage Chargeback data, as described in 13.17, Charging for storage usage on page 700.
592
Predefined reports provided by TotalStorage Productivity Center for Data Reports customized and saved by user tpcadmin Schedule reports to run in batch mode Interactive reporting options
Figure 13-80 TotalStorage Productivity Center for Data main screen showing reporting options
Asset Reporting
Asset data is collected by Probe processes and reports on physical components such as
systems, disk drives, and controllers. Currently, Asset Reporting down to the disk level is only available for locally attached devices. Asset Reporting is available for both operating system and database reporting.
593
Availability Reporting
Availability data is collected by Ping processes and allows you to report on the availability of your storage resources and computer systems. Availability Reporting is provided for operating system reporting only.
Capacity Reporting
Capacity Reporting shows how much storage you have and how much of it is being used.
You can report at anywhere from an entire network level down to an individual filesystem. Capacity Reporting is provided for both operating system and database reporting.
Usage Reporting
Usage Reporting goes down a level from Capacity Reporting. It is concerned not so much with how much space is in use, but rather with how the space is actually being used for. For example, you can create a report that shows usage by user, or a wasted space report. You define what wasted space means, but it could be for example files of a particular type or files within a certain directory, which are more than 30 days old. Usage Reporting is provided for both operating system and database reporting.
Backup Reporting
Backup Reporting identifies files that have not been backed up. Backup Reporting is provided
for operating system reporting only.
594
To demonstrate the reporting capabilities of TotalStorage Productivity Center for Data, we installed the Server code on a Windows 2000 system called Colorado, and deployed these Windows Agents: Gallium Wisla Lochness Colorado is also an Agent as well as being the Server. The host GALLIUM has both Microsoft SQL-Server and Oracle database installed to demonstrate database reporting. The Agent on LOCHNESS also provides data for a NAS device call NAS200. The Agent on VMWAREW2KSRV1 also provides data for a NetWare server called ITSOSJNW6. The lab setup is shown in Figure 13-81.
Ethernet
NetWare
IBM NAS200
ibm.com/redbooks
595
By Cluster View
Click By Cluster to drill down into a virtual server or cluster node. You can drill down further to a specific controller to see the disks under it and/or drill down on a disk to see the file systems under it.
By Computer view
Click By Computer to see a list of all of the monitored systems (Figure 13-83.)
596
From there we can drill down on the assets associated with each system. We will take a look at node GALLIUM. In Figure 13-84 we have shown most of the items for GALLIUM expanded, with the details for Disk 2 displayed in the right-hand bottom pane. You will see a detailed level of information, both in terms of the type of objects for which data is collected (for example, Exports or Shares), and the specific detail for a given device.
By OS Type view
This view of the Asset data provides the same information as the By Computer view, with the difference that the Agent systems are displayed sorted by operating system platform.
For IBM SAN Volume Controllers, you can view information about Managed disk groups Managed disks Virtual disks
597
System-wide view
The System-wide view however does provide additional capability, as it can give a System-wide view rather than a node-by-node view of some of the data. A graphical view of some of the data is also available. Figure 13-85 shows most of the options available from the System-wide view and in the main panel, the report of all exports or shares available.
Each of the options available under the System-wide view are self explanatory with the possible exception of Monitored Directories. Data Manager can monitor utilization at a directory level as well as a device or filesystem level. However, by default, directory level monitoring is disabled. To enable directory monitoring, define a Directory Group by selecting Data Manager Monitoring Groups Directory, right-click Directory and choose Create Directory Group. The process of setting up Directory Groups is discussed in more detail in 13.3.2, Groups on page 535. Once the Directory Group is created it must be assigned to a Scan job, and that job must be run on the systems where the directories to be monitored exist. By setting up a monitored directory you will get additional information for that directory. Note that the information collected includes any subdirectories. Information collected about the directory tree includes the number of files, number of subdirectories, total space used, and average file size. This can be graphed over time to determine space usage patterns.
Very specific information regarding an individual database is available as shown in Figure 13-86 for the database DMCOSERV on node COLORADO.
Or you can see rollup information for all databases on a given system (using the System-wide view) as shown in Figure 13-87.
599
All of the database Asset Reporting options are quite straightforward with the exception of one. In order to receive table level asset information, one or more Table Groups needs to be defined. This is a similar process to that for Directory Groups as described in System-wide view on page 598. You would not typically include all database tables within Table Groups, but perhaps either critical or rapidly growing tables. We will set up a group for UDB. To set up a Table Group, Data Manager - Databases Monitoring Groups Table, right-click Table and choose Create Table Group (Figure 13-88).
We have entered a description of Colorado Table Group. Now we click New Instance to enter the details of the database and tables that we want to monitor. From the drop down box, we select the database instance, in this case the UDB instance on Colorado. We then enter three tables in turn. For each table, we entered the database name (DMCOSERV), the creator name (db2admin) and a table name. After entering the values, click Add to enter more tables or finish. We entered the table names of BASEENTITY, DMSTORAGEPOOL, and DMVOLUME, as shown in Figure 13-89. Once all of the tables have been entered click OK.
600
Now we return to the Create Table Group panel, and we see in Figure 13-90 the information about the newly entered tables.
Now we Save by clicking the floppy disk icon and when prompted, we entered the Table Group name of ColoradoTableGroup. In order for the information for our tables to be collected, the Table Group needs to be assigned to a Scan job. We will assign it to the default database scan job called Tivoli.Default DB Scan by choosing Data Manager - Databases Monitoring Scans TPCUser.Default Db Scan.
601
The definition for this scan job is shown in Figure 13-91 and in particular we see the Table Groups tab. Our new Table Group is shown initially in the left hand pane. We moved it to the right hand pane by selecting it and clicking >>. We then save the updates to the Scan job by choosing File Save (or with the floppy disk icon from the tool bar). Finally, we can execute the Scan job by right-clicking it and choosing Run Now. Figure 13-91 shows the Scan job definition after the Table Group had been assigned to it.
Example 13-4 is an extract from the Scan job log showing that the table information is now being collected. You can view the Scan job log through the TotalStorage Productivity Center for Data GUI by first expanding the particular Scan job definition. A list of Scan execution reports will be shown; select the one of interest. You may need to right-click the Scan job definition and choose Refresh Job List. The list of Scan executions for the Tivoli.Default DB Scan is shown in Figure 13-92.
602
Once you have the actual job chosen you can click the detail icon for the system that you are interested in to display the job log. The actual file specification of the log file on the Agent system will be displayed at the top of the output when viewed through the GUI. Example 13-4 shows the actual file output.
Example 13-4 Database scan job showing table monitoring 09-19 18:01:01 DBA0036I: The following databases-tablespaces will be scanned: MS SQLServer gallium/gallium Databases: master model msdb Northwind pubs tempdb Oracle itsrm Tablespaces: ITSRM.DRSYS ITSRM.INDX ITSRM.RBS ITSRM.SYSTEM ITSRM.TEMP ITSRM.TOOLS ITSRM.USERS 09-19 18:01:01 DBA0041I: Monitored Tables: .CTXSYS.DR$OBJECT Northwind.dbo.Employees Northwind.dbo.Customers Northwind.dbo.Suppliers
Finally, we can produce table level asset reports by choosing for example, Data Manager Databases Reporting Asset System-wide All DBMSs Tables By Total Size. This is shown in Figure 13-93.
603
604
605
Often there is a one-to-one relationship between devices and filesystems as seen in Figure 13-97, particularly on Windows systems. However, if a single physical disk has two partitions the detailed description will show two partitions at the bottom of the right-hand pane.
606
Data Manager
With Usage Reporting, you will be able to: Identify orphan files and either update their ownership or delete them to free up space Identify the largest files and determine whether they are needed or whether parts of the data could be archived Identify obsolete files so that they can be either deleted or archived
607
There are a few restrictions on Usage Reporting: In order to report by directory or by Directory Group you will need to set them up in Data Manager Monitoring Groups Directory. UNIX systems do not record file create dates, so no reporting by creation time is available for these systems.
608
A Monitored Tables by RDBMS Type report is shown in Figure 13-100. In this case, only tables which are part of a Table Group, which is included in a Scan job will be reported on.
609
610
We then need to specify in the File Types tab, what a forbidden file is. You can define the criteria as either inclusive or exclusive; that is, you can specify just those files types that will violate the Constraint, or you can specify that all files will violate the Constraint except those specified. There are a number of predefined file types included; you can also chose additional files by entering appropriate values in the Or enter a pattern field at the bottom of the form. We have chosen MP3 and AVI files. The completed File Types tab is shown in Figure 13-102.
The Users tab is very similar to the File Types tab - you can specify which users should be included or excluded from the selection criteria. We have taken the default, which is to include all users. In the Options tab, we nominate a maximum number of rows to be returned. We can also apply some more specific selection criteria here such as only including files that are larger than a defined size. Note, however that these criteria are added to the file list. For example, if we specified here that we only wanted to include files greater than 1 MB, the search criteria would be changed to ((NAME matches any of ('*.AVI', '*.mp3') AND TYPE <> DIRECTORY) OR SIZE > 1 MB). So the returned list of files would be any file greater than 1 MB in size plus any *.MP3 or *.AVI files.
611
If you wish to change the selection criteria so that instead you select any *.MP3 or *.AVI files that are larger than 1 MB, you can enter 1 MB against the bigger than option, and then click the Edit Filter button shown in Figure 13-105. You will then see the file filter as shown in Figure 13-103. To add the size criteria to the file type criteria, click the Size > 1MB entry and drag it up to the All of tag. The changed filter is shown in Figure 13-104. You can also see the Boolean expression for the filter has changed to reflect this condition.
612
In this case we did not want to apply a size criteria, so we left the Option tab entries at their defaults as shown in Figure 13-105.
Finally, we can specify that we want an Alert generated if a triggering condition is met. The only choice here is to specify a maximum amount of space consumed by the files that meet our selection criteria. We left all of the Alert tab options at their defaults other than specifying an upper limit of 100 MB for files that have met our selection criteria.
613
The Alert tab is shown in Figure 13-106. Alerting is covered in more detail in 13.4, OS Alerts on page 555.
We then clicked the Save button and entered a name of Forbidden Files as shown in Figure 13-107.
Before we can report against the Constraint, we need to ensure that a Scan job has been run to collect the appropriate information. Once the Scan has completed successfully, you can go ahead and produce Constraint Violation Reports. Note that you cannot produce a report of violations of a particular Constraint - the report will include entries for any Constraint violation. However, once the report is generated, you can drill down into specific Constraint violations. We produced the report by choosing Data Manager Reporting Usage Violations Constraint Violation By Computer. You will see a screen like Figure 13-108 where you can select a subset of the clients if appropriate - after selecting, click Generate Report.
614
You will then see a list of all of those instances of Constraint violations as shown in Figure 13-109. The report shows multiple types of Constraints. Some of these Constraints were predefined (Orphaned File Constraint and Obsolete File Constraint) and others (ALLFILES and forbidden files) we defined. An orphaned file is any file that does not have an owner. This allows you to easily identify files that belonged to users who have left your organization or have had an incorrect ownership set.
615
From there you can drill down on a specific Constraint, then filesystems within the Constraint, and finally to a list of files that violated the Constraint on that filesystem by selecting the magnifying glass icon next to the entry of interest. Or, as shown in Figure 13-110, by clicking the pie chart icon next to the entry for forbidden files, you can produce a graph indicating what proportion of capacity is being utilized by files violating the Constraint. Position the cursor over any segment of the pie chart to show the percentage and number of bytes consumed by that segment.
Constraint violations are also written to the Data Manager Alert Log. Figure 13-111 shows the same list of violations as if you had produced a Constraint Violations by computer report.
616
617
On the Computers tab we chose our Windows group: tpcadmin.Windows Systems (Figure 13-113).
We then had to specify when and how often we wanted the Quota job to run. We chose to run the job weekly under the When to CHECK tab as shown in Figure 13-114.
618
On the Alert tab, shown in Figure 13-115, we accepted all of the defaults other than to specify the limit under User Consumes More Than, in this case, 1 GB. No Alerts will be generated other than to log any exceptions in the Data Manager Alert Log.
Finally, we save the Quota definition, calling it Big Windows Users as shown in Figure 13-116.
619
The new Quota now appears under Data Manager Policy Management Quotas User > Computer as tpcadmin.Big Windows Users (where tpcadmin is our Data Manager username). We right-clicked the Quota and chose Run Now as in Figure 13-117.
This job will collect data related to the Quota, and add any Quota Violations to the Alert Log as shown in Figure 13-118.
620
We then drilled down on one of the Alerts to see the details (Figure 13-119).
And finally we can create a Quota Violation report by choosing Data Manager Reporting Usage Violations Quota Violations Computer Quotas By Computer. The high-level report is shown in Figure 13-120.
621
We can then drill down further for additional detail or to produce a graphical representation of the data behind the violation. The graph in Figure 13-121 shows a breakdown of the users data by file size.
622
On the Instances tab, expand Computer Groups, select tpcadmin.Windows DB Systems and then click >> to add it to the Current Selections as shown in Figure 13-123.
623
On the When to Run tab shown in Figure 13-124, we chose to run the Quota job weekly and chose a time of day for the job to run. Other values were left at the defaults.
On the Alert tab (shown in Figure 13-125) we specified the actual Quota that we wanted enforced, which was a 100 MB per user Quota. Other values were left as defaults.
624
We saved the new Quota definition with a name of Big DB Users as shown Figure 13-126.
We now run the Quota by right-clicking it and choosing Run Now as seen in Figure 13-127.
625
To check if any user has violated the Quota, navigate Data Manager - Databases Alerting Alert Log All DBMSs All. We see one violation as shown in Figure 13-128.
We can also now run a database Quota violation report by choosing Data Manager Databases Reporting Usage Violations Quota Violations All Quotas By User Quota. This report can be seen in Figure 13-129.
626
627
By default, information on only 20 files will be returned. Figure 13-131 shows the selection screen for the report. You will notice that the report uses the Profile TPCUser.Most at Risk. It is in this Profile that the 20 file limit is set, although the value can be changed. You can override the value on the selection screen, but you can only reduce the value here, not increase it. By updating the Profile you can also exclude files from the report. By default, any file in the
\WINNT\system* directory tree on any device will be excluded. You can add entries to the
exclusion list if appropriate. Ideally, the exclusion list should be the same as that in your backup product.
628
To view the report, click Generate report. We choose to view it as a graphic by then clicking the pie icon and selecting Chart: Space Distribution for All. This is shown in Figure 13-133. This chart tells you the amount of space consumed by files have not been backed up since the last backup was run for this server.
629
We can also select Chart: Count Distribution for All as shown in Figure 13-134 to show the number of files in each category.
The different charts can be viewed in different ways. To select another type of chart, right-click in the chart area, select Customize this chart, and click the radio button next to the desired chart type.
630
Figure 13-136 Backup storage requirement per computer and per filesystem
631
The Incremental Backup Size option makes use of the archive bit, so it can only be used on Windows systems, and if Tivoli Storage Manager is the backup application, the resetarchiveattribute option must be used (for Version 5.2). A sample report is shown in Figure 13-137.
Figure 13-137 Incremental reporting per Node and Filesystem based on files
632
The third report type here is Incremental Range Sizes Reporting. This does not rely on the archive bit (instead, it uses the modification date) so is more generically applicable. It is possible to show through the use of this report the actual difference between a traditional weekly full/daily incremental backup process versus Tivoli Storage Managers progressive incremental approach. To generate this report, select Data Manager Reporting Backup Backup Storage Requirements Incremental Range Sizes By Computer as shown in Figure 13-138.
633
After you select the Computers of interest, click Generate Report. Figure 13-139 shows the output from this report, with the amount of data changed for different time ranges. Note that the values are cumulative, so for each time range; the values shown include the smaller time periods.
634
Intranet
635
Next, we show how to create a Probe for an ESS-F20. Select Probes Select new probe, then under the Computers tab, choose Storage Subsystems. See Figure 13-141.
On the When to PROBE tab, we selected PROBE Now because we need to populate the backend repository. See Figure 13-142.
636
Next is the Alert tab, shown in Figure 13-143. This defines the type of notification for a Probe.
After all parameters are defined, save the Probe definition. At this point the Probe is submitted and will run immediately. Note: For additional information on creating Probes, see 13.3.5, Probes on page 545. There are several ways to check the status of the Probe job. First, we can check the color of the Probe job entry in the navigation tree, then in the content panel. There are two colors that represent job status. They are: GREEN - Job successfully complete with no errors RED - Job completed with errors
637
The status of the Probe job is displayed in text and in color, as shown in Figure 13-144, after selecting the Probe job output in the navigation tree. The job at 1:55 pm is in green, indicating success.
We open the Probe job by selecting it and double-clicking the spy glass icon next to the job in the content window. We see the contents of the job, including detailed information on the status, as in Figure 13-145. Here, we have selected the successful Probe.
638
639
We drill down to the Disk Groups. The disk group contains information related to the ESS, as well as the volume spaces and disks associated with those Disk Groups. Expanding the Disk Group node, a list of all Disk Groups on the ESS displays (Figure 13-147).
Continuing, we expand the disk group DG1 to view the disks and volume spaces within it. We open Volume Space VS3, which shows the disks and LUNs associated with it. The Disks subsection shows the individual disks associated with the Volume Space (see Figure 13-148).
640
Notice the LUNs subsection for disk DD0105 (Figure 13-149). This shows the LUN to disk relationship. The LUNs shown here are just a subset of all the LUNs. You can see that the LUN is spread across all the displayed disks in the content window.
641
Figure 13-150 shows the discovery of a disk with no LUN associations. This is known as a hot spare. It can be used when one of the other seven disks in the disk group fails.
642
We now show a high level view of all disks in ESSF20. There are 32 disks in the ESS, as shown in Figure 13-146 on page 639 in the Number of Disks field. Figure 13-151 shows a partial listing of the disks.
643
We can also display a report of all the LUNs in the ESS. This report provides the physical disk association with each LUN. We have a total of 56 LUNs in the ESSF20 as shown in Figure 13-146 on page 639 (number of LUNS). A partial listing is shown in Figure 13-152.
644
By Computer
We drill down Computers Views By Computer. The report displays the association of filesystems to the storage subsystem, LUNS, and disks on ESSF20. These reports are useful for relating computers and filesystems to different storage subsystem components. There are three options available in the Relate Computers to: pull down, as shown in Figure 13-153.
We select Storage Subsystems from the pull down, select the desired computer and click Generate. Figure 13-154 shows that the generated report TSMSRV43P uses 9.24 GB in the ESS.
645
Returning to the selection screen tab (Figure 13-153 on page 645) we select LUNs. We choose the same host, and click Generate. Figure 13-155 shows the generated report; the relationship between TSMSRV43P and its assigned LUNs. TSMSRV43P has one LUN created on the ESS.
Finally, from the Selection tab (Figure 13-153 on page 645), we select Disks, our host TSMSRV43P, and click Generate. Figure 13-156 shows the report: the ESS disks assigned to the LUN on the host.
646
By Filesystem/Logical Volume
We will now drill to Computer Views By Filesystem/Logical Volume. The report displays the association of filesystems to the storage subsystem, LUNS, and disks on ESSF20. These reports are useful for relating computers and filesystems to different storage subsystem components. There are three options available in the Relate Filesystem/Logical Volumes to pull down, shown in Figure 13-157.
Select Storage Subsystem, the host (TSMSRV43P), and click Generate. Figure 13-158 shows the filesystems on the host, which are located on the ESS.
647
From the Selection tab (Figure 13-157 on page 647) we now choose LUNs, the host (TSMSRV43P), and click Generate. Figure 13-159 shows the LUN location of each filesystem on the host.
From the Selection tab (Figure 13-157 on page 647) we now choose Disks, the host (TSMSRV43P), and click Generate. Figure 13-160 shows which disks are comprising each filesystem and logical volume.
648
By Storage Subsystem
We will now drill down Storage Subsystem Views By Storage Subsystem. These reports display the relationships of the ESS components (storage subsystems, LUNs, and disks) to the computers and filesystems and logical volumes. There are two options available in the Relate Storage Subsystems to: the pull down, shown in Figure 13-161.
Select Computers from the pull down, the subsystem ESSF20, and click Generate. Figure 13-162 shows the space used by each host on the storage subsystem.
649
Now, select Filesystem/logical Volumes from Figure 13-161, the ESSF20 subsystem, and click Generate. Figure 13-163 shows each hosts filesystems and logical volumes, with their capacity and free space.
By LUN
Continuing, we drill down Storage Subsystem Views By LUNs (Figure 13-164).
650
Select Computer from the Relate LUNs to: pull down, select the subsystem (ESSF20) with the associated disks (default is all), and click Generate Report. Figure 13-165 shows the LUNs assigned to each host, with the hosts logical name for the LUN (/dev/hdisk1 in this case).
Now select Filesystem/Logical Volumes from the Relate LUNS to pull down, the ESSF20 subsystem with associated logical disks (default is all), and click. Next, we clicked Generate Report. Figure 13-166 shows the relationships between the LUNs, computers, and filesystems/logical volumes, including free space and host device logical names.
651
Disks
Now we drill to Storage Subsystem Views Disks. There are two options available in the Relate Disks to: pull down, shown in Figure 13-167.
Select Computer from the pull down, the ESSF20 subsystem with related disks (default is all), and click Generate Report. Figure 13-168 shows the relationships of the disks to the hosts.
652
Now select Filesystem/Logical Volumes from the pull down (Figure 13-167 on page 652), the ESSF20 subsystem with related disks (default is all), and click Generate Report. Figure 13-169 shows the relationship between the ESS disks and the filesystems and logical volumes.
Note: For demonstration purposes, we have reduced some of the fields in the reports.
653
Click Generate Report. The report is shown in Figure 13-171. Various columns are displayed: Storage Subsystem Storage Subsystem Type Manufacturer Model Serial Number Computer Filesystem/Logical Volume Path Capacity Free Space Physical Allocation
654
Figure 13-172 shows the right hand columns of the same report.
This report provides quick answers to how much space on the ESS is allocated to each filesystem. Select LUNs this time from the pull-down in Figure 13-170 on page 654. The report in Figure 13-173 shows the LUN to host mapping for the ESS, which filesystem is associated with each LUN, and the free space.
Figure 13-173 Computer view to the filesystem with capacity and free space
655
Click the Generate Report field - the report is shown in Figure 13-175.
656
Note that you can sort the report on a different column heading by clicking it. The current sort field is indicated by the small pointer next to the field name. Clicking again in the same column reverses the sort order.
In the Selection window, specify a date range (optional), and click Generate Report, as shown in Figure 13-177.
657
For each computer, percent availability, number of reboots, total down time, and average downtime is given, as in Figure 13-178 shows the selection. The default sort order is by descending Total Down Time.
You can also display this information graphically, by selecting the pie chart icon at the top of the report, as shown in Figure 13-179.
658
Figure 13-180 shows an unstacked bar chart of the same information (right-click and select Bar Chart).
Figure 13-181 Generate Full Backup Size report Chapter 13. Using TotalStorage Productivity Center for Data
659
Figure 13-182 shows the total disk space used by all the files, and the number of files on each computer. The top column shows the totals for all Agents.
To drill down, select all the computers (using the Shift key) so they are highlighted, then click the pie icon, and select History Chart: Space Usage for Selected. The generated report (Figure 13-183), shows how the total full backup size has fluctuated, and is predicted to change in the future (dotted lines - to disable this, click Hide Trends).
660
To display the file count graph, select History Chart: File count from the pie icon in Figure 13-182. The output report is shown in Figure 13-184, which shows trends in the number of files on each computer.
These reports will help you find potential problems (e.g. a computer system that shows an unexpected sudden upward or downward spike) and also predicts disk and backup requirements for the future.
661
The generated report shows all the filesystems on the selected computers as in Figure 13-186.
The third column shows the total number and total size of files (for all the systems, then broken down by filesystem). Then there are Last Modified columns for one day, one week, one month, two months, three, six, nine, and one year selections. Each of these gives the number and size of the modified files.
662
To generate charts, highlight all the systems, and click the pie icon. Select Chart: Count Distribution for Selected, as shown in Figure 13-187.
The chart is shown in Figure 13-188. Note that when your cursor passes over a bar, a pop-up shows the number of files associated with that bar.
663
You can display other filesystems using the Next 2 and Prev 2 buttons. Change the chart format by right-clicking and selecting a different layout. Figure 13-189 is a pie chart of the same data. The pop-ups work here also.
Figure 13-189 Pie chart selected with number of files which have modified
With these reports you can track and forecast your backups. You can also display backup behavior for the last one, three, nine, or 12 months.
664
This is a quick overview database space consumption across the network. To drill down on a particular RDBMS type, select the appropriate magnifying glass icon as in Figure 13-191.
665
Figure 13-192 shows the fields for an Oracle database. The fields for a DB2 database are as follows: Computer name Total Size Container Capacity Container Free Space Log File Capacity Tablespace Count Container Count Log File Count
666
Note you could select any RDBMS which is installed in your network. The report shows the following information for each Agent with DB2, plus a total (summary): Computer name RDBMS instance RDBMS type Total size Container Capacity Container free space Log file capacity Tablespace count Container count Log file count
667
Select the computer again, and click the magnifying glass. The report shows the entire DB2 environment running on computer Colorado. We have 10 DB2 UDB databases, shown in Figure 13-195 and Figure 13-196.
668
669
Select an Agent, and click the magnifying glass to drill down. Figure 13-198 displays. The report shows the following columns: File Type Path File Size Free Space Auto Extend of an File
670
Select now a particular data file, and click the magnifying glass. The generated pie chart is shown in Figure 13-199. We can see this data file is allocated on the C: drive.
Figure 13-199 Report DB2 File in a Pie Chart for DB2 File
Click the View Logical Volume button at the bottom to display the LUN report (Figure 13-200).
671
Using this procedure, we can find the LUNs where all the database data files are stored. This information is useful for a variety of purposes, e.g. for performance planning, availability planning, and assessing the impact of a LUN failure.
672
Now select the File Filter tab. Click in the All files selected area and right-click to create a new condition, as shown in Figure 13-202.
Enter the desired file pattern into the Match field, and click Add to bring the condition to the display window below, as in Figure 13-203. You can select from different conditions: Matches any of Matches none of Matches Does not match
When you have finished the condition, click OK. In our case we are matching Tivoli Storage Manager option files.
673
Now save the new Profile with an appropriate name, (in this instance, Search for files). The saved Profile now appears in the Profiles list, see Figure 13-205 on page 675. Tip: We recommend choosing meaningful Profile names, which reflect the content or function of the profile.
674
2. Bind new Profile into a Scan. First, create a new Scan - Data Manager Monitoring Scans. We chose TPCUser.Default Scan as shown in Figure 13-206 on page 675. Fill in a description for this Scan and select the Filesystems and Computers on which the Scan will run.
675
On the Profiles tab, select the newly created Profile and add it to the Profiles to apply to Filesystems column, as shown in Figure 13-207.
Now select the schedule time when the schedule should run, save the Scan, then check the result. 3. Generate Report with your Profile. To view the results, select Data Manager Reporting Usage Files File Size Distribution By Filesystem. Select all filesystems you which to report, select the Profile: administrator (Figure 13-208), and click Generate Report. The report contains all the option files discovered by the Scan as in Figure 13-209.
676
677
4. Define new Constraint. We would like to know where specifically these files are located. To set up this search, select Data Manager Policy Management Constraints TPCUser.Orphaned File Constraint, as shown in Figure 13-210. Enter a description, and select the Filesystem Groups and Computers where you want to locate the files.
Select the Options tab, then select Edit Filter as shown in Figure 13-211.
678
On the Edit Filter pop-up, double click the ATTRIBUTES Filter. Here we will replace the ORPHANED condition with our own filter, since we want to actually search for the text files we created, not orphaned files (Figure 13-212).
Use the Del button to delete the ORPHANED condition, then select NAME from the Attributes pull-down, and the Add button to add another Attributes condition. We will specify to search for the text files we created, as in Figure 13-213.
679
After each file pattern entry, click Add to save it. When all search arguments are entered, click OK to save the search. The selection is now complete as in Figure 13-214.
Click OK again. Save the search with a new description and name (File Save As), so that you do not overwrite the original TPCUser.Orphaned File Constraint. We called saved the file as File search. Finally, we run the Scan and check the Scan job log for correct execution, as shown in Figure 13-215.
680
5. Generate Report to find defined Constraint Now look for the results of the file name search. Select Data Manager Reporting Usage Violations Constraint Violations By Computer, select all computers and generate the report. The report will present a summary as in Figure 13-216.
Figure 13-216 Summary report of all Tivoli Storage Manager option files
To drill down, click the magnifying glass on WISLA as in Figure 13-217. This shows all the filesystems on WISLA where matching files were found.
681
Click the magnifying class on a filesystem (C drive, in this case). This will show all the files found which matched the pattern, as in Figure 13-218. Note that there was 1 file reported, which matches the summary view given in Figure 13-209 on page 677.
Figure 13-218 Report for Tivoli Storage Manager Option file searched
You can also drill down to individual files, for detailed information as in Figure 13-219.
682
683
Data Manager
Figure 13-220 shows the available System Reports for Data Manager.
Figure 13-221 shows the output from running the Storage Capacity system report. We could have generated exactly the same output by selecting Data Manager Reporting Capacity Disk Capacity By Computer Generate Report. Obviously, selecting Data Manager My Reports Storage Capacity is a lot simpler.
684
685
Data Manager
We will define a report here for tpcadmin, the username that we are currently logged in as. We will create a report that is exactly the same as the Storage Capacity system report as shown in Figure 13-221 on page 684. In practice this is not something you would normally do as a report already exists. However, this will demonstrate more clearly how the options relate to each other. We select Data Manager Reporting Capacity Disk Capacity By Computer and click Generate Report. Once the report is produced, we save the report definition, using the name My Storage Capacity. This is shown in Figure 13-223.
686
Once the report is saved you will see it available under usernames Reports for tpcadmin as shown in Figure 13-224. There are a few features of saved reports worth mentioning here. Firstly, characteristics such as sort order are not saved with the report definition; however, selection criteria are saved. Secondly, you can override the selection criteria when running your report. By default the objects selected at the time of the save only will be reported. However, you can use the Selection tab when running the saved report to include or exclude objects from the report. If you change selection criteria you can resave the report, or save it under another name to update the definition or create a new definition respectively.
687
The report is more easily run now by choosing IBM Tivoli SRM for Databases My Reports usernames Reports Monitored Tables by RDBMS Type.
688
Data Manager
To set up a new report Data Manager My Reports Batch Reports right-click Batch Reports and select Create Batch Report. You will then see the screen shown in Figure 13-226.
Now, it is a simply a matter of specifying what has to be reported, plus when and what the output should be. In this case we are going to create a system uptime report. As shown in Figure 13-227, we entered our report description of System Uptime and have then selected Availability Computer Uptime By Computer and clicked >>. Our selection is then moved into the right hand panel, Current Selections.
689
We then selected the Selection tab, which is shown in Figure 13-228. Here we are able to select a subset of available data by either reporting for a specified time range or a subset of available systems. We took the defaults here.
On the Options tab, we specified that the report should be executed and generated on the Agent called COLORADO, which is our Data Manager server. We selected HTML for Report Type Specification and then changed the rules for the naming of the output file under Output File Specification. By default the name will be {Report creator}.{Report name}.{Report run number}. In this case we do not really care who created the report and having a variable like report run number, which changes every time a new version of the report is created and makes it difficult to access the file from a static Web page. So we changed the report name to be {Report name}.html. The report will be created in <install-directory>\log\Data-agent-name\reports on the Agent system where the report job is executed. There is no ability to override the directory name. For example, C:\Program Files\tivoli\ep\subagents\TPC\Data\log\colorado\reports on our Windows 2000 Data Manager server COLORADO or /usr/tivoli/tsrm/log/brazil/reports on an AIX Data Manager Agent called BRAZIL. The Option tab is shown in Figure 13-229. Note here that it possible to run a script after the report is created to perform some type of post-processing. For example, you might need to copy the output file to another system if your Web server is on a system that is not running a Data Manager Agent.
690
On the When to REPORT tab we specified when the report should be generated. We chose REPORT Repeatedly and then selected a time early in the morning (3:00 AM) and specified that the report should be generated every day. This is shown in Figure 13-230.
691
We left the Alert tab options as default, but it is possible to generate an Alert through several mechanisms including e-mail, an SNMP trap, or the Windows event log should the generation of the report fail. Finally, we saved the report, calling it System Uptime, as shown in Figure 13-231.
692
Figure 13-233 shows the Report tab. We expanded in turn Usage All DBMSs Tables Monitored Tables By RDBMS Type and clicked >>. We also entered a Description of Monitored Tables by RDBMS Type.
693
We accepted the defaults on the Selection tab, which is to report on all RDBMS types and then went to the Options tab, shown in Figure 13-234. We set the Agent computer, which will run the report to COLORADO. Note that the system that you run the report on must be licensed for each type of database that you are reporting on. If we were to run the report on COLORADO, the Data Manager server system, we would need to have the Data Manager for Databases licences for Oracle and SQL-Server licences loaded there even though COLORADO does not run these databases.
We also set the report type to HTML and changed the output file name to be {Report
694
On the When to Report tab, shown in Figure 13-235, we chose REPORT Repeatedly and set a start time.
We did not change anything in the Alert tab. We saved the definition with the name Monitored Tables by RDBMS Type as shown in Figure 13-236.
695
We can now run the report by choosing Data Manager - Databases My Reports Batch Reports and then right-clicking tpcadmin.Monitored Tables by RDBMS Type and choosing Run Now. Figure 13-237 shows the output from the report execution.
696
697
We then used the Virtual Directory Creation Wizard within Microsoft Internet Information Server (IIS) to set up access to the reports as shown in Figure 13-240. Detailed information on using IIS is shown in 8.2.2, Using Internet Information Server on page 299.
698
We could then access the reports through a Web browser as shown in Figure 13-241.
699
In this example we have chosen to perform Chargeback by computer. It is possible to separately charge for database usage and use a different rate from the computer rate. To do this you would need to set up a Profile that excluded the database data, otherwise, it would be counted twice.
700
Chargeback is useful, even if you do not actually collect revenue from your users for the resources consumes. It is a very powerful tool for raising the awareness within the organization of the cost of storage, and the need to have the appropriate tools and processes in place to manage storage effectively and efficiently. Figure 13-243 shows the Chargeback Report being created. Currently, it is not possible to have the Chargeback Report created automatically (that is, scheduled).
tpcadmin.Linux Systems NAME SPACE GB 0 0 COST 0.104/GB 0.00 0.00 page 2 Aug 23, 2005
klchl5h group total Data Manager - Chargeback Computer Disk Space Invoice
tpcadmin.Windows DB Systems NAME SPACE GB 69 0 69 COST 0.104/GB 7.19 0.00 7.19 page 3 Aug 23, 2005
colorado senegal group total Data Manager - Chargeback Computer Disk Space Invoice
701
tpcadmin.Windows Systems NAME SPACE GB 59 75 75 209 COST 0.104/GB 6.15 7.82 7.82 21.79 page 4 Aug 23, 2005
gallium lochness wisla group total Data Manager - Chargeback Computer Disk Space Invoice
TPCUser.Default Computer Group NAME SPACE GB 137 137 COST 0.104/GB 14.28 14.28 page 5 Aug 23, 2005
415 GB
43.26 43.26
702
14
Chapter 14.
703
submap window
submap stack
704
The NetView window is divided into three parts: The submap window displays the elements included in the current view. Each element can be another submap or a device The submap stack is located on the left side of the submap window. This area displays a stack of icons representing the parent submaps that you have already displayed. It shows the hierarchy of submaps you have opened for a particular map. This navigation bar can be used to go back to a higher level with one click The child submap area is located at the bottom of the submap window. This submap area shows the submaps that you have previously opened from the current submap. You can open a submap from this area, or bring it into view if it is already opened in another window on the window.
705
Figure 14-3 shows the new display using the NetView Explorer.
From here, you can change the information displayed on the right pane by changing to the Tivoli Storage Area Network Manager view on the top pull-down field. The previously displayed view was System Configuration view. The new display is shown in Figure 14-4.
Figure 14-4 NetView explorer window with Tivoli Storage Area Network Manager view
706
Now, the right pane shows Label, Name, Type and Status for the device. You may scroll right to see additional fields.
NetView will display, with a tree format, all the objects contained in the maps you have already explored. Figure 14-6 shows the tree view.
You can see that our SAN circled in red does not show its dependent objects since we have not yet opened this map through the standard NetView navigation window. You can click any object and it will open its submap in the standard NetView view.
707
The Object Properties for that device will display (Figure 14-8). This will allow you to change NetView properties such as the label and icon type of the selected object.
Important: As TotalStorage Productivity Center for Fabric runs its own polling and discovery processes and only uses NetView to display the discovered objects, each change to the NetView object properties will be lost as soon as TotalStorage Productivity Center for Fabric regenerates a new map.
708
Yellow Red
Yellow Red
IBM Tivoli NetView uses additional colors to show the specific status of the devices, however these are not used in the same way by Productivity Center for Fabric (Table 14-2).
Table 14-2 IBM Tivoli NetView additional colors Symbol color Blue Wheat (tan) Dark green Status Unknown Unmanaged Acknowledged Status Meaning Status not determined The device is no longer monitored for topology and status changes. The device was Missing, Suspect or Unknown. The problem has been recognized and is being resolved Status not determined
Unknown
If you suspect problems in your SAN, look in the topology displays for icons indicating a status of other than normal/green. To assist in problem determination, Table 14-3 provides an overview of symbol status with possible explanations of the problem.
Chapter 14. Using TotalStorage Productivity Center for Fabric
709
Table 14-3 Problem determination Display Agents Any Device Normal (green) Link Marginal (yellow) Non-ISL explanation One or more, but not all links to the device in this topology are missing. All links to the device in this topology are missing, while other links to this device in other topologies are normal. All links to the device in this topology are missing, while all other links to devices in other topologies are missing (if any) ISL explanation One or more, but not all links between the two switches is missing All links between the two switches are missing, but the out-of-band communication to the switch is normal All links between the two switches are missing, and the out-of-band communication to the switch is missing or indicates that the switch is in critical condition This condition should not happen. If you see this on an ISL where switches on either side of the link have an out-of-band agent connected to your SAN Manager, then you are having problems with your out-of-band agent. This condition should not happen. If you see this on an ISL where switches on either side of the link have an out-of-band agent connected to your SAN Manager, then you are having problems with your out-of-band agent.
Any
Normal (green)
Critical (red)
Any
Critical (red)
Critical (red)
Both
Critical (red)
Normal (black)
All in-band agents monitoring the device can no longer detect the device. For example, a server reboot, power-off, shutdown of agent service, Ethernet problems, and soon.
Both
Critical (red)
Marginal (yellow)
At least one link to the device in this topology is normal and one or more links are missing. In addition, all in-band agents monitoring the device can no longer detect the device
710
Critical
711
712
There are three different types of views in Productivity Center for Fabric: Device Centric view, Host Centric view, and SAN view. In our configuration, the NetView display (Figure 14-12) shows two separate SANs that we are monitoring: TPC SAN and TSM SAN.
713
714
First we explore the Zone View. The Zone View displays information by zone groupings. Figure 14-14 displays the information about the three zones that have been setup on the Fibre Channel switch: the Colorado, Gallium and PQDI zones.
715
We can drill down in each zone and see which system and devices have been assigned to that specific zone. Figure 14-15 shows the Colorado zone in which there is one host and a FAStT disk subsystem.
716
Now we look at the Topology View (Figure 14-16). The Topology View draws a picture of how the SAN is configured, which devices are connected to which ports, and so on. As we drill down in the Topology View, we first see the interconnect elements. This shows you the connection between any switches. In our small environment, we have only one switch, so the only device connected is the itsosw4 switch, which is an IBM 2109-F16 switch.
717
If we had two switches in our SAN, we would see a switch icon on either side of the Interconnect elements icon. As we drill down on the switch, we see what devices and systems are directly attached to it. Figure 14-17 shows five hosts, the FAStT device, and the IBM switch in the middle.
From here, we show you several features of Productivity Center for Fabric such as: How to configure the manager and what happens when things go wrong Properties of a host with the Productivity Center for Fabric agent installed How to configure SNMP agents
718
We begin by showing what happens when things go wrong. Figure 14-18 shows that the FAStT disk system has a redundant connection. Lets see what happens when one connection goes down.
719
In Figure 14-19 you notice on the left, that all of the parent icons have turned yellow. This indicates that something has happened in your SAN environment. You can then drill down, following the yellow trail until you find the problem. Here we can see that one of the connections to the FAStT disk system has gone down.
This gives an administrator a place to start looking. After they determine what the problem is, they can take corrective action. The FAStT icon has turned Red, not because it has failed, but so you can see that it is affected. In our case, we lost access to one of the controllers of the FAStT, because it was the only path to that controller. If you right-click the FAStT icon and then select Acknowledge, it changes back to Green if the device itself is OK. The path to the icon still remains Yellow. When the problem is corrected, the topology is updated to reflect the resolution.
720
Now let us see the kind of information that we can view from a host that has a Productivity Center for Fabric agent installed on it. You select the required host, and then click SAN SAN Properties. A Properties window (Figure 14-20) opens. It shows such information as IP address, operating system, host bus adapter type driver versions, and firmware levels.
721
When you click the Connection tab on the left, you see the port on the switch to which the specific host is connected as shown in Figure 14-21.
Now that you have seen the agents and where you can define them, lets look at the manager configuration. The manager configuration is simple and enables you set the polling intervals. Figure 9-46 on page 349 shows the polling setup, in which you specify how often you want your agents to poll the SAN. You can set this to minutes, hours, days, weeks, which days you want to poll on, and the exact time. Or you can poll manually by clicking the Poll Now button. The Clear History button changes the state of an object that previously had a problem but is back up. The state appears as yellow, but the Clear History button changes it back to normal or green.
722
723
Figure 14-23 shows the management tool for the IBM 2109 after being launched from Productivity Center for Fabric.
724
725
The Storage Area Network submap (shown in Figure 14-26) displays an icon for each available topology view. There will be a SAN view icon for each discovered SAN fabric (three in our case), a Device Centric View icon, and a Host Centric View icon.
726
You can see in this figure that we had three fabrics. They are named Fabric1, Fabric3, and Fabric4, since we have changed their label using SAN SAN Properties as explained in Properties on page 736. Figure 14-27 shows the complete list of views available. In the following sections we will describe the content of each view.
Topology views
Tivoli NetView root map Storage Area Network SAN view Topology view Switches Elements Interconnect elements Elements (switches) Zone view Zones Elements Device Centric view Devices (storage servers) LUNs Host Platform Host Centric view Hosts Platform Filesystems Volumes
727
Topology view
The topology view is used to display all elements of the fabric including switches, hosts, devices, and interconnects. As shown on Figure 14-29, this particular fabric has two switches.
Now, you can click a switch icon to display all the hosts and devices connected to the selected switch (Figure 14-30).
728
On the Topology View (shown in Figure 14-29 on page 728) you can also click Interconnect Elements to display information about all the switches in that SAN (Figure 14-31).
The switch submap (Figure 14-30), shows that six devices are connected to switch ITSOSW1. Each connection line represents a logical connection. Click a connection bar twice to display the exact number of physical connections (Figure 14-32). We now see that, for this example, SOL-E is connected to two ports on the switch ITSOSW1.
729
When the connection represents only one physical connection (or, if we click one of the two connections shown in Figure 14-32), NetView displays its properties panel (Figure 14-33).
Zone view
The Zone view submap displays all zones defined in the SAN fabric. Our configuration contains two zones called FASTT and TSM (Figure 14-34).
730
Click twice on the FASTT icon to see all the elements included in the FASTT zone (Figure 14-35).
In lab1, the FASTT zone contains five hosts and one storage server. We have installed TotalStorage Productivity Center for Fabric Agents on the four hosts that are labelled with their correct hostname (BRAZIL, GALLIUM, SICILY and SOL-E). For the fifth host, LEAD, we have not installed the agent. However, it is discovered since it is connected to the switch. Productivity Center for Fabric displays it as a host device, and not as an unknown device, because the QLogic HBA drivers installed on LEAD support RNID. This RNID support gives the ability for the switch to get additional information, including the device type (shown by the icon displayed), and the WWN. The disk subsystem is shown with a question mark because the FAStT700 was not yet fully supported (with the level of code available at the time of writing) and Productivity Center for Fabric was not able to determine all the properties from the information returned by the inband and outband agents.
731
In the preceding figure, we can see the twelve defined LUNs and the host to which they have been allocated. The dependency tree is not retrieved from the FAStT server but is consolidated from the information retrieved from the managed hosts. Therefore, the filesystems are not displayed as they can be spread on several LUNs and this information is transparent to the host. Note that the information is also available for the MSS storage server, the other disk storage device in our SAN.
732
We see our four hosts and all their local filesystems whether they are locally or SAN-attached. NFS-mounted filesystems and shared directories are not displayed. Since no agent is running on LEAD, it is not shown in this view.
Starting discovery
You can discover and manage devices that use the iSCSI storage networking protocol through Productivity Center for Fabric using IBM Tivoli NetView. Before discovery, SNMP and the iSCSI MIBs must be enabled on the iSCSI device, the Tivoli NetView IP Discovery must be enabled. See 14.11, Real-time reporting on page 786 for enabling IP discovery. The IBM Tivoli NetView nvsniffer daemon will discover the iSCSI devices. Depending on the iSCSI operation chosen, a corresponding iSCSI SmartSet will be created under the IBM Tivoli NetView SmartSets icon. By default, the nvsniffer utility runs every 60 minutes. Once nvsniffer discovers a iSCSI device, it creates an iSCSI SmartSet located on the NetView Topology map at the root level. The user can select what type of iSCSI device is discovered. From the menu bar, click Tools iSCSI Operations menu and select Discover All iSCSI Devices, Discover All iSCSI Initiators or Discover All iSCSI Targets, as shown in Figure 14-38. For more details about iSCSI, refer to 14.12, Productivity Center for Fabric and iSCSI on page 810.
733
Double-click the iSCSI SmartSet icon to display all iSCSI devices. Once all iSCSI devices are discovered by NetView, the iSCSI SmartSet can be managed from a high level. Status for iSCSI devices is propagated to the higher level, as described in 14.1.9, Status propagation on page 711. If you detect a problem, drill to the SmartSet icon and continue drilling through the iSCSI icon to determine what iSCSI device is having the problem. Figure 14-39 shows an iSCSI SmartSet.
734
This will display a SAN Properties window that is divided into two panes. The left pane always contains Properties, and may also contain Connection and Sensors/Events, depending on the type of object being displayed. The right pane contains the details of the object. These are some of the device types that give information in the SAN Properties menu: Disk drive Hdisk Host file system LUN Log volume OS Physical volume Port SAN
735
Properties
The first grouping item is named Properties and contains generic information about the selected device. The information that is displayed depends on the object type. This section shows at least the following information:
Label: The label of the object as it is displayed by Productivity Center for Fabric. If you
update this field, this change will be kept over all discoveries.
Icon: The symbol representing the device type. If the object is of an unknown type, this
field will be in read-write mode and you will be able to select the correct symbol.
Figure 14-42 shows the Properties section for a host. You can see that it displays the hostname, the IP address, the hardware type, and information about the HBA. Since the host does not give back sensor related events, only the Properties and Connections sections are available.
736
Figure 14-43 shows the Properties section for a switch. You can see that it displays fields including the name, the IP address, and the WWN. The switch is a connection device and sends back information about the events and the sensors. Therefore, all three item groups are available (Properties, Connections, and Sensors/Events).
737
Figure 14-44 shows the properties for an unknown device. Here you can change the icon to a predefined one by using the pull-down field Icon. You can also change the label of a device even if the device is of a known type.
Connection
The second grouping item, Connections shows all ports in use for the device. This section appears only when it is appropriate to the device displayed switch or host. On Figure 14-45, we see the Connection tab for one switch where six ports are used. Port 0 is used for the Inter-Switch Link (ISL) to switch ITSOSW2. This is a very useful display, as it shows which device is connected on each switch port.
738
Sensors/Events
The third grouping item, Sensors/Events, is shown in Figure 14-46. It shows the sensors status and the device events for a switch. It may include information about fans, batteries, power supplies, transmitter, enclosure, board, and others.
739
740
After this, you can launch the Web application by right-clicking the object and then selecting Management Page, as shown in Figure 14-49.
Important: This definition will be lost if your device is removed from the SAN and subsequently rediscovered, since it will be a new object for NetView.
741
The user properties file contains an SRMURL setting that defaults to the fully qualified host name of Tivoli Storage Area Network Manager. This default assumes that both TotalStorage Productivity Center for Disk and TotalStorage Productivity Center for Fabric are installed on the same machine. If TotalStorage Productivity Center for Data is installed on a separate machine, you can modify the SRMURL value to specify the host name of the TotalStorage Productivity Center for Data machine. For instructions on how to do this, please refer to the manual IBM Tivoli Storage Area Network Manager Users Guide, SC23-4698. If the following conditions are true, you can start the TotalStorage Productivity Center for Data graphical interface from the Tivoli NetView console: TotalStorage Productivity Center for Data or the TotalStorage Productivity Center for Data graphical interface is installed on the same machine as TotalStorage Productivity Center for Fabric, or the SRMURL value specifies the hostname of TotalStorage Productivity Center for Data. The TotalStorage Productivity Center for Fabric is currently running. For more information on TotalStorage Productivity Center for Data, see the redbook IBM Tivoli Storage Resource Manager: A Practical Introduction, SG24-6886.
NEW GREEN
Clear History
Device down
NORMAL GREEN
Device down Device up
MISSING RED
Figure 14-51 IBM Tivoli SAN Manager normal status cycle
If you do not manually use NetView capabilities to change status, the status of a Tivoli SAN Manager object goes from green to red and from red to green. Note that the only difference between an object in the NORMAL/GREEN and NEW/GREEN status is in the Status field under SAN Properties (see Figure 14-42 on page 737 for an example). A new object will have New in the field and a normal object will show Normal. The icon displayed in the topology map will look identical in both cases. You can encounter situations where your device is down for a known reason such as an upgrade or hardware replacement and you dont want it displayed with a missing/red status. You can use the NetView Unmanage function to set its color as tan to avoid having the yellow or red status reported and propagated in the topology display. See Figure 14-52.
743
NORMAL GREEN
Device up Device down
MISSING RED
Manage / Unmanage
Clear History
MISSING TAN
Clear History
However, when a device is unmanaged and you do a SAN Configure Manager Clear History to remove historical data, the missing device will be removed from the Productivity Center for Fabric database and will no longer be reported until it is up back with a new/green status. If you have changed the label of the device, and it is re-discovered after a Clear History, it will reappear with the default generated name, as this information is not saved. See Figure 14-53.
MISSING RED
Ack/Unack
Device up
You can use the NetView Acknowledge function to specify that you have been notified about the problem and that you are currently searching for more information or for a solution. This will set the devices color as dark green to avoid having the yellow or red status reported and propagated in the topology display. Subsequently, you can use the Unacknowledge function to return in the normal status and colors cycle. When the device becomes available, it will automatically return to the normal reporting cycle.
744
Cisco 9509
Intranet
Lochness SAN Manager
Sanxc1
Sanxc2
Sanan3
We first deployed an Productivity Center for Fabric Agent to SANAN. Once the agent was installed, it registered with the Productivity Center for Fabric - LOCHNESS and discovered the CISCO1 (MDS 9509). The topology in Figure 14-55 was displayed after deploying the agent. Note: In order to discover the MDS 9000, at least one Productivity Center for Fabric Agent must be installed on a host attached to the MDS 9000. Outband management is not supported for the MDS 9000.
745
To display the properties of CISCO1, right-click the CISCO1 icon to select it and select SAN SAN Properties. See Figure 14-56.
746
The Connection option (Figure 14-57) displays information about the slots and ports where the hosts SANXC1, SANXC2 and SANXC3 are connected, as well as the status of each port.
We see that ITSOSW1 sent a trap to signal that FCPortIndex4 (port number 3) has a status of 2 (which means Offline).
747
The correlation between the inband information and the trap received is then made correctly and only the connection is shown as missing. You can see in Figure 14-59 that the connection line has turned red, using the colors referenced in Table 14-1 on page 709.
We then restored the connection, and following the status cycle explained in Figure 14-51 on page 743, the connections returned to normal (Figure 14-60).
748
Next, we removed one out of the two connections from the host TUNGSTEN to ITSOSW3. One link is lost, so the connection is now shown as suspect (yellow) Figure 14-61.
NetView follows its status propagation rules in Table 14-4 on page 711. This connection links to a submap with the two physical connections. The bottom physical connection is missing (red), and the other (top) one is normal (black), resulting is propagated status of (yellow) on the parent map (left hand side). See Figure 14-62.
749
After removing the link on LEAD and we received a standard Windows missing device popup (Figure 14-64) indicating it could no longer see its FC-attached disk device.
750
Productivity Center for Fabric shows the device as Missing (the icon changes to red see the color status listing in Table 14-1 on page 709) as it is no longer able to determine the status of the device (see Figure 14-65).
In Figure 14-66, the host is Unmanaged (tan) status since we decided to unmanage it.
751
We finally select SAN Configure Manager Clear History (See Figure 14-67).
After the next discovery, as explained in Figure 14-52 on page 744, the host is no longer displayed (Figure 14-68), since it has been removed from the Productivity Center for Fabric database.
Lab 2 environment
For demonstration purposes in the following sections, this lab is referenced as Lab 2. The configuration consists of: Two IBM 2109-S08 (ITSOSW1 and ITSOSW2) switches with firmware V2.6.0g
752
One IBM 2109-S16 (ITSOSW3) switch with firmware V2.6.0g One IBM 2109-F16 (ITSOSW4) switch with firmware V3.0.2 One IBM 2107-G07 SAN Data Gateway Two pSeries 620 (BANDA, KODIAK) running AIX 5.1.1 with: Two IBM 6228 cards One IBM pSeries F50 (BRAZIL) running AIX 5.1.1ML4 with: One IBM 6227 card with firmware 02903291 One IBM 6228 card with firmware 02C03891 One HP Server running HP-UX 11.0 One FC HBA Four Intel servers (TONGA, PALAU, WISLA, LOCHNESS) Two Intel servers (DIOMEDE, SENEGAL) with: Two QLogic QLA2200 card with firmware 8.1.5.12 One IBM xSeries 5500 (BONNIE) with: Two QLogic QLA2300 card with firmware 8.1.5.12 One IBM Ultrium Scalable Tape Library (3583) One IBM TotalStorage FAStT700 storage server Figure 14-69 shows the SAN topology of our lab environment.
Easter - HPUX 11
wisla - Win2k Srv sp3 bonnie - Win2k Srv sp3 sw2 palau- Win2k Srv sp3 sw1
iSCSI
Figure 14-69 SAN lab - environment 2 Chapter 14. Using TotalStorage Productivity Center for Fabric
753
We have powered off the switch ITSOSW4, with managed host SENEGAL enabled. The topology map reflects this as shown in Figure 14-70. The switch and all connections change to red.
The agent running on the managed host (SENEGAL) has scanners listening to the HBAs located in the host. Those HBAs detect that the attached device, ITSOSW4, is not active since there is no signal from ITSOSW4. The information is retrieved by the scanners and reported back to the manager through the standard TCP/IP connection. Since the switch is not active, the hosts can no longer access the storage servers. The active agent (SENEGAL) sends the information to the manager which triggers a new discovery. Since the switch does no longer responds to outband management, Productivity Center for Fabric will correlate all the information and as a result, the connections between the managed hosts and the switch, and the switch itself, are shown as red/missing. The storage server is shown as green/normal because of a second Fibre Channel connection to ITSOSW2. ITSOSW2 is also green/normal because of the outband management being performed on this switch. The active agent host is still reported as normal/green as it sends its information to the Manager through the TCP/IP network. Therefore the Manager can determine that only the agents switch connections, not the host itself, is down.
754
Now, we powered the switch on again. At startup, the switch sends a trap to the manager. This trap will cause the manager to ask for a new discovery. The result is shown in Figure 14-71.
Now, following the status propagation detailed in 14.6, Status cycles on page 743, all the devices are green/normal.
755
You can see under the SAN Properties window, Figure 14-73, that the RNID support only provides the device type (Host) and the WWN. Compare this with the SAN Properties window for a managed host, shown in Figure 14-42 on page 737.
756
To have a more explicit map, we put CLYDE in the Label field (using the method shown in Figure 14-44) and the host is now displayed with its new label.
757
When configuring the agents, we also used the Advanced button to enter the administrator userid and password for the switches. This information is needed by the scanners to obtain administrative information such as zoning for Brocade switches.
758
Productivity Center for Fabric discovers the topology by scanning the three registered switches. This is shown in Figure 14-76. The information about the attached devices is limited to the WWN of the device since this information is retrieved from the switch and there is no other inband management. Note the - signs next to Device Centric and Host Centric Views this information is retrieved only by the inband agent so is not available to us here.
Figure 14-77 shows the information retrieved from the switches (SAN Properties).
759
We reconnected the Fibre Channel connections to all agents into the switch and forced a new polling. We now see that all agents reported information about their filesystems. Since the agents are connected to a switch, the inband agents will retrieve information from it using inband management. That explains why we see all the devices including those without agents installed. Figure 14-79 shows that: Our four inband agents (BRAZIL, GALLIUM, SICILY, SOL-E) are recognized. The two switches ITSOSW1 and ITSOSw2 are found, since agents are connected to them. Device 1000006045161FF5 is displayed since it is connected to the switch ITSOSW1. The device type is Unknown, as there is no inband nor outband agent on this device.
760
We now have no zoning information available since this is retrieved from the switch outband Agent for the 2109 switch. This is indicated by the sign next to Zone View on Figure 14-79.
761
Figure 14-81 Discovered SAN with no LUNS defined on the storage server
762
Figure 14-82 shows that the host CRETE is not included in the MSS zone (we have enabled the outband agent for the switch in order to display zone information). This zone includes TUNGSTEN, which has no LUNs defined on the MSS.
We changed the MSS zone to include the CRETE server. We run cfgmgr on CRETE so that it scans its configuration and finds the disk located on the MSS as shown in Example 14-1.
Example 14-1 cfgmgr to discover new disks # lspv hdisk0 hdisk1 hdisk2 hdisk3 # cfgmgr # lspv hdisk0 hdisk1 hdisk2 hdisk3 hdisk4 00030cbf4a3eae8a 00030cbf49153cab 00030cbf170d8baa 00030cbf170d9439 rootvg None datavg datavg
763
Now, the agent on CRETE is able to run SCSI commands on the MSS and discovers that it is a storage server. Productivity Center for Fabric maps it correctly in Figure 14-83.
Figure 14-83 MSS zone with CRETE and recognized storage server
764
The agent will use inband management to: Query the directly attached devices. Query the name server of the switches to get the list of other attached devices. Launch inband management to other devices to get their WWN and device type (for RNID compatible supported drivers). Launch SCSI request to get LUN information from storage servers. You can see in Figure 14-85 that the agent on GALLIUM has returned information on: Directly attached switches (ITSOSW1 and ITSOSW4) Devices attached to those switches (if they are in the same zones) LUNs defined on the FAStT for this server Its own filesystems Because of the other hosts, only CLYDE runs with RNID compatible drivers, all other devices excluding switches and FAStT storage server are displayed with an unknown device icon. However, we have shown how we can get a complete map of our SAN by deploying just one inband agent.
765
14.8 Netview
In this section we describe how to use the NetView programs predefined performance applications and how to create your own applications to monitor the Storage Area Network performance. The NetView program helps you manage performance by providing several ways to track and collect Fibre Channel MIB objects. You can use performance information in any of the following ways: Monitoring the network for signs of potential problems Resolving network problems Collecting information for trend analysis Allocating network resources Planning future resource acquisition The data collected by the NetView program is based on the values of MIB objects. The NetView program provides applications that display performance information: NetView Graph displays MIB object values in graphs. Other NetView tools display MIB object values in tables or forms.
766
767
SNMP architecture
iSCSI MIB
Ethernet
A typical SNMP manager performs the following tasks: Queries agents Gets responses from agents Sets variables in agents Acknowledges asynchronous events from agents A typical SNMP agent performs the following tasks: Stores and retrieves management data as defined by the MIB Signals an event to the manager
768
TOP
CCITT (0)
JOINT-ISO-CCITT (2)
ISO (1)
STD (0) REG AUTHORITY (1) MEMBER BODY (2)
ORG (3)
DOD (6)
INTERNET (1)
PRIVATE (4)
DIRECTORY (1) MGMT (2) EXPERIMENTAL (3)
MIB (1)
ENTERPRISE (1)
RESERVED (0)
IBM (2)
bcsi (1588)
iSCSI
769
We downloaded the MIBs below and copied them to the directory. v2_6trp.mib (Enterprise Specific trap) v2_6sw.mib (Fibre Channel Switch) v2_6fe.mib (Fabric Element) v2_6fa.mib (Fibre Alliance) Note: If you have unloaded all the MIBs in the MIB description file (\usr\ov\snmp_mibs), you must load MIB-I or MIB-II before you can load any enterprise-specific MIBs. These are loaded by default in NetView. In Example 14-2 we show the \usr\ov\snmp_mibs directory listing with our newly added MIBs.
Example 14-2 MIB directory Directory of C:\usr\ov\snmp_mibs 04/13/2002 08/27/2002 04/13/2002 04/13/2002 09:33a 02:45p 09:33a 09:33a 4 File(s) 81,253 v2_6FA.mib 79,095 v2_6FE.mib 60,139 v2_6SW.mib 5,240 v2_6TRP.mib 225,727 bytes
770
0 Dir(s) C:\usr\ov\snmp_mibs>
IBM 2109
The IBM 2109 comes configured to use the MIB II-private MIB (TRP-MIB), FC Switch MIB (SW-MIB), Fibre Alliance MIB (FA-MIB) and Fabric Element MIB (FE-MIB). By default, the MIBs are not enabled. Here is a description of each MIB and their respective groupings.
771
To enable the MIBs for the IBM/Brocade switch, log into the switch via a telnet session, using an ID with administrator privilege (for example, the default admin ID). We enabled all four of the above MIBS using the snmpmibcapset command. The command can either disable or enable a specific MIB within the switch. Example 14-3 shows output from the snmpmibcapset command.
Example 14-3 snmpmibcapset command on IBM 2109 itsosw2:admin> snmpmibcapset The SNMP Mib/Trap Capability has been set to support FE-MIB SW-MIB FA-MIB SW-TRAP FA-TRAP SW-EXTTRAP FA-MIB (yes, y, no, n): [yes] SW-TRAP (yes, y, no, n): [yes] FA-TRAP (yes, y, no, n): [yes] SW-EXTTRAP (yes, y, no, n): [yes] no change itsosw2:admin>
NetView
The purpose of loading a MIB is to define the MIB objects so the NetView programs applications can use those MIB definitions. The MIB you are interested in must be loaded on the system where you want to use the MIB Data Collector or MIB Tool Builder. Some vendors specific MIBs are already loaded into NetView. Since we want to collect performance MIB objects types for the Brocade 2109 switch, we will load its MIB. On the NetView interface, select Tools MIB Loader SNMP V1. This will launch the MIB Loader interface as shown in Figure 14-90.
772
Each MIB that you load adds a subtree to the MIB tree structure. You must load MIBs in order of their interdependencies. We loaded the v2_6TRP.MIB first by clicking Load then selecting the TRP.MIB from the \usr\ov\snmp_mibs directory see Figure 14-91.
Click Open and the MIB will loaded into NetView. Figure 14-92 shows the MIB loading indicator.
773
We then loaded the v2_6 SW.MIB, v2_6FE.MIB and v2_6FA.MIBs in turn using the same process. You must load the MIBs in order of their interdependencies. A MIB is dependent on another MIB if its highest node is defined in the other MIB. After the MIBs are loaded, we now verify that we are able to traverse the MIB tree and select objects from the enterprise-specific MIB. We used the NetView MIB Browser to traverse the branches of the above MIBs. Click Tools MIB Browser SNMP v1. to launch the MIB browser and use the Down Tree button to navigate down through a MIB (see Figure 14-93).
774
775
Figure 14-96 describes the MIB tree where this object identifier resides. The root of the tree, bcsi, stands for Brocade Communication Systems Incorporated. The next several pages describe the step-by-step process for defining a Data Collection on the swFcPortTxFrames MIB object identifier using NetView.
bcsi (1588)
commDev (2)
fcSwitch (1)
sw (1)
swFCPort (6)
swFCPortTable (2)
swFCPortEntry (1)
swFCPortTxFrames (13)
1. To create the NetView Data Collection, select Tools MIB Collect Data from the NetView main menu. The MIB Data Collector interface displays (Figure 14-97). Select New to create a collection.
776
2. If creating the first Data Collection, you will also see the pop-up in Figure 14-98 to start the Data Collection daemon. Click Yes to start the SNMPCollect daemon.
3. The Data Collection Wizard GUI then displays (Figure 14-99). This is the first step in creating a new Data Collection. By default NetView has navigated down to the Internet branch of the tree (.iso.org.dod.internet). See Figure 14-88 on page 769 for the overall tree structure. Highlight private and click Down Tree to navigate to the private MIB.
We have now reached the private branch of the MIB tree (.iso.org.dod.internet.private). See Figure 14-100.
777
4. Continue to navigate down the enterprise branch of the tree by clicking Down Tree. Figure 14-101 shows the enterprise branch of the tree (.iso.org.dod.internet.private.enterprise).
778
5. We reach the bcsi branch of the tree by clicking Down Tree. Figure 14-102 shows the bcsi (Brocade) branch of the tree (.iso.org.dod.internet.private.enterprise.bcsi).
779
6. We continue to navigate down the tree, using the path shown in Figure 14-96, and, as shown in Figure 14-103 on page 780, eventually reaching: .iso.org.dod.internet.private.enterprise.bcsi.commDev.fibrechannel.fcSwitch.sw.swFCport. swFCPort.swFCPortEnrty.swFCPortTxFrames.
7. We selected swFCPortTxFrames and clicked OK. We received the following pop-up (Figure 14-104) from the collection wizard. This pop-up occurs because this will be the first node added to this collection. NetView then adds the swFCTxFrames MIB Data Collection definition as a valid data collector entry.
780
This launches the Add Nodes to the Collection Dialog, which is the second step in creating a new Data Collection. See Figure 14-105.
8. We proceeded to customize the section Collect MIB Data from fields, using the following steps: a. We entered the switch node name for which we wanted to collect performance data (in this case, ITSOSW2.ALMADEN.IBM.COM) and clicked Add Node. You can add a node either by selecting it on the topology map or typing in the field as the IP address or hostname for the device. Also, you can select multiple devices on the topology map and click Add Selected Nodes from Map. This adds all the selected nodes selected on the topology map to the Collect MIB Data From field. We also added several nodes to the collection by adding one device at a time in the Node field and clicking Add Node. To remove the node, just click the node name in the list and click Remove. b. We then customized the section Set the Polling Properties for these Nodes, using the following steps: i. We changed the Poll Nodes Every field to 5 minutes. This specifies the frequency in which the nodes are polled. Important: Before setting the polling interval, you should have a clear understanding of available and used bandwidth in your network. Shorter polling intervals generate more SNMP data on the network. ii. We checked Store MIB Data. This will store the MIB data that is collected to C:/usr/ov/databases. iii. The Check Threshold if box was checked. This will define the arm threshold. We want to collect data and signal an event each time more than 200 frames are sent on a particular port. Since we checked this box, we will be required to define the trap value and rearm number fields. iv. The option then send Trap Number was configured. We used the default setting, which is the MIB-II enterprise-specific trap.
781
v. We then configured and rearm When. We specified a rearm value of greater than or equal to 75%. of the arm threshold value. This means that a trap will be generated and sent when the number of TX frames reaches 150. Note that these traps are NetView-specific traps (separate from Productivity Center for Fabric traps) and will therefore be sent to the NetView console. 9. Click OK to create the new Data Collection, shown in Figure 14-106. Select the swFCPortTxFrames Data Collection and click Collect.
Note: It could take up to 2 minutes before the newly defined Data Collection is being collected by NetView. To verify that data is being captured, navigate to: c:\usr\ov\databases\snmpcollect. If there are files present, then the Data Collection is functioning properly. 10.Click Close and the Stop and restart Collection dialog is displayed as in Figure 14-107. Click Yes to recycle the snmpcollect daemon. At this point the Data Collection status (Figure 14-106 above) should change from Suspended to To be Collected
We are now collecting the data swFCTxFrames on ITSOSW2. Depending upon the level of granularity that is required for your reporting needs, you may want to collect data over shorter or longer periods. In our lab we collected every 5 minutes, but you may want to collect data once every hour for a week or once every hour for a month. We will now use the NetView Graph tool to display the data collected as described in 14.10.4, NetView Graph Utility on page 784.
782
Note: We followed the same procedure to add the remaining metrics for Data Collection swFCRxFrames, swFCTxErrors, and swFCRxErrors. For demonstration purposes we used a of 50 for an arm threshold and a value of 75% for re-arm. Your values for arm/re-arm may differ from what we used.
Important: There are documented steps on how to perform important maintenance of Tivoli NetView. Refer to the IBM Redbook Tivoli NetVIew and Friends, SG24-6019.
783
If the snmpcollect daemon is not running, you will see a state value of NOT RUNNING from the ovstatus snmpcollect command as shown in Example 14-5.
Example 14-5 snmpcollect daemon stopped C:\>ovstatus snmpcollect object manager name: snmpcollect behavior: OVs_WELL_BEHAVED state: NOT RUNNING PID: 1536 last message: Exited due to user request. exit status: Done C:\>
The snmpcollect daemon can be started manually. At a command prompt, we typed in ovstart snmpcollect. You will see the output shown in Example 14-6. We then issued an ovstatus snmpcollect for verification, as shown in Example 14-4.
Example 14-6 snmpcolllect started C:\>ovstart snmpcollect Done C:\>
Note: If no Data Collections are currently defined to the MIB Data Collector tool, the snmpcollect daemon will not run.
784
We used the NetView Graph tool to display the collected data. This provides a convenient way to display numerical performance information on collected data. We now show how to display the collected data from the previous Data Collection that was built for ITSOSW2 (swFCPortTxFrames). We start by single-clicking ITSOSW2 on the NetView topology map (Figure 14-109).
Select Tools MIB Graph Data to launch the graph utility This will report on the historical data that has been collected on ITSOSW2. After selecting this, NetView takes some time to process the data and present it in the graphical display. The graph build time depends on the amount of data collected. Figure 14-110 shows the progress indicator.
After the graph is built, it displays the swFCTxFrames data that was collected (Figure 14-111). Note there are multiple instances of the object ID mapped that is, swFCPortTxFrames.1, swFCPortTxFrames.2 and so on. In this case they represent the data collected for each port in the switch.
785
For viewing purposes, we adjusted the x-axis for Time by clicking Edit Graph Properties in the open graph window. This allowed us to zoom into shorter time periods. See Figure 14-112.
Any MIB object identifier that has been collected using the NetView MIB Data Collector can be graphed using the NetView Graph facility using the above process.
786
Important: Depending on the configuration, some advanced functionality may be initially disabled in NetView under Tivoli SAN Manager. This section requires this functionality to be enabled. To enable all functionality required, in NetView, click Options Polling and check the Poll All Nodes field. This is shown in Figure 14-114.
787
We will demonstrate how to build a a MIB application that will query the swFCPortTxFrames MIB object identifier in the SW-MIB. This process can used to query any SNMP enabled device using NetView. With the switch ITSOSW2 selected, we start building the MIB Application by launching the Tool Builder. Select Tools MIB Tool Builder New. The MIB Tool Builder interface is launched as in Figure 14-115. Click New to create a new Tool Builder entry for collecting data on ITSOSW2.
The Tool Builder Wizard Step1 window is displayed (Figure 14-116). We entered FCPortTxFrames in the Title field and clicked in the Tool ID field to auto populate the remaining fields. We clicked Next to continue with the wizard.
788
The Tool Wizard Step 2 interface displays. You can see our title of FCPortTxFrames has carried over. We are now ready to select the display type. We can choose between Forms, Tables, or Graphs. We will choose Graph and click New as shown in Figure 14-117.
The NetView MIB Browser is now displayed. We will use the MIB Browser to navigate down to the FCPortTxFrames object identifier. Use the Down Tree button to navigate through the MIB tree. Figure 14-118 shows the path through the SW-MIB error table. Click OK to add the object identifier.
SW MIB - Port Table group private... enterprise... bcsi... commDev... fibrechannel... fcSwitch... sw... swFcPort... swFcPortTable... swFCPortTxFrames
Figure 14-118 SW-MIB Port Table
789
The newly created MIB application is displayed in the Tool Builder Step 2 of 2 window. See Figure 14-119 for the completed MIB Application. Click OK to complete the definition.
Now, the final window for the Tool Builder is displayed. It shows the newly created MIB application in the window, Figure 14-120. Click Close to close the window. The new MIB Application has been successfully created.
790
Clicking on the FCPortTXFrames option, launches a graph utility, shown in Figure 14-122.
The collection of MIB data starts immediately after selecting the swFCPortTXFrames MIB application from the Monitor Other menu. Figure 14-123 shows the data being collected and displayed for each MIB instance of the ITSOSW2.
791
The polling interval of the application can be controlled using the Poll Nodes Every field located under Edit Graph Properties. See Figure 14-124.
792
This launches a dialog to specify how often NetView Graph receives real-time data for graphing, shown in Figure 14-125. This determines how often the nodes are asked for data.
We continued to use the Tool Builder process defined in 14.11.1, MIB Tool Builder on page 787 to build additional MIB applications for real-time performance monitoring. We used the following MIB objects: swFcPortTXWords swFcPortRXC2Frames swFCPortRXC3Frames fcFXPortLinkFailures fcFXPortSyncLosses fcFXPortSigLosses Figure 14-126 shows the newly defined MIB Applications as they appear in the Tool Builder.
793
Figure 14-127 shows all the above MIB objects as they appear in the NetView Monitor pull-down menu. Note we have abbreviated the names of the MIB applications listed in the Monitor Other menu for ease of use.
14.11.3 SmartSets
With Productivity Center for Fabric (Tivoli SAN Manager) providing the management of the SAN, we can further extend the management functionality of the SAN from a LAN and iSCSI perspective. NetView SmartSets gives us this ability. This section describes the concept of the NetView SmartSet. See Figure 14-128 below. For an overview, we provide details on how to group and mange your SAN attached resources from an TCP/IP (SNMP) perspective. By default, the iSCSI SmartSet is created by Productivity Center for Fabric when nvsniffer is enabled. SmartSets for iSCSI initiators and targets can be created using the process described here.
What is a SmartSet? Why SmartSets? Defining a SmartSet SmartSets and Data Collections
Figure 14-128 SmartSet Overview
794
In NetView a SmartSet is used to monitor a set of objects (devices). NetView allows for user-defined SmartSets. We use this to define and manage our SAN devices as one item. SmartSets can be used to group together systems that support a specific operating system, device type or business function. The symbol status displayed for nodes appearing in user-defined SmartSets is based solely on the IP status, not Fibre Channel status. You can customize the attributes available for creating a SmartSets. Refer to the manual Tivoli NetView for Windows Users Guide, SC31-8888 for more information. With Productivity Center for Fabric using the TCP/IP and Fibre Channel protocols to manage the SAN, we will demonstrate how to complement this by using SNMP to manage the same components of the SAN using SmartSets. Important: Depending on the configuration, some advanced functionality required for SmartSets may be disabled in NetView in Productivity Center for Fabric. This section requires this functionality to be enabled. To enable all functionality required, in NetView, click Options Polling and check the Poll All Nodes field. This is shown in Figure 14-114 on page 787. We will demonstrate how to group all the IBM 2109 Fibre Channel switches (ITSOSW1, ITSOSW2 and ITSOSW3) in our configuration, into one SmartSet called IBM2109. 1. On the NetView topology display select the switches ITSOSW1, ITSOSW2 and ITSOSW3. See Figure 14-129 for selected switches. Each symbol can be selected by holding down the Shift-Key and clicking once on each symbol.
795
2. Select Submap New Smartset from the main menu. The Find window is displayed, as in Figure 14-130.
796
3. Click the Advanced tab this will allow the selected switches on the topology map to be added to the SmartSet. See Figure 14-131.
797
4. Click the Add Selected Objects to add ITSOSW1, ITSOSW2, and ITSOSW3 to the Combined Functions field (Figure 14-132).
5. Click Create SmartSet. This launches the New SmartSet dialog. We entered the name of our SmartSet as IBM2109, and added a description. See Figure 14-133. Note that no spaces are allowed in the SmartSet Name field.
798
6. At this point, the SmartSet definition is complete. Click the SmartSets tab to verify that the IBM2109 SmartSet was created as shown in Figure 14-134.
799
3. Clicking on the IBM2109 SmartSet, we find its members ITSOSW1, ITSOSW2, and ITSOSW3, as shown in Figure 14-136. Note: Symbols on the topology map have links back to their respective objects, since the same symbol can reside in more than one location in NetView. In the case of the switch discussed here, the same symbol in the SmartSet also resides on the IP Internet map. Propagation of status occurs to all symbols regardless of their location on the topology. For example, if there is a problem with the switch, causing it to change to a critical (RED) status, this will be reflected in both the SmartSet and on the IP Internet map.
800
SmartSets can be used to group your devices using a logical taxonomy for the enterprise. For our setup, we categorized our SAN resources by Fabric and Operating System. This allows us to easily manage those devices at a high level. Alternatively, we could have grouped the devices by SAN fabric, or by Application or Business Function. We created the following SmartSets, shown in Figure 14-137: IBM 2109 contains all IBM 2109 Fibre Channel switches SANfabricA_AIX contains all AIX SAN attached hosts SANfabricA_HPUX contains all HP-UX SAN attached hosts SANfabricA_Solaris contains all Solaris SAN attached hosts SANfabricA_Win2k contains all Windows 2000 SAN attached hosts TivoliSANManager contains all the Tivoli SAN Manager hosts.
801
Now we can manage our SAN attached devices from both SAN and LAN perspectives from a single console.
802
3. After allowing the Data Collection to collect data, we then graph the data using Tools MIB Graph Data All. The NetView Graph dialog (Figure 14-139) is displayed while the information is collected this can take some time, depending on the amount of data returned.
4. A window displays, presenting all MIB instances of the swFCPortTxFrames MIB object (Figure 14-140) for all three switches in the SmartSet. Since the total number of entries is greater than 15, we get a message on the menu bar indicating that Maximum Graph Lines Exceeded. The NetView Graph utility can only graph 15 lines at a time.
803
5. Next, we need to select the desired instance of the MIB object for each switch that we want to graph. We then clicked Add to add the selected MIB labels to the Lines To Graph panel, then we clicked OK. For this example, we chose the first 5 instances for each of the three switches, shown in Figure 14-141. Click OK to start the graph.
804
The NetView Graph for the fifteen MIB instances we selected is shown in Figure 14-142.
805
NetView uses the default template located in \usr\ov\conf\netmon.seed. We modified the netmon seed file to include the specific IP addresses of all the LAN attached SAN devices. for more details on the seed file please refer to the comments section of the \usr\ov\conf\netmon.seed. Example 14-7 shows a partial listing of the seed file. Note: iSCSI discovery requires that IP discovery in Tivoli NetView that is shipped with Productivity Center for Fabric be enabled. Be aware that when you turn on IP discovery, there can be a lot of network activity depending on how many devices are in your IP network. For this reason we advise the use of the seed file.
Example 14-7 Modified seed file for limited discovery ## All seed file errors are logged in the \usr\OV\log\nv.log file. Any # entry that is invalid will be ignored. # # If the <SystemRoot>\system32\drivers\etc\networks file has entries # for subnets that are contained in your network, the network names # as specified in the file will appear on the map instead of the network # numbers. # ############################################################################ 9.1.38.188 9.1.38.184 9.1.38.186 9.1.38.187 9.1.38.189 9.1.38.153 9.1.38.154 9.1.38.191 9.1.38.155 9.1.38.152 9.1.38.157 9.1.38.158 9.1.38.159 9.1.38.201 !*
Once the seed file is updated and saved, we then need to clear out the NetView databases where the current topology information is stored. Start Server Setup by clicking Options Server Setup as in Figure 14-143. Important: Performing the Clear Databases on NetView will delete all previously saved NetView object and topology information only. This does affect the Tivoli SAN Manager and WebSphere Application Server databases.
806
Now we want to configure NetView to use the updated seed file. Click the Discovery tab in the Server Setup options window. Under Discovery, check Use Seed File, shown in Figure 14-144, and click OK.
807
Click the Databases tab. Click the pull-down, select Clear Databases, shown in Figure 14-145, and click OK. This starts the process to clear the databases.
NetView prompts one last time to verify that you want to clear the databases. Click Yes. Figure 14-146 shows the warning message.
808
Clearing the databases typically takes a minute, however, this will vary depending on the size of the NetView databases being cleared. The NetView console will automatically shut down and restart when complete. See Figure 14-147.
When NetView restarts, it will discover and display the nodes that we defined in our netmon.seed file, shown in Figure 14-148.
809
To demonstrate the difference in the discovered IP topologies, Figure 14-149 shows the NetView display without using a seed file for discovery. In this case, NetView discovers itself and all other nodes on the subnet.
This completes our demonstration of how existing NetView capabilities can be leveraged to further extend the capabilities of Productivity Center for Fabric.
810
iSCSI Components
Application Server
initiator
IP
initiator
target
Client Desktop
Storage
iSCSI uses standard Ethernet switches and routers to move the data from server to storage. It also allows the IP and Ethernet infrastructure to be used for expanding access to SAN storage and extending SAN connectivity across any distance.
811
FC SAN
Database Application Block I/O
iSCSI
Database Application Block I/O
Pooled Storage
Pooled Storage
Below we list some common iSCSI terms: iSCSI Adapter - iSCSI Adapters combine the functions of Network Interface Cards (NICs) with the function of a storage Host Bus Adapter (HBA). These adapters take the data in block form, and perform processing on the adapter card with TCP/IP processing engines, and then send the IP packets across an IP network. The implementation of these functions enables users to create an IP based SAN without lowering the performance of the server. I iSCSI Drivers - before the introduction of iSCSI adapters, some vendors released software versions of iSCSI adapters. These software-enabled adapters accept block level data from applications, but still require CPU cycles for the TCP/IP processing. The advantage of such adapters is that they can work on existing Ethernet NICs. The main disadvantage is that they require heavy CPU utilization for TCP/IP processing. iSCSI Name - The name of the iSCSI initiator or iSCSI target. iSCSI Node - This represent either an iSCSI initiator or iSCSI target. The iSCSI node is identified by its iSCSI name.
812
The iSCSI MIBs and iSNS MIBs are pre-installed into the c:\usr\ov\snmp_mibs directory. This is performed so that the NetView MIB browser can be used to query the iSCSI MIBs.
Restriction: Note that IBM Tivoli NetView does not currently support MIB Tool Builder and Data Collections against SNMP V2.
The iSCSI MIB trap definition files are used by Tivoli NetView for event processing.
iSCSI MIBs
Before managing the iSCSI device, the MIBs must be loaded. By default, the MIBs are not loaded in to Tivoli NetView at installation time. You have to load these MIBs using the NetView MIB loading function. The purpose of loading a MIB is to define the MIB objects so NetViews applications can use those MIB definitions. You load the iSCSI MIB files one at a time into Tivoli NetView.
813
iSNS MIB - The Internet Storage Name Service iSNS defines a mechanism for IP based storage devices to register and query for other storage devices in the network. The iSNS MIB is designed to allow SNMP to be used to monitor and manage iSCSI devices.
See 14.9.3, Loading MIBs on page 771 for detailed instructions on loading MIBs.
14.16.1 Overview
SANs are becoming more and more critical in the corporate infrastructure, therefore they should be made as highly available as possible, just like other IT components. SANs are complex network environments, with potentially hundreds or even thousands of individual devices. Hardware outages cause disruptions to the business environment, leading to lost revenue and reduced customer satisfaction. Minimizing outages due to hardware failures is therefore a goal of SAN management and one way to do this is by predicting and detecting likely errors before they cause outages. Typically, servers have multiple redundant paths to devices. Determining a root error cause in such an environment is usually problematic. Some of the most important factors in complex root cause analysis: Error data can be inconsistent and sparse Complexity of error counter implementations Error indications can be dispersed from the source - they can propagate across the SAN Error Detection and Fault Isolation (ED/FI - SAN Error Predictor) is implemented in Productivity Center for Fabric Version 1.2 to provide a way to predict errors on the optical links that are used to connect SAN components (including HBA to switch, switch to switch, and switch to storage connections.
814
By using Predictive Failure Analysis (PFA), downtime of SAN components can be significantly decreased as it is possible to remove problematic components before failure. This can significantly reduce operational cost of SANs. The ED/FI function collects data from Productivity Center for Fabric agents, outband and/or inband as available. The polling interval is every 15 minutes. The data is stored in the ITSANMDB database.This data is then analyzed using various statistical methods and from this future errors are predicted. The predicted errors are presented in the NetView interface by adorning the appropriate icons as shown in Figure 14-153. The adornment means that the exclamation point is superimposed on the icon representing the device where the error is predicted. A TEC event and SNMP trap are also generated.
Figure 14-154 shows an example of a failing device, in this case, the host SENEGAL. Although in this case, the icon is actually red, indicating a SAN Manager detected failure, note that typically, adorned icons will still show green, indicating they are available. This is because the ED/FI function is designed to flag potential problems before they have escalated to an actual failure. This allows you to replace hardware preemptively at a convenient time, rather than incurring an unplanned outage due to failure.
815
Data is collected from the following counters: FA MIB Counters FE MIB Counters
816
Brocade Switch MIB Counters HBA APIs (Request Port Status, Read Link Status) - inband only
Note: Not all the switch vendors collect data on all the defined counters in the MIB schema. This depends on the particular implementation and adherence to the various standards. At the time of writing, the fullest ED/FI functionality is available on Brocade switches. Fewer counters are available for monitoring on other switch vendors.
Predictive Failure Analysis is build on a Stochastic model called Dispersion Frame Technique (DFT), which was developed and tested at Carnegie Mellon University. The method eliminates complexity through simple and effective pattern recognition of error occurrences. DFT involves a set of rules for predicting failures, based on the proximity of error occurrences to each other in time. ED/FI uses a set of these rules to determine when a set of counters exceeding a threshold will indicate an error. While the specific rules are internal to ED/FI, they are used to detect the difference between normal and abnormal behavior by using an increase in error rate and a decrease in time intervals between error occurrences. An example rule might be to trigger if a counter exceeds a threshold 3 times within a defined interval. When the PFA process sees that counters have changed, along with previous data, it evaluates the counters. If the counter changes meet the criteria using DFT rules, an indication is created. An Indication Record is created for each port/counter/rule group. These indications are them passed on to the Fault Isolation process (FI). The FI process analyzes the indications by further filtering the errors. FI also uses topology and attribute information provided by Productivity Center for Fabric and with this data isolates faults to the specific Fibre Channel (FC) link. If all requirements are met, FI will create a Fault Record. After a defined number of faults occurs (as defined in the FI rules), a Notification Record will be created. The Notification Record will be presented in NetView by adorning the corresponding device as shown in Figure 14-154 on page 816. The Notification Record is permanent and can only be removed with explicit user intervention (via the GUI). When a user clears the adornment, a Cleared Record will be created in the ITSANMDB database and the device port will be set to a cleared state. If another fault occurs on the same port it may be immediately upgraded to a Notification. The whole FI flow is shown in Figure 14-156.
Cleared
FI Upgrade
Fault
User Input
Notification
FI Upgrade
817
Fault Isolation will adorn the transmitter of the link (rather than the sender), because it is most likely that the faulty component in the group of link, transmitter, cable and receiver, is the transmitter.
Note: (1) Switches cannot be adorned if inband agents are not active, except in the case of cascaded switches using outband management only. (2) Endpoint devices cannot be adorned if outband agents are not active. Important: Error counters can also change for non error conditions including:
Rebooting the system Configuration changes Clearing of counter manually As Fault Isolation mechanism will count them as error conditions, it is recommended that Error Detection/Fault Isolation is disabled in such cases to avoid spurious adornments.
818
In this window you can enable or disable ED/FI using the Enable Error Detection and Fault Isolation radio button.
Tip: As stated in the window, it is recommended that you disable the error prediction in case of service actions so that false notifications can be avoided.
819
In the Rule Set Selection you can see the available rules and which rules are active. The active rules here are used in error processing as described in 14.16.2, Error processing on page 816. To see the notes for the specific rule select the rule and click View, you will see a window similar to Figure 14-159.
820
In our example we simulated errors by disabling and enabling a port on the switch ITSOSW1 over a period of time. As well as the graphical display of the adornments, they are also listed under SAN ED/FI Configuration in the Properties tab as shown in Figure 14-161.
821
This window displays the list of potentially faulty SAN devices, using the following columns:
Clear - check this box to clear the adornment on a particular device. Time - the time when the error was identified by FI rules. Faulted Device - the device which was predicted by FI to be failing. The rule here is that the device with the transmitter will be marked as failed, as explained in 14.16.2, Error processing on page 816. If the device has a Productivity Center for Fabric agent installed and running it will appear with its Global Unique Identifier (GUID) similar to the first entry in Figure 14-161 on page 821. If there is no agent running or the device is a switch, the device will be identified by its node WWN. In our example the fifth entry in Figure 14-161 on page 821 is a server without an agent and the sixth is a switch. Faulted Port - if the device has several ports, the actual faulting port WWN will be displayed here. Indicated Device - the device which actually detected the errors. It is identified in the same way as the faulted device. Figure 14-162(which is simply the same as Figure 14-161 on page 821, scrolled to the right), shows an example. Indicated Port - if the device has several ports, the actual port WWN on which errors where detected will be displayed. PD Reference - the reference to Problem Determination guides which can be used by IBM Support to diagnose the problem (if it is an IBM-supported piece of hardware).
822
To find the object with the corresponding GUID or port WWN, enter it in the Object Name field. NetView uses both GUID and port WWNs for the Object Names. As the GUID and ports are usually uniquely identified by less then the whole numeric string, you can use wildcards, rather than the entire string, as shown in Figure 14-163.
In our example we used the last four numbers of the GUID displayed in the first entry shown in Figure 14-161 on page 821. The search string is actually the least significant digits of the GUID, which is truncated in that figure. The full string for the GUID, including the searched string is displayed in Figure 14-166 on page 825. After entering the search string, click OK. The search results are displayed in Figure 14-164.
823
If you double-click on a returned object, NetView will open the topology map, highlighting the device, as shown in Figure 14-165. We can see the notification is for the host SENEGAL, which is adorned.
Now you can clearly see where the faulted device is located in the SAN, and you can start planning the necessary action to diagnose or repair the faulting device. ED/FI isolates faults only to the link level. Therefore, either side of the link or the cable itself might be the faulty component. Before replacing hardware, you should consult your service contracts and product problem determination guides for direction. Cleaning, cable seating, and diagnostic execution are some of the steps that might be recommended that lead to a definitive decision on parts repair or replacement. IBM Service can use ED/FI information in conjunction with problem determination guides to advise what/if part replacements are necessary. If you can identify a component, you should diagnose the problem and repair or replace the component as soon as possible before a permanent failure occurs. If you cannot identify a component, at a minimum you should monitor the link for further errors. In environments where high systems availability is a requirement or service level agreements are in place, you can contact service representatives about replacing the Fibre Channel component.
824
In Figure 14-167 you can see that the selected entry is now removed.
825
The removal is also reflected in the topology map as shown in Figure 14-168. The host SENEGAL is no longer adorned.
826
15
Chapter 15.
827
or:
http://www.ibm.com/storage/support
Then select Storage software in Product family select TPC for Replication select the Install and use tab. The ESS Copy Services supported with TotalStorage Productivity Center for Replication V2.3 include: ESS PPRC Synchronous remote copy Add / delete paths Add / delete volume pairs Full background copy Freeze / Run Suspend / resume Query status of the session, paths, and pairs
828
PPRC
PPRC is a function of a storage server that constantly updates a secondary copy of a volume to match changes made to a primary volume. The primary and the secondary volumes can be on the same storage server or on separate storage servers. PPRC differs from FlashCopy in two essential ways. First, as the name implies, the primary and secondary volumes can be located at some distance from each other. Second, and more significantly, PPRC is not aimed at capturing the state of the source at some point in time, but rather aims at reflecting all changes made to the source data at the target. PPRC is application independent. Because the copying function occurs at the disk subsystem level, the hosts operating system or application has no knowledge of its existence. In contrast to that, host-based mirroring is controlled by software at the operating system or file system level: The storage subsystem does not know about that. Table 15-1 summarizes characteristics of both approaches.
Table 15-1 Comparison of PPRC and host-based mirroring
Host-based mirroring
Operation is performed by host software or host bus adapter, transparent for storage subsystem. The functionality depends on capabilities of the operating system or host bus adapter. Write operations are sent to both volumes. Read operations are sent to any volume, depending on read policy. The relationship between the volumes is symmetric. Failure recovery is identical for both volumes.
FlashCopy
FlashCopy makes a single point-in-time copy of a LUN. This is also known as a time-zero copy. The target copy is available once the FlashCopy command has been processed. FlashCopy provides an instant or point-in-time copy of an ESS logical volume. Point-in-time copy functions give you an instantaneous copy, or view of what the original data looked like at a specific point-in-time.The point-in-time copy created by FlashCopy is typically used where you need a copy of production data to be produced with minimal application downtime. It can be used for backup, testing of new applications, or for copying a database for data mining purposes. The copy looks exactly like the original source volume and is an instantly available. TotalStorage Productivity Center for Replication provides a user interface for creating, maintaining, using volume groups and for scheduling copy tasks. The User Interface populates lists of volumes using the Device Manager interface. TotalStorage Productivity Center for Replication uses different names for copy services than ESS: Point-in-Time Copy is equivalent to FlashCopy on ESS Continuous Synchronous Remote Copy is equivalent to Peer to Peer Remote Copy on ESS Refer to Figure 15-1 for an illustration of these concepts.
829
Figure 15-1 illustrates a list of the tasks you can perform from Manage Replication group, which represents TotalStorage Productivity Center for Replication: Create and manage groups, which are collections of volumes grouped together so that they can be managed concurrently. Create and manage paths between storage subsystems which are required for remote copy functionality. Create and manage pools which are collections of target volumes. Add Replication Devices for improved performance (ESS model 800 only in TPC 2.3) Run the wizard for creating a session: Select copy type Select source group Select target pool Save session or start a replication session
Monitor, terminate, or suspend running sessions A user can also perform these tasks with the TotalStorage Productivity Center for Replication command-line interface, which is described in 15.3, Using Command Line Interface (CLI) for replication on page 884.
830
TotalStorage Productivity Center for Replication supports the session concept in which multiple pairs are handled as a consistent unit. You can create and manage copy relationships between source and target volume pairs or source volume groups, and among target pools through a Replication Manager copy session. The Replication Manager Sessions panel shows sessions and their associated status. The status indicates if the volume is a source, target, or both; and it shows the copy mode of the volume. You can also use this panel to assess if current replication activities are proceeding normally or abnormally. When you are creating a replication session, you can select source and target volume pairs or volume groups, then establish a continuous synchronous remote copy (remote copy) or point-in-time copy (flash copy) relationship between them. The Sessions panel includes the following options:
Create - invokes the Create Session wizard, which you can use to create copy relationships for a new session. Delete - deletes an existing session. Flash - starts a created or terminated session (for Point-in-Time only) Start - starts a created, suspended or terminated session (for Remote Copy only). Properties - displays the Session Properties panel for an existing session. Suspend (consistent) - suspends an existing session, which results in a consistent target copy if there are no errors. Suspend (immediate) - stops an existing session with no guarantee of consistency. Terminate - stops an existing session and withdraws the relationships.
831
TotalStorage Productivity Center for Replication provides the ability to copy a group to a pool, in which case it creates the valid mappings for source and target volumes and optionally presents them to the user for verification that the mapping is acceptable. Sessions are a set of multiple pairs that are managed as a consistent unit from which freeze and run functions can be performed when errors occur. The session can also be viewed as a consistency group. The following chart (see Figure 15-2) graphically depicts the interactions of groups, pools and a session. It shows one group of related volumes on a source ESS (volumes S1 and S2) that we want to copy to another target pool of volumes (the T volumes). Once we have identified and created the source volumes in the group and the target volumes in a pool, we can then establish the relationship.
S1
Remote or Flash Copy
T1
T3
S2
T2
T4
Session
Figure 15-2 Relationship of a group, pool and session
Our example session shows that S1 is associated with T1, similarly S2 with T2. The T1 and T2 volumes are now persistently bound to the relationship whereas T3 and T4 are still available for use. TotalStorage Productivity Center for Replication can automatically create the source to target relationship. Once created, these volumes are now part of a session or consistency group. This means again that any error on any of the volumes in this session could trigger a suspend across all the volumes to ensure data consistency. Events such as loss of access to a source subsystem or the loss of the PPRC links could be examples of such conditions to trigger a freeze event.
832
T1
T2
Sequences will be further utilized in subsequent releases of TotalStorage Productivity Center as more complex copy types are supported.
833
Create a Session
Manage Replication
The ESS Copy Services servers are defined to the CIMOM using the addessserver command. Each ESS cluster which acts as copy services server must be defined to the ESS CIMOM. Refer to Register ESS server for Copy services on page 214. Verify the ESSs you will use are at the required LIC level. TotalStorage Productivity Center for Replication V.2.3 requires for ESS 750 and ESS 800 LIC level 2.4.3.38 or above. ESS models F10 and F20 require LIC of 2.3.256 or above. The paths between ESSs you want to replicate are defined.
Start a Session
Create a Group
4. The Replication Manager Device List panel opens (Figure 15-6). If devices were previously added, they will be displayed in this panel. Click Add... to add a device.
5. The Add New Device wizard launches (Figure 15-7) to guide you through the process. Click Next.
835
6. In step 2 of the wizard, select the storage subsystem you wish to add from the drop down list (Figure 15-8). After choosing a device, click Next.
Note: A successful device Discovery must be completed before adding a replication device.
7. In step 3, you must enter the Specialist user ID and password for Cluster 1 and 2 of the device (Figure 15-9). Click Next.
836
8. In the final step, click Finish to add the device (Figure 15-10).
9. After adding a replication device, the Device List panel opens (Figure 15-11). The device is listed with the connection status information.
837
4. Click Create. The Create Group wizard opens (see Figure 15-13). 5. Click the device shown in Device Components pane and select logical storage subsystem (LSS).
Note: Device Components shown in the group window do not use the same names defined in the Group Contents pane of the IBM Director console (see Figure 15-1 on page 830). The Device Component pane uses the format: device_type.serial_number. In our example Device Component ESS.2105-16603 in Figure 15-13 indicates ESS 2105 F20 16603 in Figure 15-1 on page 830.
838
6. Select one or more volumes (press Ctrl and click for multiple volumes selection) from the Available Volumes pane of the Create group pane. Click Add (see Figure 15-13). You can also click Select all if you want to add all available volumes to a group. In our example we chose two volumes from ESS F20 (16603) and two volumes from ESS 800 (22513).
Note: Although you can only select volumes from one LSS at a time, you can select different LSSs within the same Create Group session. As you select each LSS, the Available volumes pane updates the list of volumes that are available for the selected device.
7.
If you want to remove a volume from the Selected volumes panel, select it, and then click
Remove.
8. Click Next. The Save group window opens (see Figure 15-14).
839
9. Enter a name for the new group in the Name field. The name is required and must not exceed 250 characters and may not contain special characters such as spaces. 10.Enter a description for the new group in the Description field. The Description is optional and can be 0 - 250 characters. 11.Click Finish to save the new group and close the wizard.
Result
The new group appears in the Groups window (see Figure 15-15). In our example we created two groups which will be used for Point-in-time copy and Remote Copy.
840
841
6. To change volumes which belong to the group, click Update. The Group properties window with volumes opens (similar to the one shown in Figure 15-17).
7. To add volumes to the group: Select one or more volumes (using Ctrl) in the Available volumes panel. Click Add. 8. To remove volumes from the group: Select one or more volumes (using Ctrl) in Selected volumes panel. Click Remove.
Attention: Check if existing defined sessions use volumes which you want to remove from the group you are updating.
842
Perform the following steps to view the properties of an Replication Manager group: 1. Expand Multiple Device Manager in the IBM Director Console Tasks pane. 2. Click Manage Replication. 3. Click Groups, the Groups panel opens. 4. In the Groups table, select the group that you want to view (see Figure 15-15 on page 840). 5. Click Properties. The Properties panel opens for the selected group. You can view the following information: Group name Description of the group The table of the volumes that are managed by the group which shows: Volume ID Device (for example ESS.2105-16603) Volume location - logical storage subsystem Volume type (FB for open systems) Volume size
Perform the following steps to delete a Replication Manager group: 1. Expand the Multiple Device Manager tab in the IBM Director Console Tasks pane. 2. Click Manage Replication. 3. Click Groups (see Figure 15-1 on page 830). The Groups panel opens. In the Groups list, select the group that you want to delete (see Figure 15-18).
843
4. Click Delete. A window opens asking to verify the delete request (see Figure 15-19).
5. Click Yes to delete the group. Alternatively, click No to cancel the delete.
844
4. Click Create. The Create Pool Wizard opens (see Figure 15-21).
5. Click the device shown in the Device Component pane and select a logical storage subsystem (LSS).
Note: Device Components shown in the Group window do not use the same names defined in the Group Contents panel in the IBM Director Console (see Figure 15-1 on page 830). The Device Component pane uses the format device_type.serial_number. In our example Device Component ESS.2105-16603 in Figure 15-13 indicates ESS 2105 F20 16603 in Figure 15-1 on page 830.
6. Select one or more volumes (press Ctrl and click for multiple selection) in the Available volumes pane and click Add. You can also click Select all if you want to add all available volumes to a pool. In our example we chose two volumes from ESS F20 (16603) and two volumes from ESS 800 (22513).
Important: The size of a source and target volume of a copy relationship has to be equal.
7.
If you want to remove a volume from the Selected volumes panel, select it, and then click
Remove.
8. Click Next. The Save pool window opens (see Figure 15-22). 9. Enter a name (required), description (optional) and location (optional).
Note: We recommend that you enter a Location name, which helps in automatic allocation of target volumes during creating a session.
845
846
Note: You do not have to use all volumes of pool when you create a session. Additionally, a pool and even the same volume from a pool can be defined as a target for multiple sessions.
847
6. To change volumes which belong to the pool, click Update. The Pool properties window with volumes opens (similar to the one shown in Figure 15-21 on page 845). 7. To add volumes to the group: Select one or more volumes (using Ctrl) in the Available volumes panel. Click Add. 8. To remove volumes from the group: Select one or more volumes (using Ctrl) in Selected volumes panel Click Remove.
Attention: Check if any defined sessions use volumes which you want to remove from the pool you are modifying.
9. Click OK to commit changes and close the window or click Cancel if you want to cancel the modifications.
848
4. Click Delete. A window with the message Are you sure you want to delete pool pool_name? opens as shown in Figure 15-26.
1. Expand the Multiple Device Manager in the IBM Director Console Tasks panel. 2. Click Manage Replication. 3. Click Pools. The Pools panel opens. 4. In the Pools table, select the pool that you want to view. 5. Click Properties. The properties window opens for the selected pool (see Figure 15-24 on page 847). You can view the following information: Pool name Description of the pool Location name The table of the volumes that are managed by the group which shows: Volume ID Device (for example ESS.2105-16603) Volume location - logical storage subsystem Volume type (FB for open systems) Volume size
849
The ability to create paths is now supported within TotalStorage Productivity Center for Replication V.2.3. At the time a Replication Manager session is initiated, the paths in effect at the time are retained, and restored on subsequent restarts of the session. To create a path: 1. From the IBM Director Console Tasks pane, expand the Multiple Device Manager tab. 2. Expand Manage Replication. 3. Click Paths. The Paths panel opens. 4. In the Paths panel, click Create. 5. The Create Path wizard launches (Figure 15-27). Step 1 welcomes you to the wizard. Click Next.
6. In step 2 of the Create Path wizard, select the source storage device and LSS from the drop down menus as seen in Figure 15-28. Click Next.
850
7. In step 3 of the Create Path wizard, select the target storage device and LSS from the drop down menus as seen in Figure 15-29. Click Next.
8. In step 4, select the path type and number of paths to create. In the example below, we create three Fibre Channel paths between our source and target (Figure 15-30).
9. In the final step (Figure 15-31), review the information and confirm the choices by clicking Next.
851
10.The Paths panel is displayed (Figure 15-32) and you can see the status of the path you just created. The path we created above is highlighted and is Established. Click Close to close the panel.
852
4. Select Create session action. The Create session window opens (see Figure 15-34). Choose Point-in-Time Copy and click Next.
Note: You can define another session which uses the same group.
5. The Choose source group window opens (see Figure 15-35). Choose the Group name which you want to copy and click Next. If you ran the wizard from the Groups window you can see only one Group which you selected before.
853
7. In the Location filter field enter the name of the location of the target pool. You can enter an asterisk (*) as a wildcard for the first or last character of the filter. 8. Click Apply to see volumes of all locations which meet the criteria. 9. Select the All listed locations radio button if you want to use volumes from more than one location or the Select single location radio button then select the location from the Location pane and click Next.
Note: We recommend that you enter the entire location name in the Location filter field instead of using wildcards. Remember, location name is case sensitive
10.Enter the session Name and Description in the Create session - Set session settings panel (see Figure 15-37).
854
Select one of the following options in the Session approval pane: Automatic - indicates that you allow Replication Manager to automatically create relationship between source and target volume Manual - indicated that you want to select volumes and approve relationships 11.Click Next, the Review session properties window opens. Verify your input and click Finish to submit (see Figure 15-38).
12.The session will be created and a new window opens with a message that the command completed successfully. If you get a message as shown in Figure 15-41 on page 856, refer to 15.2.13, Creating a session - verifying source-target relationship on page 856. 13.In the Sessions pane you can see the newly created session (see Figure 15-39).
855
If the session was created successfully, select Flash from the Session actions pull down to run a Point-in-time copy session (see Figure 15-40). We recommend that you verify the source-target volumes before running a session. To verify relationships, refer to 15.2.13, Creating a session - verifying source-target relationship on page 856.
14.Now you can see in the ESS Specialist interface that FlashCopy is running as shown in Figure 15-50 on page 861. In our example there are two pairs of FlashCopy on two different ESS devices running in the same session.
Perform the following steps to verify source-volume pairs. 1. If you got a message that creating command was completed with errors, click Details (see Figure 15-41). A window with messages opens, and you can see detailed messages (see Figure 15-42).
856
Close both windows. You can see the created session in the Sessions pane. In our example in Figure 15-43, we created a session named FC_F20_800.
2. Click a session you want to verify in the Sessions panel. 3. Click the box, Please select one drop down, and choose Properties. The window Session properties opens. Click the Copyset tab. See Figure 15-44.
857
4. The number under Non-approved copysets: indicates that during creating a session or relationships they could not be created automatically. In our example, we chose the Automatic Session Approval method; two pairs were set automatically, however the next two were not approved (see Figure 15-44). Click Copyset details. The Copyset window opens as shown in Figure 15-45.
5. Select the Invalid Copyset to see details of the last result and click Modify copyset target. In our example two pairs are approved and two are not valid and should be modified as shown in Figure 15-45.
858
6. The Choose Target window opens. Select target volumes to create copy pair with source volume and click Next. In our example (see Figure 15-46) source volume is 1300 and we have two available targets 1304 and 1305.
7. The Choose Target Verify window opens. If it shows the correct target volume for modifying the copyset click Finish to approve. 8. Perform steps 5 - 7 for all copysets which are invalid, which means that source-target pairs were not set and approved. 9. If all copysets are correct you will see status as shown in Figure 15-47. Select modified copyset to verify the last result says that the relationship was successfully created.
859
10.Go back to Session properties windows, Copyset tab (see Figure 15-48 on page 860) and click Refresh. If you modified all copysets correctly you should get results as shown in Figure 15-48.
11.Go back to the main Sessions window. Select Session pull down actions and click Flash to run a Point-in-time copy session as shown in Figure 15-49. The Confirmation window opens, click Yes to run or No to cancel.
860
12.You can see in the ESS Specialist interface that FlashCopy is running as shown in Figure 15-50. In our example there are two pairs of FlashCopy on two different ESS devices running in the same session.
Figure 15-50 FlashCopy pairs created and run by TotalStorage Productivity Center for Replication
861
Figure 15-51 Create session window with Continuous Synchronous Remote Copy selection
- or 1. In the IBM Director Task panel, click Multiple Device Manager. 2. Click Manage Replication. 3. Double-click Sessions. The Session window opens (see Figure 15-52). 4. Select Create session action. The Create session window opens (see Figure 15-51).
5. Choose Continuous Synchronous Remote Copy and click Next. The Choose source group window opens Choose the Group name which you want to copy and click Next (see Figure 15-53). If you ran the wizard from Groups window you can see only one Group which you selected before. Choose a target pool window opens as shown in Figure 15-54 on page 863.
862
6. In the Location filter field enter the name of the location of the target pool. You can enter an asterisk (*) as a wildcard for the first or last character of the filter. 7. Click Apply to see volumes of all locations which meet criteria. 8. Select All listed locations if you want to use volumes from more than one location or select a single location, then select correct location, and click Next.
Note: Remember, location name is case sensitive
863
9. The Set session settings window opens. Enter the name and description (see Figure 15-55). 10.Select one of the following options in the Session approval panel: Automatic - indicates that you allow Replication Manager to automatically create a relationship between the source and target volume Manual - indicates that you want to select volumes and approve the relationship
11.Click Next, the Review session window opens. Validate the information and click Finish to submit (see Figure 15-56).
12.The session will be created and new window opens with a message that command completed successfully as shown in Figure 15-57. If you get a message as shown in Figure 15-41 on page 856, read 15.2.13, Creating a session - verifying source-target relationship on page 856.
864
13.In the Sessions window you can see a new created session (see Figure 15-58).
Figure 15-58 Sessions window with created Continuous Synchronous Remote Copy session.
14.If session was created successfully, select the session you want to run, select Session actions and click Start to run a Remote Copy session (see Figure 15-59). However, we recommend that you verify source-target volumes before running a session. To verify relationships, read 15.2.13, Creating a session - verifying source-target relationship on page 856.
865
15.You can see in the ESS Specialist interface that a Remote Copy is running as shown in Figure 15-60. In our example there are two pairs of Remote Copy between volumes on two different ESSs running in the same session.
Figure 15-60 Remote copy pairs created and run by TotalStorage Productivity Center for Replication
Using any copy services requires that you create an accurate plan before running and a detailed plan for future management. Any mistakes can cause loss of data, for example if you use a wrong volume as the target of a copy session. Therefore we recommend that you verify all pairs in a session before staring the copy process, see 15.2.13, Creating a session verifying source-target relationship on page 856.
Sessions window
When you create, verify and run a session you can monitor its status in the main Session window, which gives you basic information about a given session. Each session can include many pairs of volumes which are in copy relationships and create a consistent group. If the status of a given session is not optimal, you need to review the properties for a given session to check if there is a general problem or if it is related to a certain pair of volumes.
866
IBM TotalStorage Productivity Center: Getting Started
Perform the following steps to check a basic status of a session: 1. Click Multiple Device Manager in the IBM Director Task panel. 2. Click Manage Replication. 3. Click Sessions. The Sessions window opens (also called the main Session window). There are eight fields in a Sessions window: a. Name of a session b. Status field can have one of the following status:
Normal (green icon): Point-in-Time Copy was invoked successfully. Medium (yellow icon): A session is not started or was terminated Severe (red icon): An error occurred. Defined - a session is created and not started or was terminated Active - a session is running
d. Group - name of a Group of volumes which are sources of copy pairs e. Copy Type - Point-in-Time Copy or Continuous Synchronous Remote Copy f. Recoverable - indicates if any sequences in a session are considered recoverable g. Shadowing - indicates if any part of a session is shadowing data h. Volume Exceptions - shows the total number of volumes which are in an exception state. Before starting a created session, you should see the following field values as shown in Figure 15-61: Status - Medium State - Defined Recoverable - No Shadowing - No Volume Exceptions - No
867
When you successfully flashed a new or terminated session you will see the values for the following parameters shown in Figure 15-70: Status - Normal (green) State - Active (changed from Defined) Recoverable - Yes
Properties window
The Sessions window shows a status of session as a group of volume pairs. If you want to see details, perform following steps to use Properties window: 1. In the IBM Director Task panel, click Multiple Device Manager. 2. Click Manage Replication. 3. Double-click Sessions. The Session window opens 4. Select a session which you want to manage, select the Properties session action. The Properties window opens. 5. There are three tabs in the Properties window a. General - shows general information about a session. Comparing the information to the Session window you get additional information like Description, number of volumes and approval status. The only parameter you can change is the approval status which can be automatic or manual. b. Copyset - you can check if all pairs are valid and approved. This panel is mostly used during verification of a session, see Creating a session - verifying source-target relationship on page 856. c. Sequence - this tab is mostly used to see detailed information about the status of a session, especially when used together with the Pairs window. The General tab shows basic information, like the Session window. For example, when you flashed a session, you should see the following values (see Figure 15-63): Copy type - Point-in-Time Copy State - Active Status - Normal Group - name of group used for this session
868
The Copyset tab generally does not change while managing a session unless some error occurs. Figure 15-64 shows the status in our environment.
Figure 15-64 Copyset tab for correctly defined session with 4 pairs of volumes
869
To see more information about copysets, especially If some of them are invalid, click Copyset details. The Copyset window opens, displaying the table of copysets in the session. You can check for problems in the following tables: The Copyset table indicates if the copyset is invalid. The Last Result column displays the latest message issued for a copyset and indicates why it is invalid. The Last Result column of the Copyset Relationships table displays the last message issued for a copyset pair. If a message ends in E or W, the pair is considered an exception pair. For more details refer to Creating a session - verifying source-target relationship on page 856. The Sequence tab is the most useful when you manage replication sessions, especially during synchronization. You can see which volume pairs are synchronized and the status of the others. In the Sequence panel the following columns are available: Recoverable - true or false. Indicates if all pairs in a sequence are recoverable Exception - yes or no. Indicates if at least one pair is in exception state. Shadowing - yes or no. Indicates if all pairs are in shadowing state. Exception volumes - shows number of volumes which are in exception state. Recoverable pairs - shows number of volume pairs which are recoverable Shadowing pairs - shows number of volume pairs which are in shadowing state Total pairs - shows total number of pairs in a sequence Recoverable timestamp - shows time when a session was suspended Following is an example from our environment of different states of replication session. In our example, after you created or terminated a session, you will see the Sequence tab as shown in Figure 15-65.
Figure 15-65 Sequence tab in Session properties window for defined Point-in-Time Copy session
870
When the session is created or terminated, it is in defined state. You can see in the Sequences pane: Name - Local point in time copy sequence Recoverable - false - it is not recoverable Exception - No - there are no exceptions Shadowing - No - sequence is not shadowing Exception volumes - 0 - no volume is in exception state Recoverable pairs - 0 - no pair is recoverable Shadowing pairs - 0 - no pair is shadowing Total pairs - 4 (in our example) - total number of pairs is two Recoverable timestamp - n/a - is non-available In the Sequence states panel you see that four pairs are in Defined state. To see more details, select Sequence in the Sequences panel and click Pairs. A new window opens as shown in Figure 15-66.
The Sequence Flashed Target pairs window contains the following information:
Source Volume - source volume of a pair, includes type and number of ESS and volume number Target Volume - target volume of a pair State - Defined - means a session is created or terminated but not running Recoverable - No - indicates if a pair is flashed Shadowing - No New - Yes - indicates it is new session
Timestamp
Last result - the code of the last result, you can see description in Last result panel if you click one pair in Pairs panel
871
When you flash a new or terminated session you will see the Sequence tab as shown in Figure 15-77.
Notice the values of the following columns: Recoverable - true Recoverable timestamp - time, when Point-in-Time Copy session was successfully flashed The Sequence Flashed Target pairs window shown in Figure 15-68 shows successfully flashed volumes.
872
Using any copy services requires that you create an accurate plan before running and a detailed plan for managing the copy services. Any mistake can cause loss of data, for example when you use the wrong volume as a target of a copy session. Therefore we recommend that you verify all pairs in a session before starting the copy process. Refer to Creating a session - verifying source-target relationship on page 856.
Sessions window
When you create, verify, and run a session, you can monitor its status in the main Session window, which gives you basic information about a given session. Each session can include many pairs of volumes which are in copy relationships and create a consistent group. If the status of a given session is not optimal, you need to review the properties for a given session to check if there is a general problem or if it is related to a certain pair of volumes. Perform the following steps to check a basic status of a session: 1. Click Multiple Device Manager in the IBM Director Task panel. 2. Click Manage Replication. 3. Click Sessions. The Sessions window opens (also called as main Session window) 4. There are eight fields in a Sessions window: a. Name of a session b. Status field can have one of the following status:
Normal (green icon): All source volumes are replicating in both directions and copy is active. All volumes were established successfully and are synchronized. Medium (yellow icon): A session is not started, was terminated or is synchronizing but at least one volume is not synchronized with a source. Severe (red icon): An error caused a hardware device to respond at multiple addresses or, for a fibre-channel connection, a volume failed to be established. Defined - a session is created and not started or was terminated Active - a session is running
d. Group - name of a Group of volumes which are sources of copy pairs e. Copy Type - can be a Point-in-Time Copy or Continuous Synchronous Remote Copy (as described in this chapter) f. Recoverable - indicates if any sequences in a session are considered recoverable g. Shadowing - indicates if any part of a session is shadowing data h. Volume Exceptions - shows the total number of volumes which are in an exception state.
Chapter 15. Using TotalStorage Productivity Center for Replication
873
After you have created a session, before starting it, you should see the following values for several fields (see Figure 15-69): Status - Medium State - Defined Recoverable - No Shadowing - No Volume Exceptions - No
When you start a new session or resume suspended session you will see the following values (see Figure 15-70): Status - Medium (is still not optimal) State - Active (changed from Defined) Recoverable - No Shadowing - Yes (changed) Volume Exceptions - No
874
If all pairs in a session are synchronized you should see the following values (see Figure 15-71): Status - Normal (changed, now it is optimal state) State - Active Recoverable - Yes (changed, now you can recover data in case of disaster) Shadowing - Yes Volume Exceptions - No
If a session is suspended you should see the following values (see Figure 15-72): Status - Normal State - Active Recoverable - Yes Shadowing - No Volume Exceptions - No
875
Properties window
The Session Properties window shows a status of the session as a group of volume pairs. If you want to see details, perform the following steps to use Properties window: 1. In the IBM Director Task panel, click Multiple Device Manager. 2. Click Manage Replication. 3. Double-click Sessions. The Session window opens 4. Select a session which you want to manage, select Properties session action. The Properties window opens. 5. There are three tabs in the Properties window: General - This tab shows general information about a session. Comparing to Session window you get additional information such as Description, number of volumes and approval status. The only parameter you can change is the approval status which can be automatic or manual. Copyset - This tab lets you check if all pairs are valid and approved. This panel is mostly used during verification of a session, see Creating a session - verifying source-target relationship on page 856. Sequence - This tab is mostly used to see detailed information about the status of a session, especially together with the Pairs window. The General tab shows basic information, like the Session window. For example, when you create a session, before starting it, you should see the values in Figure 15-73: Copy type - Continuous Synchronous Remote Copy State - Defined Status - Medium Group - name of group used for this session Source Volumes - number of volumes in a group Approval status - Automatic or Manual
876
The Copyset tab information generally does not change while managing a session unless some error occurs. You should see the following status as shown in Figure 15-74.
Figure 15-74 Copyset tab in Properties window for correctly defined session
To see more information about copysets, especially If some of them are invalid, click Copyset details. The Copyset window opens, displaying the table of copy sets in the session. You can check for problems in the following tables: The Copyset table indicates if the copy set is invalid. The Last Result column displays the latest message issued for a copyset and indicates why it is invalid. The Last Result column of the Copyset Relationships table displays the last message issued for a copyset pair. If a message ends in E or W, the pair is considered an exception pair. For additional details, refer to Creating a session - verifying source-target relationship on page 856.
Sequence tab is the most useful when you manage replication sessions, especially during synchronization you can see which volume pairs are synchronized and the status of the others. In the Sequence panel the following columns are available:
Recoverable - true or false. Indicates if all pairs in a sequence are recoverable Exception - yes or no. Indicates if at least one pair is in exception state. Shadowing - yes or no. Indicates if all pairs are in shadowing state. Exception volumes - shows number of volumes which are in exception state. Recoverable pairs - shows number of volume pairs which are recoverable Shadowing pairs - shows number of volume pairs which are in shadowing state Total pairs - shows total number of pairs in a sequence Recoverable timestamp - shows time when a session was suspended The following example is taken from our environment, showing different states of a replication session.
877
If you created session or terminated running session, you will see the Sequence tab as shown in Figure 15-75.
Figure 15-75 Sequence tab in Session properties window for defined session
When the session is created or terminated, it is in a defined state. You can see in Sequences panel the following values: Recoverable - false - it is not recoverable Exception - No - there are no exceptions Shadowing - No - sequence is not shadowing Exception volumes - 0 - no volume is in exception state Recoverable pairs - 0 -no pair is recoverable Shadowing pairs - 0 -no pair is shadowing Total pairs - 2 - total number of pairs is two Recoverable timestamp - n/a - is non-available In the Sequence states panel you see that two pairs are in Defined state. For more details, select Sequence in Sequences panel and click Pairs. The Sequence Remote Target pairs window opens as shown in Figure 15-76.
878
The Sequence Remote Target pairs window contains the following columns:
Source Volume - source volume of a pair, includes type and number of ESS and volume number Target Volume - target volume of a pair State - Defined - means a session is created or terminated but not running Recoverable - No - indicates if a pair is synchronized Shadowing - No New - Yes - indicates it is new session Timestamp Last result - the return code of the last result, you can see a description in the Last result panel if you click one pair in the Pairs panel
When you start a new session or resume suspended session you will see the Sequence tab as shown in Figure 15-77
Figure 15-77 Sequence tab in Session properties window for just started session
879
The following columns have changed their state: Shadowing - yes Shadowing pairs - 2 It looks similar to the Sequence Remote Target pairs window as shown in Figure 15-78.
If one volume is synchronized but another is still synchronizing you will see the status as shown in Figure 15-79.
Figure 15-79 Sequence tab in a Session properties window for partially synchronized session
880
One pair is in Duplex state which means synchronized and another pair is in synchronizing state. Notice that Recoverable state is still false, because not all pairs are synchronized. To see which pair is in full duplex state, click Pairs (see Figure 15-80).
In our example, one pair is in Duplex state (volume 1703 on ESS F20 16603 is synchronized with volume 1301 on ESS 800 22513) while the second pair is still synchronizing. When all pairs in a session are synchronized, you will see the status as shown in Figure 15-81.
Notice that Recoverable status is true which means that all pairs are in Duplex state which means synchronized. The same status is shown for all pairs separately in the Sequence Remote Target pairs window as shown in Figure 15-82.
881
When a session is fully synchronized you can suspend a session to have a consistent state of data on remote server. If you successfully suspended a session you will see Sequence tab information as shown in Figure 15-83.
Figure 15-83 Sequence tab in Session properties window in successfully suspended state
Exception - No - there are no exceptions Shadowing - No - sequence is not shadowing Exception volumes - 0 - no volume is in exception state
Recoverable pairs - 2- two pairs are recoverable
Shadowing pairs - 0 -no pair is shadowing Total pairs - 2 - total number of pairs is two
Recoverable timestamp - time, when a session was successfully suspended
In the Sequence states panel you see that two pairs are in Suspended state. To see more details, select Sequence in Sequences panel and click Pairs. A new window opens as shown in Figure 15-84.
882
For a successfully suspended session, you should see the following values in Pairs window: State - Suspended Recoverable - Yes Shadowing - No New - No
Important: Remember to check the status of a session if it is successfully synchronized before you invoke the suspend command. Otherwise you will get invalid and inconsistent data on the remote site.
If you suspended a session which was not synchronized you will get the information in the Sequence tab as shown in Figure 15-85.
883
Shadowing - No - sequence is not shadowing Exception volumes - 0 - no volume is in exception state Recoverable pairs - 0- no pair is recoverable Shadowing pairs - 0 -no pair is shadowing Total pairs - 2 - total number of pairs is two
Recoverable timestamp - n/a, recovery is not possible so there is no time information
Notice that state is Suspended but a session is not recoverable. In a Sequence states panel you see that two pairs are in Suspended state but the Recoverable value is false. To see more details, select Sequence in Sequences panel and click Pairs. A new window opens as shown in Figure 15-86.
When a session was suspended in an inconsistent state, you will see the following values in the Pairs pane: State - Suspended Recoverable - No Shadowing - No New - Yes
884
repcli utility
To use the CLI, you have to run the repcli utility. The default folder location of CLI for Replication Manager is c:\Program Files\IBM\mdm\rm\rmcli. This utility can also run commands in interactive mode, a single command, or a set of commands from a script. Syntax of repcli command:
repcli [ { -ver|-overview|-script file_name|command | - } ] [ { -help|h|-? } ]
Where:
-ver
Displays the overview information about the repcli utility, including command modes, standard command and listing parameters, syntax diagram conventions, and user assistance.
-script filename
Runs the set of command strings in the specified file outside of a repcli session. You must specify a file name. The format options specified using the setoutput command apply to all commands in the script. Output from successful commands routes to stdout. Output from unsuccessful commands route to stderr. If an error occurs while one of the commands in the script is running, the script exits at the point of failure and return to the system prompt. example:
repcli -script start_backup.scr
-command_string
885
mkcpset
rmtgtpool
stopsess suspendsess
In this section we focus on a few of these commands, which are mostly used for managing replication sessions:
flashsess - start Point-in-Time Copy session lspair - show information about a copy pair for a session lsseq - show information about a sequence for a session lssess - show details about all or filtered sessions setoutput - change default format for output showsess - show details about certain session startsess - start Continuous Synchronous Remote Copy session stopflashsess - terminate Point-in-Time Copy session stopsess - terminate Continuous Synchronous Remote Copy session suspendsess - suspend Continuous Synchronous Remote Copy session
lssess command
lssess [ { -help|-h|-? } ] [ { -l (long)|-s (short) } ] [-fmt default|xml|delim|stanza] [-p on|off] [-delim char] [-hdr on|off] [-r #] [-v on|off] [-cptype flash|pprc] [-state defined|active] [-status norm|warn|sev|unknown] [-recov yes|no] [-shadow yes|no] [-err yes|no] [session_name ... | -] -s
Displays more details - default output plus approval type, pool criteria, copysets, non-approved, invalid, and description.
-cptype copytype
An optional parameter that displays only the sessions with the copy type specified.
-state defined | active
886
An optional parameter that is set to yes or no to indicate whether the session can be considered recoverable, based on whether any sequences in the session can be considered recoverable.
-shadow yes | no
An optional parameter that indicates whether any part of the session is shadowing data.
-err yes | no
An optional parameter that displays only the sessions with the session name specified. Separate multiple session names with a comma between each name. If no session name is specified, all sessions are displayed unless another filter is used. In our example you should see the following results for created sessions using the lssess command as shown in Example 15-1.
Example 15-1 lssess - defined sessions repcli> lssess Name Status State Group Type Recover Shadow Err ================================================================= PPRC_800_to_F20 warning Defined PPRC_src pprc No No No FC_F20_800 warning Defined FC_src flash No No No
showsess command
You can see details about certain sessions using lssess with the -l parameter or the showsess command.
showsess session_name
This command shows the following information (like lssess -l): Name - Session Name. Copy type - Point-in-Time Copy or Continuous Synchronous Remote Copy. State - Defined or Active. Status - Unknown, Normal, Low, Medium, Severe, or Fatal. Group - Name of group of source volumes. Source volumes - Shows number of volumes in the group being replicated by this session. Approval status - Automatic or Manual. Copysets - Shows number of copysets that the session is managing. Non-approved - Indicates the number of copysets that have yet to be verified. Invalid copysets - Indicates the number of copysets that were determined to be invalid. Seq - Valid sequence names are Remote Target for remote copy and Flashed Target for point-in-time copy. Use quotes around the entire flag e.g. Flashed Target:location=RTP.
Chapter 15. Using TotalStorage Productivity Center for Replication
887
Pool Criteria - Location exact name or filter. Shadow - Yes or no. Indicates if a session is shadowing data. Recov - Yes or no. Indicates if all pairs in a session are recoverable. Approve - Yes or no. Indicates if all copysets are approved. Description - User-defined session description. In Example 15-2 you can see the result of the showsess command for defined sessions. You can compare the parameters to the information you can get using the graphical interface as described in Managing a Continuous Synchronous Remote Copy on page 873.
Example 15-2 showsess - defined sessions repcli> showsess PPRC_800_to_F20 Name PPRC_800_to_F20 Type pprc State Defined Status warning Group PPRC_src Source Volumes 2 Approval Status Automatic Copysets 2 Non-approved 0 Invalid 0 Seq "Remote Target" Pool Criteria F20 Shadow No Recover No Err No Approve Yes Description Remote copy of 2 volumes from ESS 800 to F20 AWN007080I Command completed successfully. repcli> showsess FC_F20_800 Name FC_F20_800 Type flash State Defined Status warning Group FC_src Source Volumes 4 Approval Status Automatic Copysets 4 Non-approved 0 Invalid 0 Seq "Flashed Target" Pool Criteria P% Shadow No Recover No Err No Approve Yes Description Point in time copy of 4 volumes, 2 on ESS F20 and 2 on ESS 800 AWN007080I Command completed successfully.
888
flashsess command
To run a created or terminated Point-in-Time Copy session, invoke flashsess command:
flashsess [-quiet] session_name [. . .] session_name
specifies the session name to be activated. Separate multiple session names with a white space between each name. Alternatively, use the dash (-) to specify that input for this parameter comes from an input stream (STDIN).
-quiet
An optional parameter that turns off the confirmation prompt for this command.
Note: In a batch program use the quiet parameter where available, otherwise the program will wait for your confirmation
To start a created, terminated, or suspended Continuous Synchronous Remote Copy session, invoke the startsess command as shown in Example 15-4.
Example 15-4 startsess command repcli> startsess PPRC_800_to_F20 AWN007100I Command completed successfully.
Example 15-5 shows the status of started sessions. Point-in-Time Copy was created successfully which confirms normal status and yes value of recover parameter. However, Continuous Synchronous Remote Copy is running (Active State) but not synchronized, which shows in the Recover and Status parameters.
Example 15-5 lssess - started sessions repcli> lssess Name Status State Group Type Recover Shadow Err ================================================================ PPRC_800_to_F20 warning Active PPRC_src pprc No Yes No FC_F20_800 normal Active FC_src flash Yes Yes No
lsseq command
You should use two additional commands, lsseq and lspair, to get more details about the current state of sessions.
lsseq [ { -l |-s } ] [-recov yes|no] [-shadow yes|no] [-err yes|no] session_name -s
An optional parameter that displays all valid output. This is the default.
889
-recov yes | no
An optional parameter that indicates whether any sequences in the session can be considered recoverable.
-shadow yes | no
An optional parameter that indicates whether or not the sequence is shadowing (copying) the data.
-err yes | no
Specifies the session name to be activated. In Example 15-6 you can find the time when the Point-in-Time Copy was run in the Timestamp column. For Continuous Synchronous Remote Copy session, one pair is synchronized, which shows in the Recov Pairs parameter. The state of the Recov parameter will change to yes when all pairs are synchronized.
Example 15-6 lsseq - started sessions
repcli> lsseq FC_F20_800 Name Recov Err Shadow Err Vols Recov Pairs Shadow Pairs Total Pairs Recov Timestamp ===================================================================================================== Flashed Target Yes No Yes 0 4 4 4 2005/04/12 16:34:00 PDT repcli> repcli> lsseq PPRC_800_to_F20 Name Recov Err Shadow Err Vols Recov Pairs Shadow Pairs Total Pairs Recov Timestamp ============================================================================================ Remote Target No No Yes 0 1 2 2 n/a
lspair command
lspair [ { -l |-s } ] { -seq sequence_name|-cpset source_vol_id } [-state defined |active|duplex|suspended|synch|flashed] [-recov yes|no] [-shadow yes|no] [-new yes|no] [-err yes|no] session_name | -
You can use the lspair command to list the source and target of the copy service pairs and their status.
-s
Displays only pairs of the sequence name specified. Mutually exclusive with -cpset.
-cpset source_vol_id
Specifies the source volume ID of the copy set on which you want a list of pairs. Mutually exclusive with -seq.
-state defined | active | duplex | suspended | synch | flashed
An optional parameter that displays the state. The state can be defined, active, duplex, suspended, synch, or flashed.
-recov yes | no
An optional parameter that displays only pairs in the corresponding recoverable state.
890
-shadow yes | no
An optional parameter that displays only pairs that are in the new state.
-new yes | no
An optional parameter that displays only pairs that are in the new state specified.
-err yes | no
An optional parameter that displays only pairs that are in the error state. session_name The session name by which the pairs are identified. In Example 15-7 you can see details about volume pairs. For a Continuous Synchronous Remote Copy session, one pair of volumes is synchronized, which shows Duplex State, but the second one is still synchronizing.
Example 15-7 lspair - started sessions
repcli> lspair -seq 'Flashed Target' FC_F20_800 Source Target State Recov Shadow New Copyset Timestamp Last result ==================================================================================================================================== ESS:2105.16603:VOL:1702 ESS:2105.16603:VOL:1706 Flashed Yes Yes No ESS:2105.16603:VOL:1702 2005/04/12 16:34:00 PDT IWNR2016I ESS:2105.16603:VOL:1703 ESS:2105.16603:VOL:1705 Flashed Yes Yes No ESS:2105.16603:VOL:1703 2005/04/12 16:34:00 PDT IWNR2016I ESS:2105.22513:VOL:1300 ESS:2105.22513:VOL:1305 Flashed Yes Yes No ESS:2105.22513:VOL:1300 2005/04/12 16:34:00 PDT IWNR2016I ESS:2105.22513:VOL:1301 ESS:2105.22513:VOL:1304 Flashed Yes Yes No ESS:2105.22513:VOL:1301 2005/04/12 16:34:00 PDT IWNR2016I repcli> repcli> lspair -seq 'Remote Target' PPRC_800_to_F20 Source Target State Recov Shadow New Copyset Timestamp Last result ============================================================================================================================ ESS:2105.22513:VOL:1302 ESS:2105.16603:VOL:1707 Duplex Yes Yes No ESS:2105.22513:VOL:1302 n/a IWNR2011I ESS:2105.22513:VOL:1303 ESS:2105.16603:VOL:1708 SYNCHRONIZING No Yes Yes ESS:2105.22513:VOL:1303 n/a IWNR2011I
When all volume pairs of Continuous Synchronous Remote Copy session are synchronized, you should get results as shown in Example 15-8.
Example 15-8 Duplex state of Continuous Synchronous Remote Copy session
repcli> showsess PPRC_800_to_F20 Name PPRC_800_to_F20 Type pprc State Active Status normal Group PPRC_src Source Volumes 2 Approval Status Automatic Copysets 2 Non-approved 0 Invalid 0 Seq "Remote Target" Pool Criteria F20 Shadow Yes Recover Yes Err No Approve Yes Description Remote copy of 2 volumes from ESS 800 to F20 AWN007080I Command completed successfully. repcli> repcli> lsseq PPRC_800_to_F20 Name Recov Err Shadow Err Vols Recov Pairs Shadow Pairs Total Pairs Recov Timestamp ============================================================================================ Remote Target Yes No Yes 0 2 2 2 n/a repcli> repcli> lspair -seq 'Remote Target' PPRC_800_to_F20 Source Target State Recov Shadow New Copyset Timestamp Last result ===================================================================================================================== ESS:2105.22513:VOL:1302 ESS:2105.16603:VOL:1707 Duplex Yes Yes No ESS:2105.22513:VOL:1302 n/a IWNR2011I ESS:2105.22513:VOL:1303 ESS:2105.16603:VOL:1708 Duplex Yes Yes No ESS:2105.22513:VOL:1303 n/a IWNR2011I
891
suspendsess command
You can use suspendsess to suspend a Continuous Synchronous Remote Copy session. To restart a session, invoke the startsess command.
Note: To keep data consistency use -type consist parameter for suspendsess command suspendsess [ { -help|-h|-? } ] [-quiet] -type consist|immed session_name ... | -quiet
An optional parameter that turns off the confirmation prompt for this command.
-type consist | immed
Specifies the type of session to suspend. Specify consist to freeze a PPRC session, or specify immed (for immediately) to stop a session.
session_name [...] | -
Specifies the session name to be suspended. Separate multiple session names with a white space between each name. Alternatively, use the dash (-) to specify that input for this parameter comes from an input stream (STDIN). Example 15-9 shows the suspendsess command.
Example 15-9 suspendsess repcli> suspendsess -quiet -type consist PPRC_800_to_F20 AWN007140I Command completed successfully.
When a Continuous Synchronous Remote Copy session is suspended you should see results as shown in Example 15-10 from the lssess command. Notice that a session is recoverable and is not shadowing.
Example 15-10 lssess - suspended session repcli> lssess PPRC_800_to_F20 Name Status State Group Type Recover Shadow Err ============================================================== PPRC_800_to_F20 normal Active PPRC_src pprc Yes No No
lspair command
Invoke the lspair command to see that all volume pairs are suspended and time, when the session was frozen as shown in Example 15-11.
Example 15-11 lspair - suspended session
repcli> lspair -seq 'Remote Target' PPRC_800_to_F20 Source Target State Recov Shadow New Copyset Timestamp Last result ====================================================================================================================================== ESS:2105.22513:VOL:1302 ESS:2105.16603:VOL:1707 Suspended Yes No No ESS:2105.22513:VOL:1302 2005/04/12 19:43:25 PDT IWNR2015I ESS:2105.22513:VOL:1303 ESS:2105.16603:VOL:1708 Suspended Yes No No ESS:2105.22513:VOL:1303 2005/04/12 19:43:25 PDT IWNR2015I
892
stopflashsess command
You can use the stopflashsess command at any point during the life of a Point-in-Time Copy session once that session is in active state. This command withdraws all relationships between volumes on the storage subsystem. Example 15-12 shows an example of the stopflashsess command.
Example 15-12 stopflashsess repcli> stopflashsess -quiet FC_F20_800 AWN007150I Command completed successfully. repcli> lssess FC_F20_800 Name Status State Group Type Recover Shadow Err ========================================================== FC_F20_800 warning Defined FC_src flash No No No repcli> showsess FC_F20_800 Name FC_F20_800 Type flash State Defined Status warning Group FC_src Source Volumes 4 Approval Status Automatic Copysets 4 Non-approved 0 Invalid 0 Seq "Flashed Target" Pool Criteria P% Shadow No Recover No Err No Approve Yes Description Point in time copy of 4 volumes, 2 on ESS F20 and 2 on ESS 800 AWN007080I Command completed successfully.
stopsess command
To stop Continuous Synchronous Remote Copy session you can use the stopsess command at any point during the life of a session once that session is in active state. This command withdraws the relationship on the hardware.
stopsess [-quiet] session_name [. . .] -quiet
An optional parameter that turns off the confirmation prompt for this command.
session_name [...] | -
Specifies the session name to be stopped. Separate multiple session names with a white space between each name. Alternatively, use the dash (-) to specify that input for this parameter comes from an input stream (STDIN).
893
Output format
This section details the commands to display repcli commands.
setoutput command
You can use the setoutput command to display current output settings for repcli commands. The output format set by this command remains in effect for the duration of a command session or until the options are reset.
setoutput [ { -help|-h|-? } ] [-p on|off] [-r #] [-fmt default|xml|delim|stanza] [delim character] [-hdr on|off] [-v on|off]
? | h | help
Displays a detailed description of this command, including syntax, parameter descriptions, and examples. If you specify a help option, all other command options are ignored.
fmt
Specifies the format of the output. You can specify one of the following values:
default
Specifies that output is to be displayed in a tabular format using spaces as the delimiter between the columns. This is the default value.
894
delim
Specifies that output is to be displayed in a tabular format using the specified character to separate the columns. If you use a shell meta character (for example, * or \t) as the delimiting character, enclose the character in single quotation mark () or double quotation mark (). A blank space is not a valid character.
xml
delim character
Specifies character to separate the columns when -fmt delim parameter is used
p
Specifies whether to display one page of text at a time or all text at once.
off
Displays all text at one time. This is the default value when the perftool command is run in single-shot mode.
on
Displays one page of text at time. Pressing any key displays the next page. This is the default value when the repcli command is run in interactive mode.
hdr
Specifies the number of rows per page to display when the p parameter is on. The default is 24 rows. You can specify a value from 1 to 100.
v
Enables verbose mode. Example 15-14 shows the different formats of output.
Example 15-14 Default output settings repcli> setoutput Paging Rows Format Headers Verbose Banner ========================================== On 22 Default On Off Off
895
If you want to use an output format other than the default format only once per repcli session, use the setoutput command. There are output parameters for the commands:
lssess lspair lsseq
The syntax is the same as for the setoutput command. You can see the different formats of output for the lssess command in Example 15-15, Example 15-16, Example 15-17, and Example 15-18.
Example 15-15 default output format repcli> lssess PPRC_800_to_F20 Name Status State Group Type Recover Shadow Err ============================================================== PPRC_800_to_F20 normal Active PPRC_src pprc Yes Yes No
Example 15-16 XML output format repcli> lssess -fmt xml PPRC_800_to_F20 <IRETURNVALUE> <INSTANCE CLASSNAME="RM_Session"><PROPERTY NAME="session_name" TYPE="string"><VALUE TYPE="string">PPRC_800_to_F20</VALUE></PROPERTY><PROPERTY NAME="cptype" TYPE="string"><VALUE TYPE="string">pprc</VALUE></PROPERTY><PROPERTY NAME="state" TYPE="string"><VALUE TYPE="string">Active</VALUE></PROPERTY><PROPERTY NAME="status" TYPE="string"><VALUE TYPE="string">normal</VALUE></PROPERTY><PROPERTY NAME="srcgrp" TYPE="string"><VALUE TYPE="string">PPRC_src</VALUE></PROPERTY><PROPERTY NAME="shadow" TYPE="string"><VALUE TYPE="string">Yes</VALUE></PROPERTY><PROPERTY NAME="recov" TYPE="string"><VALUE TYPE="string">Yes</VALUE></PROPERTY><PROPERTY NAME="err" TYPE="string"><VALUE TYPE="string">No</VALUE></PROPERTY></INSTANCE> </IRETURNVALUE>
Example 15-17 stanza output format repcli> Name Status State Group Type Recover Shadow Err repcli> Name Status State Group Type Recover Shadow lssess -fmt stanza PPRC_800_to_F20 PPRC_800_to_F20 normal Active PPRC_src pprc Yes Yes No lssess -l -fmt stanza PPRC_800_to_F20 PPRC_800_to_F20 normal Active PPRC_src pprc Yes Yes
896
Err Approval Status Pool Criteria Copysets Non-approved Invalid Description Seq Source Volumes Approve
No Automatic F20 2 0 0 Remote copy of 2 volumes from ESS 800 to F20 "Remote Target" 2 Yes
Example 15-18 delim output format repcli> lssess -fmt delim -delim ',' PPRC_800_to_F20 Name,Status,State,Group,Type,Recover,Shadow,Err =============================================== PPRC_800_to_F20,normal,Active,PPRC_src,pprc,Yes,Yes,No
Note: Use lssess -l command instead of showsess in batch programs, because showsess shows results only in one, stanza format. See Example 15-19.
Example 15-19 is a sample lssess -l command which can be easily used in a batch program.
Example 15-19 Using lssess with -l (long) parameter in delim format
repcli> lssess -l -fmt delim PPRC_800_to_F20 Name,Status,State,Group,Type,Recover,Shadow,Err,Approval Status,Pool Criteria,Copysets,Non-approved,Invalid,Description,Seq,Source Volumes,Approve ================================================================================================================================================== PPRC_800_to_F20,normal,Active,PPRC_src,pprc,Yes,Yes,No,Automatic,F20,2,0,0,Remote copy of 2 volumes from ESS 800 to F20,"Remote Target",2,Yes
897
898
16
Chapter 16.
899
Router configuration
Configure the routers in the network to enable general multicasting or to allow multicasting for the SLP multicast address, port 239.255.255.253, and port 427. The routers of interest are those that are associated with subnets that contain one or more storage devices that are to be discovered and managed by Productivity Center for Disk. To configure your router hardware and software, refer to your router reference and configuration documentation.
Environment configuration
It may be advantageous to configure SLP DAs in the following environments: Where there are other non-Productivity Center for Disk SLP UAs that frequently perform discovery on the available services: This ensures that the existing SAs are not overwhelmed by too many service requests. Where there are many SLP SAs: A DA helps decrease network traffic that is generated by the multitude of service replies. It also ensures that all registered services can be discovered by a given UA. We particularly recommend the configuration of an SLP DA when there are more than 60 SAs that need to respond to any given multicast service request.
900
Here, myhost.com is the name of the server hosting the CIMOM, and port is the port number of the service, such as 5989.
Resource Manager
You can find the locations of the configured user ID for the Resource Manager in the files listed in Table 16-1.
Table 16-1 Resource Manager user ID and password
Component/server
Tivoli Agent Manager Productivity Center for Data Productivity Center for Fabric
File
...\AgentManager\config\Authorization.xml ...\TPC\Data\config\ep_manager.config ...\TPC\Fabric\manager\conf\AgentManager\config\endpoint.properties
The password of the Resource Manager is stored in the same files, but the password is in a readable format only on the Tivoli Agent Manager. The Resource Manager certificate subdirectory contains a file called pwd, which contains the agent registration password, which is required to open the certificate files.
901
Common Agent
The Common Agent does not have a user ID. Instead it has a context name which the agent uses for communication (see Table 16-2).
Table 16-2 Common Agent user ID and password
Component/Server
Tivoli Agent Manager Common Agent
File
...\WebSphere\AppServer\installedApps\<server>\AgentManager.ear\Agent Manager.war\WEB-INF\classes\resourcesl\AgentManager.properties ...\Tivoli\ep\config\endpoint.properties
AgentManager.properties has the context name and the password stored in clear text, so it is easy to find the values if you did not note them during the installation. On the Common Agent, the password is encrypted and stored in the pwd file in the same directory as the certification files, which are located in ...\Tivoli\ep\cert. If you go through the procedure of replacing the current certificates with a new one, do not forget to delete the pwd file, because it no longer matches the certificate file.
Ikeyman.exe
On all systems that have the one of the Tivoli Common Agent Services components installed, you will find a tool called ikeyman.exe. You can use this tool to open the certificate file, if you know the agent registration password. This is a quick way to verify that you still know the password that was used to lock a certificate file.
Message
BTACS0031I The Fabric Manager server is not registered with the Agent Manager BTACS0034I The Fabric Manager credentials are current
Meaning
Registration failed Fabric manager was restarted, Agent Manager could be contacted, everything is OK Attempt to register Registration was successful
BTACS0032I Registering with the Agent Manager at 9.1.38.36:9511. BTACS0037I Fabric Manager successfully registered with the Agent Manager. AgentManagerClient register
Open the DB2 control center, navigate to the tables view and open a table as shown in Figure 16-1.
In the table named IP_ADDRESS (Figure 16-2), you find the IP addresses of all registered Common Agents and Resource Managers.
903
904
4. In the SAN Configuration window, click the Agent Configuration tab. In the right pane, you see a table with the IP addresses, the names and the state of the Fabric Agents, similar to the example in Figure 16-4.
905
16.3 Launchpad
The TotalStorage Productivity Center comes with a Launchpad that is used to launch the interfaces to the individual managers. This includes the IBM Director for both Productivity Center for Disk and Productivity Center for Replication, Productivity Center for Data, and Productivity Center for Fabric. When you plan to install the remote management interfaces for the different managers you may also want to have to Launchpad installed on that machine. We explain the manual installation of the Launchpad. In 16.3.2, Launchpad customization on page 909, you see how you can integrate other applications into the Launchpad so that the Launchpad really becomes the central point for your storage management.
906
To install the Launchpad on a different machine, use the following procedure: 1. Start setup.exe. 2. After a few seconds, you see the Welcome display (Figure 16-7). Click Next.
907
3. In the window shown in Figure 16-8, enter the installation directory name and click Next.
4. The next panel (Figure 16-9) summarizes the installation information. Click Install.
5. The Launchpad installation takes less than five minutes. When it is completed, you see the installation status (see Figure 16-10). Click Finish to exit the installer.
908
6. Start the Launchpad by clicking the IBM TotalStorage Productivity Center desktop icon.
Note: The start of the Launchpad take some time. Be patient because there is no indication as to whether it really started.
Label
Manage Disk Performance and Replication Manage Disk Performance and Replication Manage Storage Network Fabric
Component
Productivity Center for Disk
Application
IBM Director
Configuration
No configuration necessary No configuration necessary No configuration necessary No configuration necessary Specify URL Specify URL
IBM Director
Manage File System and Database Utilization Manage Data Availability Automate Provisioning
909
Every time the Launchpad starts, it looks for installed TotalStorage Productivity Center components. If a component is found or if Manage Data Availability or Automate Provisioning is configured, TotalStorage Productivity Center changes the icon from the standard icon to an application specific icon. In addition, Productivity Center provides the capability for to add your own launchable application to the Productivity Center user interface. The readme file (readme.html), which is located in C:\Program Files\IBM\ProductivityCenter, contains information about how to integrate your own applications into the Launchpad in the Customizing the Productivity Center User Interface section. You need to provide a properties file and a launcher program. The launcher program is an executable program that the launchpad invokes that knows how to launch the management application. Launching may involve starting another executable or opening a Web page. The key to the support is the properties file specification. Each Productivity Center user interface application must have a separate properties file located in the executable path com/ibm/storage/launchpad/resources/extensionN.properties. In this path, N must be an integer value starting at 0. Any number of extension applications can be specified as long as sequential N values are used. That is, if there are four extension applications, the properties files must be named: extension0.properties extension1.properties extension2.properties extension3.properties The basic format of the content of the extension file is as follows:
[product] product.title
String to be used as the application title on the interface Name of graphic file to be used as the icon on the interface. The graphic file must be in a path relative to the execution path for the productivity center user interface executable. The main launchpad application uses a default icon if this is not specified or cannot be found. Program that is invoked when the extension application is selected. The application prompts the user for the launcher executable if this value is not specified or cannot be found.
product.description String to be used as the description of the product on the interface product.icon
product.exepath
Example 16-1 shows a sample extension file with the name extension0.properties.
Example 16-1 Example of a extension0.properties file [PartnerProduct] PartnerProduct.title = Vendor Extension 1 PartnerProduct.description = Description of the product. PartnerProduct.icon = ./images/otherProd1_16.gif PartnerProduct.exepath = LaunchTest
Attention: We used explorer.exe http://some.url.com as a Partner Product. At the time of this writing, we received an error message every time we tried to start the application. Figure 16-11 shows an example of that message. The error message was triggered as a result of explorer.exe always returning an error code of 9009. Nevertheless, the application was launched without any problems.
910
If the common agent is running, it listens for requests on that port and opens a connection. You simply see an empty screen. The common agent is not running if you see the message Connecting To 9.1.38.104...Could not open a connection to host on port 9510 : Connect failed.
911
Delete the data rows relating to old CIMOMs. Commit changes to DMCIMOM table. Locate the BASEENTITY table. Filter rows DISCRIM_BASEENTITY = DMCIMOM. Delete the data rows relating to the old CIMOMs. Commit changes to BASEENTITY table. Locate the DMREFERENCE table. Delete the data rows relating to the old CIMOM(s). Commit changes to DB2 table. The following screen captures illustrate the process. Before deleting a non-existing CIMOM through DB2 tables, you should first delete any storage devices that are associated with the CIMOM in TotalStorage Productivity Center. Right-click the device and select Delete as shown in Figure 16-12.
912
Launch DB2 Control Center (Figure 16-13). This is a general administration tool for managing DB2 databases and tables.
Attention: DB2 Control Center is a database administration tool. It gives you direct and complete access to the data stored in all the TotalStorage Productivity Center databases. Altering data through this tool can cause damage to the TotalStorage Productivity Center environment. Be careful not to alter data unnecessarily using this tool.
Navigate down the structure in the left hand panel to open up the DMCOSERV database, then click on the Tables option. A list of tables for this database will appear in the right hand upper panel as seen in Figure 16-14.
913
Locate the DMCIMOM table and double click to open a new window (Figure 16-15) showing the data rows. Identify the CIMOM rows to be deleted by their IP address.
Click once on the row to be delete to select it. Click on the Delete Row button to remove it from the table. When you have made your changes you must click the Commit button for the table changes to be made effective. Click Close to finish with this table.
914
If you make any mistakes before you have clicked the Commit button, you can click the Roll Back button to undo the changes. Next locate the BASEENTITY table from Control Center panel as seen in Figure 16-14 on page 914. Open it with a double click. This table contains many rows of data. Filter the data to show only entries that relate to CIMOMs (see Figure 16-16). Click the Filter button to open the filter panel as shown in Figure 16-17.
Enter DMCIMOM in the values field as shown in Figure 16-17 and click OK.
915
The table data is now filtered to show only CIMOM entries as shown in Figure 16-18.
Use a single click to select the entries by IP address that relate to the non-existent CIMOMs. Click Delete Row to remove them. Click Commit to make the changes effective, then Close. You can used Roll Back to undo any mistakes before a Commit. Now locate the DMREFERENCE table from Control Center panel as seen in Figure 16-14 on page 914. Open it with a double click (see Figure 16-19).
Note: The DMREFERENCE table may contain more than one entry for each of the non-existent CIMOMs. It may not contain any rows at all for the CIMOMs. Delete all relevant rows for the non-existent CIMOMs if they exist. If there are no rows in this table for the CIMOMs you are deleting they are not linked to any devices and this is OK.
916
917
In our case, note the large quantity of User ID/Password Incorrect from Server on CIMOM messages. For us, these messages were indicative of the CIM agents (not ours) that resided in the local subnet which were not configured to provide data to our TotalStorage Productivity Center for Disk administrative username and password (superuser/password). The Director console reports back that it has been rejected from accessing a CIM agent.
918
919
920
921
Figure 16-25 is an example of the raswatch during the trace of TotalStorage Productivity Center discovery.
The raswatch output can be very verbose and will scroll very quickly off the screen. Consider logging the output into a file using raswatch -dev_mgr -high > c:/testlog.txt. This will allow you to open the raswatch file in the Notepad editor and search for IP or hostname strings that will validate the TotalStorage Productivity Center discovery process.
922
923
Additional information about viewing and managing events logged in the DB2 journal can be referenced by accessing the help menus. By default, DB2 instance health checking is disabled. It may be advisable to enable the health monitor. Even the default alert thresholds that are in place when the health monitor is enabled will provide for the TotalStorage Productivity Center administrator at least some idea of any issues with the DB2 instance. You can open the DB2 Health Center by clicking Start Programs IBM DB2 Monitoring Tools Health Center (Figure 16-27).
924
Remember that by default, the Health Center monitoring is disabled. You can tell by the green circle on our DB2 instance and associated databases that we have enabled monitoring. At present we have no issues that have generated any alerts. Figure 16-28 shows the typical default threshold settings for each of the TotalStorage Productivity Center for Disk databases.
925
Aside from viewing the DB2 events, or ensuring that event monitoring is enabled, we can also review the contents of specific database tables to ensure that we are receiving data and that the appropriate tablespaces are being populated. For example, if we have just performed an ESS data collection task, we should have entries in the following three tables in the PMDATA database: VPCCH - Volume data VPCRK - Array data VPCLUS - Cluster data We can review the contents of these tables from within the DB2 Control Center, by navigating through the tree on the left-hand side, from system, instance, to the database, and tables, as shown in Figure 16-29.
926
To view the contents of a specific table, right-click the table you want to view and select Sample Contents. You should see a table like the one in Figure 16-30.
The presence of these rows in this table tells us that we have successfully performed a data collection task against the ESS with serial number 22219. We should also see data in the other tables cited above.
927
A considerable amount of application level information can be obtained from within the IBM WebSphere framework, using the Administrative Console.
928
For SVC, the TotalStorage Productivity Center for Disk superuser name and password must be synchronized by creating the same account credentials in the SAN Volume Controller Console GUI. 5. Follow the TotalStorage Productivity Center for Disk discovery process using raswatch (Figure 16-25 on page 922) for evidence that the new CIM: a. Has been detected b. Has allowed TotalStorage Productivity Center for Disk to authenticate correctly The same activities may be traced in the IBM Director logfiles (Figure 16-20 on page 917).
#---------------------------------------------------------------------------# Tracing and Logging #---------------------------------------------------------------------------# A boolean controlling printing of messages about traffic with DAs. # Default is false. ;net.slp.traceDATraffic = true # A boolean controlling printing of details on SLP messages. The fields in # all incoming messages and outgoing replies are printed. ;net.slp.traceMsg = true # A boolean controlling printing details when a SLP message is dropped for # any reason. Default is false. Default is false.
;net.slp.traceDrop = true # A boolean controlling dumps of all registered services upon registration # and deregistration. If true, the contents of the DA or SA server are Default is false.
The more detailed output from SLP tracing is in the slp logfile, located in c:/WINNT/slpd.log.
929
Important: After the required tracing information has been gathered, you should disable slp trace. The logfile can become very large.
930
Seeing trace entries of this type shows that Replication Manager is receiving indications properly. If Replication Manager is not receiving indications, then no entries of this type will be seen surrounding the event in question. If Replication Manager is not seeing indications, then usually one or more layers of the software stack need to be restarted.
931
This will redirect the browser to the secure login page, and afterward the login goes to the WebSphere Application Server Administrative root page (see Figure 16-33).
932
Figure 16-34 WebSphere Application Server trace tool - select servers example
933
Figure 16-35 WebSphere Application Server trace tool - select application servers example
934
Figure 16-36 WebSphere Application Server trace tool - select server 1 example
935
Figure 16-37 WebSphere Application Server trace tool - select logging and tracing example
936
Figure 16-38 WebSphere Application Server trace tool - select diagnostic trace example
7. Check the enable trace box: Enter the required trace entries into the trace specification box, separated by colons. Insert all of the trace specifications in this table which might be used byTotalStorage Productivity Center.
937
Table 16-5 provides the default TotalStorage Productivity Center for Replication specifications (see Figure 16-39 for the trace).
Table 16-5 Default trace specifications
Component
general format
Default
Comp=level=state where: Comp is the component to trace Level is the amount of trace *State is enabled or disabled ELEMCAT=all=enabled HWLAYER=all=enabled REPMGR=all=enabled DMINT=all=enabled
Replication Manager Element Catalog Replication Manager Hardware layer Replication Manager Session Manager Replication Manager integration with Device Manager
*This is the value which should be set all the time, unless otherwise specified. For TotalStorage Productivity Center for Replication, the full setting is:
REPMGR=all=enabled:HWLAYER=all=enabled:DMINT=all=enabled:ELEMCAT=all=enabled
The remaining settings on this page are used to control how much trace is captured before it is overwritten. The actual settings to use depend on the actual server configuration, but here are some guidelines to use: Always choose the setting which sends the trace to a file. 20 MB seems to be a good size to use for each trace file. Enable at least one historical file. The more history is available, the better, since many TotalStorage Productivity Center are long-running and may result in a lot of trace data. Be sure to leave sufficient free space. The total trace will take up the number of historical files plus 1 multiplied by the size of each file.
Recommended settings:
20 MB per file 10 historical files (unless there is not enough disk space on the server)
Tip: The default file name is ${SERVER_LOG_ROOT}/trace.log and it is best to keep this default whenever possible. This will ensure that the automated tools to collect log and trace information can find the files. If the log files need to be written to a different location, for example to a different disk to manage free disk space, it is better to change the environment variable SERVER_LOG_ROOT. Refer to the WebSphere Application Server documentation for information about how to change this environment variable.
938
Figure 16-39 is a sample window showing several values have been changed.
Figure 16-39 WebSphere Application Server trace tool - several trace values changed
8. After making all changes, press the OK button, and then click save the changes.
Tip: To change trace settings immediately without restarting WebSphere Application Server, make the equivalent change in the Runtime tab instead of the Configuration tab. When Apply or OK is clicked from the Runtime tab, the change takes effect immediately.
939
3. Copy mdmpmconsole.jar to the same directory as previous step. 4. Restart the IBM Director server from the TotalStorage Productivity Center for Disk Server Start Menu Control Panel Administrative tools Services IBM Director Server. 5. You may need to wait for some time till the IBM Director has re-started.
You will be requested to confirm whether to stop the task. You may respond Y (yes).
940
IBM TotalStorage Productivity Center V2.3: Getting Started
Alternatively, you may launch the SVC Console Web browser interface. After logging into the SVC console, choose Clusters under My Work column and click in the check box for the respective SVC cluster in the Clusters column and click Go. In the next panel select Manage Cluster. You will see a panel similar to Figure 16-41.Choose Stop Statistics Collection.
941
Click Yes. This will stop all the performance data collection for SVC.
942
17
Chapter 17.
943
The sizing function on this panel shows used space and free space in the database. Space status advisor: The Space status advisor monitors the amount of space used by the Performance Manager database and advises you as to whether you should purge data. The advisor levels are: Low: You do not need to purge data now High: You should purge data soon Critical: You need to purge data now Disk space thresholds for status categories: The thresholds are as follows: Low if utilization <0.8 High if 0.8 <= utilization <0.9 Critical otherwise The delimiters between low/high/critical are 80% and 90% full.
Database purging:
You use the Performance Manager database panel to specify properties for a performance database purge task. You can purge performance data based on the age of the data, the type of data, and the storage devices associated with the data. After you specify the database purge information, it is saved as a noninteractive IBM Director task. You schedule all performance data-collection tasks using the IBM Director scheduler function.
944
945
The current database information is shown. Use this panel to specify the properties for a new performance database purge task. The fields are:
Name: Type a name for the performance database purge task, from 1 - 250 characters. Description (optional): Type a description for the performance database purge task, from 1 to 250 characters. Device type: Select one or more storage device types for the performance database purge. Purge performance data older than: Select the maximum number of days or the number of years that you want the performance data to reside in the database before it is purged. Purge data containing threshold exception information: When you select this check box, you choose to purge exception data.
946
Save as task: When you click Save as task, the information you specified is saved and the panel closes. The newly created task is saved as a noninteractive task to the IBM Director Task pane under the Performance Manager Database. All performance database tasks can be scheduled using the IBM Director scheduler function as seen in Figure 17-3.
Right-click the newly created database purge task (Figure 17-3) to schedule it for execution. Execution is either immediate or scheduled as seen in Figure 17-4.
947
The following main menu options are available to you for use with the TotalStorage Productivity Center databases or any other DB2 database instance you may have on your TotalStorage Productivity Center server. We will be putting more emphasis on the Command Line Tools in a TotalStorage Productivity Center reporting framework. Command Line Tools Development Tools General Administration Tools Information Monitoring Tools Set-up Tools
Tip: For detailed information, use the DB2 Tool Suite Help Screens or DB2 Online Information at the following URL:
http://publib.boulder.ibm.com/infocenter/db2help/index.jsp
948
Before accessing a database, you must perform preliminary tasks, such as starting DB2 with START DATABASE MANAGER. You must also connect to a database before it can be queried. Connect to a database by doing one of the following: Issue the SQL CONNECT TO database statement (see Figure 17-5 or Data Extraction using DB2 Command Line Processor Interface on page 981). Establish an implicit connection to the default database defined by the environment variable DB2DBDFT. If a command exceeds the character limit allowed at the command prompt, a backslash (\) can be used as the line continuation character. When the command line processor encounters the line continuation character, it reads the next line and concatenates the characters contained on both lines. Alternatively, the -t option can be used to set a line termination character. In this case, the line continuation character is invalid, and all statements and commands must end with the line termination character. For more information, use the DB2 UDB online help. In current releases of DB2 UDB CLP starts in the interactive mode, which is indicated by a DOS-looking command prompt db2=>. In this mode, end-users may enter one DB2 UDB command or one SQL statement by typing it in at the prompt, and then pressing the enter key. Figure 17-5 shows an example query in an IBM DB2 Command Line Processor window.
In this example, a connect DB2 UDB command was executed to connect to the TotalStorage Productivity Center Performance Manager database named PMDATA (the TotalStorage Productivity Center performance database alias). When this command is executed, you can then enter a SELECT SQL statement against any of the PMDATA tables in the database. The commands are not case sensitive but the user ID (MDMSUID) and password (MDMSPW) are case sensitive based on how they were defined in the database setup during installation or thereafter. The interactive mode is exited by typing QUIT and pressing enter. The DB2 UDB Tool Suite also has another CLP which operates in a non-interactive mode; Command Window. It may be opened up from the path:
Start IBM DB2 Command Line Tools Command Window
949
The SQL queries are invoked by starting each SQL statement with the characters db2, for example db2 connect to pmdata. This CLP has the same case sensitivity requirements of the Command Line Processor. For additional examples of the Command Line Processor, refer to Data Extraction using DB2 Command Line Processor Interface on page 981.
The Development Tools options are Development Center Project Deployment Tools DB2 Development Center provides an easy-to-use development environment for creating, installing, and testing stored procedures. It allows you to focus on creating your stored procedure logic rather than the details of registering, building, and installing stored procedures on a DB2 server. Additionally, with Development Center, you can develop stored procedures on one operating system and build them on other server operating systems. Development Center is a graphical application that supports rapid development. Using Development Center, you can perform the following tasks: Create new stored procedures. Build stored procedures on local and remote DB2 servers. Modify and rebuild existing stored procedures. Test and debug the execution of installed stored procedures.
Control Center
A GUI for snapshot and event monitoring. For snapshots, it allows you to define performance variables in terms of the metrics returned by the database system monitor and graph them over time. For example, you can request that it take a snapshot and graph the progression of a performance variable over the last eight hours. Alerts can be set to notify the DBA when certain threshold are reached. For event monitors, it allows you to create, activate, start, stop, and delete event monitors. See the online help for the Control Center for more information (also see Control Center on page 976).
Journal
You can start the Journal by selecting the icon from the Control Center toolbar. The Journal allows you to monitor jobs and review results. From the Journal, you can also display the recovery history and DB2 messages. The Journal allows you to monitor pending jobs, running jobs, and job histories; review results; display recovery history and alert messages; and show the log of DB2 messages.
950
Replication Center
The Replication Center stores the initial information about registered sources, subscription sets, and alert conditions in the control tables. The Capture program, the Apply program, and the Capture triggers update the control tables to indicate the progress of replication and to coordinate the processing of changes. The Replication Alert Monitor reads the control tables that have been updated by the Capture program, Apply program, and the Capture triggers to understand the problems and progress at a server.
Task Center
Use the Task Center to create, schedule, and run tasks. You can create the following types of tasks: DB2 scripts that contain DB2 commands OS scripts, which has operating system commands MVS shell scripts to run on OS/390 and z/OS operating systems JCL scripts to run in a host environment Grouping tasks, which contain other tasks Task schedules are managed by a scheduler, while the tasks are run on one or more systems, called run systems. You define the conditions for a task to fail or succeed with a success code set. Based on the success or failure of a task, or group of tasks, you can run additional tasks, disable scheduled tasks, and other actions.
Tip: You can also define notifications to send after a task completes. You can send an e-mail notification to people in your contacts list, or you can send a notification to the Journal.
The DB2 Monitoring Tools options are: Event Analyzer Health Center Indoubt Transaction Manager Memory Visualizer
Event Analyzer
The Event Analyzer GUI is sued for viewing file event monitor traces. Information collected on connections, deadlocks, overflows, transactions, statements, and subsections is organized and displayed in a tabular format. See the online help for the Event Analyzer for more information.
Health Center
Use the Health Center GUI tool to set up thresholds that, when exceeded, will prompt alert notifications, or even actions to relieve the situation. In other words, you can have the database manage itself!
951
952
Tip: Alternatively, if the Control Center is open, click the Center opens.
You can use the toolbar icons (see Figure 17-7) to open DB2 tools, view the legend for Command Center objects, and view DB2 information.
Execute
Executes the SQL statements, DB2 CLP commands, scripts, or MVS system commands that you enter on the Interactive or Script page. The results are displayed on the Query Results and the Access Plan pages.
953
Control Center
Opens the Control Center so that you can display all of your systems, databases, and database objects and perform administration tasks on them.
Replication Center
Opens the Replication Center so that you can design your replication environment and set up your replication environment.
Opens the Satellite Administration Center so that you can set up and administer satellites and the information that is maintained in the satellite control tables.
Opens the Data Warehouse Center so that you can manage Data Warehouse objects.
Task Center
Opens the Task Center so that you can create, schedule, and execute tasks.
Opens the Information Catalog Center so that you can manage your business metadata.
Health Center
Opens the Health Center so that you can work with alerts generated while using DB2.
Journal
Opens the Journal so that you can schedule jobs that are to run unattended and view notification log entries.
License Center
Opens the License Center so that you can display license status and usage information for the DB2 products installed on your system and use the License Center to configure your system for license monitoring.
954
Development Center
Opens the Development Center so that you can develop stored procedures, user-defined functions, and structured types.
Contacts
Opens the Contacts window where you can specify contact information for individual names or groups.
Tools Settings
Opens the Tools Settings notebook so that you can customize settings and properties for the administration tools and for replication tasks.
Legend
Opens the Legend window that displays all of the object icons available in the Command Center by icon and name.
Retrieves the data for the table you have executed SQL statements against and displays it on the Query Results page.
Creates the access plan for the current SQL statement and displays it on the Access Plan page.
Information Center
Opens the Information Center so that you can search for help on tasks, commands, and information in the DB2 library.
Help
955
Next, display three columns with Date over the left column, Cluster 1 over the middle column, and Cluster 2 over the right column Sort by date/time in the left column (VPCLUS:PC_DEV_DATE_E/PC_DEV_TIME_E). Left column is keyed by VPCLUS:P_TASK VPCLUS:M_MACH_SN VPCLUS:PC_DEV_DATE_E/PC_DEV_TIME_E
Under each cluster column, display sub-column headers of I/O Rate, Avg Cache Hold Time, and NVS % Full; Center and right columns are keyed by VPCLUS:PC_DEV_DATE_E/PC_DEV_TIME_E VPCLUS:M_CLUSTER_N
I/O rate, (VPCLUS:Q_CL_IO_RATE) Average cache hold time, (VPCLUS:Q_CL_AVG_HOLD_TIME) NVS % full, (VPCLUS:Q_CL_NVS_FULL_PRCT)
Under the center and right columns, display rows with the following:
2. Select a cluster from step 1 to investigate further. 3. Then, build a report broken by Logical SubSystem (LSS) displaying the following information in the header to reflect the row data;
Type, M_MACH_TY Model, M_MODEL_N Serial, M_MACH_SN Cluster, M_CLUSTER_N from-date/time, PC_DATE_B/PC_TIME_B to-date/time, PC_DATE_E/PC_TIME_E
Sort by date/time and sub-sort by Device Adapter number (DA #) PC_DEV_DATE_B/PC_DEV_TIME_B PC_DEV_DATE_E/PC_DEV_TIME_E M_CARD_NUM
DA ID #, M_CARD_NUM Loop A or B, M_LOOP_ID Array ID, M_ARRAY_ID Array Type, M_STOR_TYPE Average ms to satisfy all requests to this array, PC_IOR_AVG % Time Array Busy, Q_SAMP_DEV_UTIL Total I/O read/writes to this array, Q_IO_TOTAL Total sequential read/writes to this array, Q_IO_SEQ
4. Select an LSS from step 2 to investigate further. 5. Then, build a report broken by loop displaying the following row information with the corresponding column headers;
Type, M_MACH_TY Model, M_MODEL_N
957
Serial, M_MACH_SN Cluster, M_CLUSTER_N LSS, M_LSS_LA DA #, M_CARD_NUM Loop, M_LOOP_ID from-date/time,PC_DATE_B/PC_TIME_B to-date/time, PC_DATE_E/PC_TIME_E
6. Select an array from step 3 to investigate further. 7. Build a report broken by volume displaying the following row data with the corresponding column names;
Type Model Serial Cluster LSS DA# Loop Array from-date/time to-date/time
M_MACH_TY M_MODEL_N M_MACH_SN M_CLUSTER_N M_LSS_LA M_CARD_NUM M_LOOP_ID M_ARRAY_ID PC_DATE_B/PC_TIME_B PC_DATE_E/PC_TIME_E
958
Tip: The proceeding global to granular reporting sequence example can be achieved through the DB2 Tool Suite Command Center (for instance). The individual component reports can be scheduled in the DB2 Tool Suite and the output parsed and compiled into a spreadsheet format (table data output exported in .WKS format for example) such as Lotus 123 or Excel.
Once the data is formatted into the worksheet, the native macro functionality of the spreadsheet could be used to process the data further to output graphical reports, summary reports, problem analysis document for root cause analysis, and performance analysis of SAN components, in order to meet the particular needs of your organization and SAN environment.
959
2. Using the Interactive option of the Command Center, click that button in the middle left of the window (see Figure 17-8). This will give you the option of Executing single SQL statements and DB2 CLP or executing MVS system commands. You also have the option to run an existing command script or one that you create in this example.
Figure 17-8 Open Command Center window select Interactive button example
960
3. Utilize performance data collected from your storage servers using the PM database tables (PMDTATA). With the Command Center window open, click the Database Connection button ..... to the right of the Database Connection bar. The Select Database window opens. Figure 17-9 shows an example of selecting the PM database tables (PMDATA) in the Command Center.
4. Once you have selected the database you want to work with, you can use the previously described functions to manage information within the database, extract data, or set up SQL queries using the Interactive or SQL Assist options. For our example, we use only the PMDATA database information. You could just as easily utilize the DM (DMCOSERV database) data to retrieve related information (such as Asset data) or the IBMDIR database to manage information in the IBM Director tables.. We will proceed to run a base report against a specific Model 800 ESS.
Note: We can also run this report against all similar storage server types for which we currently have performance previously collected and stored in the PMDATA database. If we preferred to do this, we would include specific, or all, M_MACH_SN (and their associated) values available in the database.
961
5. We have connected to the PMDATA database and will utilize the SQL Assist function within the Command Center for our query example. Click the SQL Assist button to begin our SQL query script definition (see Figure 17-10).
Note: Use SQL Assist to create SQL statements. With SQL Assist and some knowledge of SQL, you can create SQL SELECT statements. In some environments, you can also use SQL Assist to create INSERT, UPDATE, or DELETE statements. SQL Assist is a tool that uses an outline and details panels to help you organize the information that you need to create an SQL statement.
SQL Assist is available in the Control Center, the Command Center, the Replication Center, and the Development Center. See the Online Help for more information. SQL Assist and other functions within the DB2 Tool Suite incorporate button sensitive (mouse over) help pop-up windows to aid you in navigating and making your menu selections within the tools.
962
6. The SQL Assist window will now open (see Figure 17-11). Click the Select radio button, in the middle right area of the window, since we are only going to retrieve data from the tables with our database queries in this example. The radio button options available in this window include the SQL query options; Select, Insert, Update, Delete. It is not recommended to do any SQL commands on your production database other than Select statements since with this function, you are only reading data from the database. If you want to manipulate your database further, it is recommended to make a copy of the database (not the production backup) and work with the backup.
963
7. Notice that in the lower pane of the SQL assist window, there is the initial syntax of a select statement. We are now going to go through the steps to complete a comprehensive SQL Select query statement to view (or extract) data from our PMDATA database. In the upper left-hand pane called Outline, double-click the FROM (Source Tables) icon and the Details pane will open in the center of the SQL Assist window. The Available Tables tree is listed. Select DB2ADMIN to find the tables that you want to use in our query (see Figure 17-12).
Figure 17-12 Selecting the DB2ADMIN table button pull-down listing of available tables
964
8. We will now select the VPVPD table which contains a storage server configuration snapshot. Use the slider on the Available tables pane until you can click on the VPVPD table. Now click the > button and the table name will be populated into the Selected Source Tables in the upper right corner of the SQL Assist window (see Figure 17-13). Notice how the VPVPD table selection has now been entered automatically into the rudimentary Select statement in the lower pane of the window. This window will show you the SQL validated statement you have created thus far and keep track as you proceed through the SQL Assist function.
965
9. Now that you have selected the table you want to query for data, click the SELECT (Result Columns) icon in the Outline pane on the left of the window. The DB2ADMIN.VPVPD (Instance.Table) icon will appear in the pane. Select the + button in this icon and a list of the VPVPD columns will list in the table tree column listing (see Figure 17-14).
966
10.Now select the VPVPD columns we want in our view with our SQL statement. You can select the columns in several ways. One way is to select the columns one at a time, or by clicking on the column name, selecting multiple column names by clicking the column names while holding the shift key down, or you can select all the columns by clicking the >> button. In our example we elected the columns we want while holding down the shift key and clicking the column names. Now click the > button to populate the Result columns pane (which is greyed out until you make your selections. Once you click the > button, a moment will pass and the column names will appear in the right-hand pane and in the SQL validated query statement that is building in the lower SQL Assist pane (see Figure 17-15). User defined field variables are found in the SQL validated statement.
967
11.Next click the Where (Row filter) icon in the outline pane. This will present a table list in the Available columns pane from which you can select where, and how, we want to filter the query (see Figure 17-16).
Figure 17-16 Where (Row filter) for column M_MACH_SN values (note mouse over help)
968
12.Now define ther statement to return results where the M_MACH_SN value (ESS serial number) equals 2105.22219. Place the cursor in the Value field. You can either type a specific value in or use the pull-down arrow to being up other options. One of the options presents you the opportunity to see field results already in the table. This will open a subsequent screen showing current column results. You can select from that screen how many results to display. The default is to show 25 rows. You can increase of decrease this value through the menu. After you have made your selection for the Value, click the > button to enter the value into the Search Condition pane and in the SQL validated pane (see Figure 17-17).
969
13.You will not be using the Group By or Having SQL query functions for this simple query example. See the help screens for further information in for those query options. Now click Order By (Sort Criteria) in the outline pane. In the Available columns pane is the VPVPD table content tree (column names). Click value P_CDATE (performance collection date). Now click the > button and the column name appears in the Sort Columns pane. You have the option to select ASC (ascending) or DSC (descending) sort order. ASC is setup as a default sort order. Leave the ASC as-is (see Figure 17-18).
14.Now that you have completed building a query statement, click the Run button to view the results (see Figure 17-19).
970
15.After reviewing the results of query, click the OK button to return the SQL code to the main Command Center, Interactive window. Notice the mouse over, pop-up window (see Figure 17-20). The SQL code from this example is shown in Example 17-1.
Example 17-1 SQL code sample SELECT VPVPD.M_MACH_SN, VPVPD.M_MACH_TY, VPVPD.M_MODEL_N, VPVPD.M_CLUSTER_N, VPVPD.M_RAM, VPVPD.M_NVS, VPVPD.P_CDATE, VPVPD.P_CTIME FROM DB2ADMIN.VPVPD AS VPVPD WHERE VPVPD.M_MACH_SN = '2105.22219 ' ORDER BY VPVPD.P_CDATE ASC
Figure 17-20 Return SQL code we created to the Command Center, Interactive window
971
16.You can now save your SQL code for future use. You can schedule this as a recurring task or use it ad hoc. Click the Interactive tab on the menu bar at the top of the window and then click Save Command As... (see Figure 17-21).
Figure 17-21 Save Command As... an ascii file for later use
972
973
2. Drill down with the next query for the suspect time period to see which arrays were involved. This could be done with any of the table data you want to examine that have data associated with them. The SQL query is shown in Example 17-4. The query result is shown in Figure 17-24.
Example 17-4 VPCRK SQL query for specific time period SELECT VPCRK.M_MACH_SN, VPCRK.PC_DEV_DATE_E, VPCRK.PC_DEV_TIME_E, VPCRK.M_LSS_LA, VPCRK.M_ARRAY_ID, VPCRK.M_DDM_NUM, VPCRK.M_CARD_NUM, VPCCH.M_VOL_NUM, VPCCH.M_VOL_ADDR, VPCCH.M_VOL_TY, VPCRK.PC_MSR_AVG FROM DB2ADMIN.VPCRK AS VPCRK, DB2ADMIN.VPCCH AS VPCCH WHERE VPCRK.M_MACH_SN LIKE '%2105%' AND VPCRK.M_MACH_SN LIKE '%2105%' AND VPCRK.PC_DEV_DATE_E = '2004-06-08' ORDER BY VPCRK.PC_DEV_DATE_E ASC, VPCRK.Q_SAMP_DEV_UTIL DESC, VPCRK.PC_MSR_AVG DESC
974
From the information above, you can determine the date, time of day, rank number, volume number, and volume address for the time period we examined from the previous query. You know that all the hits for this report are indicating the volumes are OS/390 assigned storage (value C in the M_VOL_TYPE column).
Tip: You could save this query and run it as a scheduled task from the DB2 Tool Suite. You could also export the data for further manipulation and presentation with the use of a spreadsheet application. You could also setup SQL query tasks to run on a schedule, using the TotalStorage Productivity Center gauge reports to determine which areas need further investigation. The information derived from these queries of the PM database tables will correlate with the information you can derive from the ESS Specialist so you can determine which hosts and associated applications are causing performance concerns.
For further information and examples of SQL queries, you can refer to the redbook IBM TotalStorage Expert Reporting: How to Produce Built-In and Customized Reports, SG24-7016.
975
976
Tip: Within the DB Tool Suite, there are context sensitive pop-up help windows to aid you in navigating through the menu selections and tasks.
Important: Never make any modifications to your existing TotalStorage Productivity Center database. If you want to learn and experiment, create a new database or export your production database and perform operations on the exported database only.
Once you have opened the DB2 UDB Control Center, you can drill down to your TotalStorage Productivity Center database (PMDATA in this example) by using the explorer window on the left-hand side of the window (Figure 17-26).
On the right-hand side of the Control Center main window you can view the tables of the PMDATA database (since the Tables folder is highlighted on the left-hand side). In Figure 17-27 we explore the VTHRESHOLD table further by viewing the columns. This is done by double-clicking on the particular table you want to view details on. We have selected the Columns tab at the top left side of the window. The column attributes are listed under the window column headers. You can also view the tables Keys, Check Constraints, or General Table attributes.
977
From this window, you can further explore the table by selecting the tabs on the upper portion of the window. We will now view the Primary Key(s) window for the CNODE table (Figure 17-28). This is very useful information when you are creating your own query statements. It will reduce the amount of research time spent digging through hardcopy documentation.
You can also show any current SQL statement you are generating within the Control Center, Estimate the size of this table and determine the current size or add unique or foreign key associations from this window. Additional database information is available from within the Control Center and useful help is available as needed.
978
Reporting tools
In this section we outline some of the processes for getting the most out of the TotalStorage Productivity Center database using applications and tools outside of the TotalStorage Productivity Center product. This is not an exhaustive list, but these are things to keep in mind when you are trying to make decisions with respect to exporting data, creating, and disseminating custom reports: IBM DB2 Express (or any version 8 full-featured IBM DB2 product) is required on the query system (laptop, desktop) Portable programming language and other tools for data extraction/parsing REXX C and C++ (can be compiled and disseminated in AIX environment) ESSCLI (asset/capacity data only) CLI Python QMF
Spreadsheet application Microsoft Excel IBM Lotus 123 Quick print-to-screen reports Parsed and formatted SQL query output Data Output types Data to files (compressed and uncompressed) Binary, ASCII, etc.
979
Integrated replication administration. DDL statements to easily create, drop, and alter data source mappings, users, data types and functions (user-defined and built-in). Excellent performance and intelligent use of pushdown and remote query caching. Refer to the following Web site for more information about IBM DataJoiner:
http://www.ibm.com/software/data/datajoiner/
You can download the free QMF for Windows Try and Buy version from the following Web site:
http://www-3.ibm.com/software/data/qmf/reporter/june98/downloads.html
980
There are numerous other tools and applications such as IBM DB2 Intelligent Miner, IBM Object REXX, LotusScript which contain powerful scripting and/or report formatting capabilities and can access DB2 UDB on UNIX or Intel, IBM ~ iSeries, z/OS, as well as any database manager connected to DataJoiner. Refer to following Web sites for more information about these other tools: DB2 Intelligent Miner:
http://www.ibm.com/software/data/iminer
Object REXX:
http://www.ibm.com/software/awdtools/obj-rexx/
LotusScript:
http://www.ibm.com/software/data/db2/db2lotus/db2lscpt.htm
There are no direct solutions to print the built-in reports or save the report files directly from TotalStorage Productivity Center. However, you can issue standard SQL statements to extract the data. All asset, capacity, and performance data is available in the form of DB2 tables. DB2 UDB management tools will be useful in utilizing your table data in the most efficient manner.
981
Exports, imports, or loads user-created tables. If a database is to be duplicated from one operating system to another operating system, db2move only helps you to move the tables. You also need to move all other objects associated with the tables, such as: aliases, views, triggers, user-defined functions, and so on. db2look is another DB2 UDB tool to help you easily move some of these objects by extracting the Data Definition Language (DDL) statements from the database. When export, import, or load APIs are called by db2move, the FileTypeMod parameter is set to lobsinfile. That is, LOB data is kept in separate files from PC/IXF files. There are 26 000 file names available for LOB files. The lLOAD action must be run locally on the machine where the database and the data file reside. When the load API is called by db2move, the CopyTargetList parameter is set to NULL; that is, no copying is done. If logretain is on, the load operation cannot be rolled forward later. The tablespace where the loaded tables reside is placed in backup pending state, and is not accessible. A full database backup, or a tablespace backup, is required to take the tablespace out of backup pending state.
2. Create a folder to for the output files of the data extract. For example,
mkdir c:\ibmout
(Windows 2000 command window, or use Windows Explorer to create new folder) 3. Start Programs IBM DB2 Command Line Tools click Command Line Processor (CLP) 4. Connect to the DB2 using the following command in the CLP window
connect to pmdata user db2admin using db2admin Note: The screen response should be a few lines which says you are connected and the level of DB2 is 8.1.2. You should replace the word db2admin (following the word using) with the actual password you are using for your database instance).
982
5. Issue the following command to extract the data from the VPVPD table substituting the folder name with one of your choice (VPVPD table contains cluster-level and storage server-level configuration data; generated at the start of Performance Data Collection) and replacing the mmdd with the date of the extract. Export VPVPD table data to file
export to c:\ibmout\vpvpdmmdd.txt of del select * from vpvpd
6. Issue the following command to extract a specific day's worth of data from the VPCRK table (logical array-level performance data) substituting the date to be extracted and the same substitution for the filename. VPCRK discrete table output directed to file
export to c:\ibmout\vpcrkmmdd.txt of del select * from vpcrk where pc_date_b = 'mm/dd/yyyy' Note: Be patient while this process takes place. The prompt will return when the process is complete. The more complex your SQL statement is, the more data to be extracted, and the amount of background host processor load will have a limiting effect on the speed of processing the command.
The following can also be performed through the DB2 Tools Suite Command Line Interface or Command Center. The commands are not case sensitive and are presented here and include explanations. a. Sample connect to database
connect to pmdata user db2admin using db2admin
Where pmdata is the TotalStorage Productivity Center DB2 database, user db2admin, using the password of db2admin b. Select * from CNODE command example
select * from vpclul
The previous SQL query will select all (column) information from table vpclul, all rows are implied by the asterisk *
Note: Data is stored as a matrix with Columns (Field names) and Rows (Field Values). For more information about relational database tables, see the redbook IBM TotalStorage Expert Reporting: How to Produce Built-In and Customized Reports, SG24-7016.
983
Tip: You can use the FETCH clause in your SQL statement when testing your queries and scripts. This will limit your large script output to how ever many rows you have defined in the FETCH clause. Remember to remove the FETCH clause when your testing is completed to eliminate the erroneous output from your scripts. This can be placed as the last line of your SQL statement (note the semicolon which indicates to the CLP the end of the SQL statement); fetch first 10 rows only;
Planning considerations
Planning is one of the most important areas for consideration before beginning to do database backups. We cover the factors which should be weighed against one another in planning for recovery, for example, type of database, backup windows and relative speed of backup and recovery methods. We also introduce various backup methods. In general terms, DB2 can offer a number of options for backup and recovery management to meet the needs of a wide range of applications. The more simple backup and recovery options provide data protection with minimal administrator skill or effort. Other more powerful options give greater levels of data protection but require more administrator skill and require more effort to maintain.
984
If your organization has existing high levels of skills with DB2 or other relational databases, you may already have standard operating procedures for protecting databases. If your organization is less skilled in this area, you may wish to chose a simple backup and recovery process that doesnt require a lot of new administrator skill or effort.
Speed of recovery
If you ask users how quickly they would like you to be able to recover lost data, they usually answer, Immediately. In practice, however, recovery takes time. The actual time taken depends on a number of factors, some of which are outside your control (for example, hardware may need to be repaired or replaced). Nevertheless, there are certain things that you can control and that will help to ensure that recovery time is acceptable: Develop a strategy that strikes the right balance between the cost of backup and the speed of recovery. Document the procedures necessary to recover from the loss of different groups or types of data files. Estimate the time required to execute these procedures (and do not forget the time involved in identifying the problem and the solution). Set user expectations realistically, for example, by publishing service levels that you are confident you can achieve.
Database logging
In DB2 UDB databases, log files are used to keep records of all data changes. They are specific to DB2 UDB activity. Logs record actions of transactions. If there is a crash, Logs are used to playback/redo committed transactions during recovery. Logging is always on for regular tables in DB2 UDB: It is possible to mark some tables or columns as NOT LOGGED. It is possible to declare and use USER temporary tables. There are two kinds of logging: Circular logging (default) This is the TotalStorage Productivity Center database default logging type. Archive logging
985
In addition, capture logging is available for replication purposes. Each type of logging corresponds to the method of recovery you want to perform. Circular logging is used if the maximum recovery you want to perform is crash or restore recovery. Archive logging is used if you want to be able to perform rollforward recovery.
Note: IBM does not recommend or support the use of Archival logging with the TotalStorage Productivity Center product database.
Circular logging
Circular logging is the default behavior when a new database is created. (The logretain database configuration parameter setting is NO). With this type of logging, only full, offline backups of the database are valid. As the name suggests, circular logging uses a ring of online logs to provide recovery from transaction failures and system crashes. The logs are used in a round-robin fashion and retained only to the point of ensuring the integrity of current transactions. Circular logging does not allow you to roll a database forward through transactions performed after the last full backup operation. All changes occurring since the last backup operation are lost. Only the crash recovery and restore recovery can be performed when this type of logging. Active logs are used during crash recovery to prevent a failure (system power or application error) from leaving a database in an inconsistent state. The data changes are recorded in the log files and when all the units of work are committed or rolled back in a particular log file, the file can be reused. The number of log files used by circular logging is defined by the logprimary and logsecondary database configuration parameters. If there are UOWs running in a database using all the primary log files and still not reaching a point of consistency, then the secondary log files are allocated one at a time. Figure 17-29 shows a Circular Logging log path.
986
Archive logging
Archive logging is used specifically for rollforward recovery. You can configure this logging mode by setting the logretain database configuration parameter to RECOVERY. Rollforward recovery can use both archived logs and active logs to rebuild a database or a tablespace either to the end of the logs, or to a specific point-in-time. The rollforward utility achieves this by reapplying committed changes found in the following three types of log files: Active logs: Crash recovery also manipulates active logs, which uses them to place the database into a consistent state. They contain transaction records that have not been committed and also the committed transaction information that has not been written to the database on disk. You can locate active log files in the LOGPATH directory. Online archived logs: When changes in the active log are no longer needed for normal processing, the log is closed, and becomes an archived log. An archived log is said to be online when it is stored in the database log path directory (see Figure 17-30).
Offline archived logs: An archived log is said to be offline when it is no longer found in the database log path directory. When you want to use archive logging (see Figure 17-31), you must make provision for the logs to be stored away from the database. This is done in DB2 UDB by specifying a userexit parameter and interfacing to a suitable archive manager. Full documentation of this is supplied in the DB2 manuals (see Online resources on page 997).
987
Database recovery
A database restore will recreate the database from a backup. The database will exist as it did at the time the backup completed. If archival logging were used before the database crash, it would then be possible to roll forward through the log files to reapply any changes since the backup was taken. It is possible to roll forward either to the end of the logs or to a specific point in time. The granularity available on the last transaction needs to be weighed against database performance.
Important: Logfiles are just as important as the backup files. It is not possible restore the database without logfiles.
988
File: database_list
This file contains a line for each database to backup. Depending on the TotalStorage Productivity Center components you have installed this list may vary. You can use DB2 Control Center to establish the full list of databases in your installation. The backup data will reside in C:\db2_backups in this example. You need to create this directory before using this process as shown in Example 17-5.
Example 17-5 Example of database_list backup backup backup backup backup backup backup database database database database database database database DIRECTOR to C:\db2_backups without prompting; DMCOSERV to C:\db2_backups without prompting; ELEMCAT to C:\db2_backups without prompting; ESSHWL to C:\db2_backups without prompting; PMDATA to C:\db2_backups without prompting; REPMGR to C:\db2_backups without prompting; TOOLSDB to C:\db2_backups without prompting;
File: TCP_backup.bat
This is the script you run (see Example 17-6). It stops the IBM Director service which will close all connection to the DB2 databases. This will allow DB2 to take an offline backup. The script then restarts the IBM Director service.
Example 17-6 Example of backup script @ECHO ON @REM This is a sample backup script @REM to backup TotalStorage Productivity Center @REM for Disk and Replication @REM ----
@REM @REM
net stop IBM Director Support Program @REM Starting backup of DB2 databases @REM -------------------------------C:\PROGRA~1\IBM\SQLLIB\BIN\db2cmd.exe /c /w /i db2 -tvf C:\scripts\database_list @REM @REM Restarting TotalStorage Productivity Center -------------------------------------------
989
990
Appendix A.
Worksheets
This appendix contains worksheets that are intended for you to use during the planning and the installation of the TotalStorage Productivity Center. The worksheets are meant to be examples. Therefore, you can decide whether you need to use them, for example, if you already have all or most of the information collected somewhere. If the tables are too small for your handwriting, or you want to store the information in an electronic format, simply use a word processor or spreadsheet application, and use our examples as a guide, to create your own installation worksheets. This appendix contains the following worksheets: User IDs and passwords Storage device information: IBM TotalStorage Enterprise Storage Server (ESS) IBM Fibre Array Storage Technology (FAStT) IBM San Volume Controller
991
Server information
Table A-1 contains detailed information about the servers that comprise the TotalStorage Productivity Center environment.
Table A-1 Productivity Center server
Server
Machine Hostname IP address
Configuration information
____.____.____.____
In Table A-2, simply mark whether a manager or a component will be installed on this machine.
Table A-2 Managers/components installed
Manager/component
Productivity Center for Disk Productivity Center for Replication Productivity Center for Fabric Productivity Center for Data Tivoli Agent Manager DB2 WebSphere
Installed (y/n)?
992
Password
Enter the user IDs and password that you used during the installation in Table A-4. Depending on the selected managers and components, some of the lines are not used for this machine.
Table A-4 User IDs used on this machine
Element
Suite Installer DB2 IBM Director Resource Manager Common Agent Common Agent TotalStorage Productivity Center universal user Tivoli NetView IBM WebSphere Host Authentication
Enter user ID
Enter password
a. This account can have any name you choose. b. This account name cannot be changed during the installation. c. The DB2 administrator user ID and password are used here. See Fabric Manager User IDs on page 68.
For details about the purpose of the user IDs, see 3.5.1, User IDs on page 65.
Appendix A. Worksheets
993
Both IP addresses
LIC level
ESS password
994
IBM FAStT
Use Table A-6 to collect the information about your FAStT devices. Check the device support matrix before you install the CIM Agent for the correct level.
Table A-6 FAStT devices
Firmware level
IP address
Appendix A. Worksheets
995
Firmware level
IP address
User ID
Password
996
Related publications
The publications listed in this section are considered particularly suitable for a more detailed discussion of the topics covered in this redbook.
IBM Redbooks
For information about ordering these publications, see How to get IBM Redbooks on page 998. Note that some of the documents referenced here may be available in softcopy only.
TCP/IP Tutorial and Technical Overview, GG24-3376 Exploring Storage Management Efficiencies and Provisioning: Understanding IBM TotalStorage Productivity Center and IBM TotalStorage Productivity Center with Advanced Provisioning, SG24-6373 IBM TotalStorage SAN Volume Controller, SG24-6423 IBM Tivoli Storage Area Network Manager: A Practical Introduction, SG24-6848 IBM Tivoli Storage Resource Manager: A Practical Introduction, SG24-6886
Other publications
These publications are also relevant as further information sources:
IBM Tivoli Storage Area Network Manager Planning and Installation Guide, SC23-4697 IBM TotalStorage Enterprise Storage Server Command-Line Interface User's Guide, SC26-7494 IBM Tivoli Storage Resource Manager Configuration and Getting Started, SC32-9067
Online resources
These Web sites and URLs are also relevant as further information sources: Tivoli software products index:
http://www-306.ibm.com/software/tivoli/products/
Engenio:
http://www.engenio.com
FibreAlliance:
http://www.fibrealliance.org
997
998
Index
A
addess command 213 addessserver command 212, 214, 834 adduser command 216 administrator ID 67 agent deployment 361 Agent Manager 30, 51, 295, 902 %WAS_INSTALL_ROOT%AgentManager logsSyste mOut.log file 53 agentTrust.jks file 77 certificates 69 common agent 168 DB2 control center 902 default password 105 host name 102 IBMCDB 85, 98 key file 69 port number 181 ports used 102 status 902 TCP/IP ports 59 Agent Manager certificates 80 Agent Manager ports 59 agent placement 764 Agent registration password 69 agent registration password 52, 105 AgentManager registration password 105 AgentManager.properties file 168 Agenttrust.jks file 181 agentTrust.jks file 5253, 69, 77, 158 archive bit 627628 archive logging 985, 987 asset reporting 593, 595, 598 associated CIM Agent device support matrix 994 at risk files 548 Attended Installation 93 availability reporting 594, 604
C
capacity reporting 594, 605, 607 Carnegie Mellon University 817 changeMe password 105 Chargeback 592 chargeback 700 CIM Agent 30, 32, 35, 4748, 66, 84, 289290, 352 agent code 47 client application 47 device 47 device provider 47 ESS CLI 196 ESS configuration 196 overview 50 CIM Browser interface 220 CIM Client 35 CIM Managed Object 35 CIM Object Manager 194 CIM Object Manager (CIMOM) 32, 35, 47, 49 CIM Provider 35 CIM request 47 CIM Server 35 CIM/OM 635 discovery 635 CIM-compliant 47 CIMOM customization 192 delete entries 258 CIMOM (CIM Object Manager) 32, 47, 49 CIMOM communication 242, 248 CIMOM configuration recommendations 195 CIMOM discovery 49 CIMOM SVC console 241 CIMS 700 circular logging 985986 Cisco MDS 9000 734 Cleared Record 817 client application CIM requests 47 Cloudscape database 76 collect daemon 777 collectLogs.bat file 917 Command Center 952, 980 Command Line Processor (CLP) 949, 980 command-line interface 884 Common Agent AgentManager.properties file 67 context name 902 demo certificates 52 password 67 subagent 52 Common Agent Services Agent Manager 51 Resource Manager 51
B
backup reporting 594, 627, 659 backup storage requirements 661 backup storage requirements reporting 630 bandwidth 781 BASEENTITY table 915 batch reporting 683, 688 batch reports 697 Brocade 768 Brocade Silkworm Fibre Channel Switch 347 BWN005921E message 486 BWN005922E message 486 BWN005996W message 491
999
Common HBA API 35 Common Information Model (CIM) 28, 34 common schema 47 components 30 computer uptime 604, 657 configured user ID locations 901 Constraint 610, 614 Constraint violation report 610 Contacted status 348 Control Center 976 core schema 47 cpthresh 475 creating gauges 443 CSV 697 CSV format report 976 CSV output 592, 697 customized reporting 683 customized reports 976
D
DA benefit 42 Data Agent 30, 160, 354, 358 Data Agents 906 data collection SVC,SVC data collection 460 data collection task 429, 431 Data Manager 30, 7879, 84, 290, 293, 304, 351, 906 administrator rights 80 CIM interface 290 DB2 database 81 fully qualified host name 303 GUI 293, 906 local subnet 290 Parameters window 180 repository 177 security certificates 80 server 357 SLP interfaces 290 subtree 293 Data Manager security issues 79 Data planning considerations 78 Data user levels 80 database Alert 588 asset reporting 585, 598 availability 591 batch reporting 693 capacity reporting 607 chained rows 592 chargeback 700 Instance Alert 588 Instance Quota 591 policy management 589 Quota 589, 591 Quota violation reporting 622 Scan 601 space usage report 685 system reports 685 Table Alert 589 usage reporting 608
usage violation report 685 used table extents 591 user specific reporting 688 utilization 591 database backup 984 database instance storage report 667 database LUN assignment 669 database name 62, 150 new Element Catalog subcomponent database 151 database purge function 22 database recovery time 985 database restore 988 database size report 667 database storage usage 665 DataJoiner 981 DB2 Agent manager registry 902 Health Center monitoring 925 view a table 927 DB2 Command Center 952 DB2 Control Center 913, 976 DB2 Cube Views 980 DB2 database connect to 949 DB2 database health 922 DB2 DataJoiner 979 DB2 Development Center 950 DB2 Development Tools 950 DB2 Event Analyzer 951 DB2 General Administration Tools 950 Control Center 950 Journal 950 Replication Center 951 Task Center 951 DB2 Health Center 924 DB2 Intelligent Miner 981 DB2 journal 924 DB2 logging 985 DB2 Monitoring Tools 951 DB2 report customized example 956 DB2 Tool Suite command line processor 948 Command Line Tools 948 DB2 tool suite 948 DB2 UDB Journal utility 922 DB2 Utilities Command Center 952 db2move tool 982 DDL 980 default directory 359 delete a gauge 456 demo keys 52 Device Centric view 713 device discovery 921 device management 37 Device Manager device discovery 249 LUN mapping 385, 392 mdisk display 384, 390, 396, 403 overview 19
1000
directors 767 Directory Agent 31, 33, 38, 40, 900 configuring for subnet 45, 900 discover all services (DAS) 900 discovery 635, 733 device 249 disk capacity 595, 605 disk subject matter expert 12 Dispersion Frame Technique 817 display gauges 447, 454 Distance Vector Multicast Routing Protocol 44 distinguished name (DN) 105 DMCIMOM table 914 DMCOSERV database 913 dmlog.txt file 139 DNS server 77 DVMRP 44
Event Action Plan 23 export and import action plans 423 Message Browser 422 Event Action Planner creating an action 417 Event Action Plans 921 Event Filter Builder 920 Event Filters 24 event forwarding 774 event logging 781 Event Services 23 exception gauges 450 extension schema 47
F
Fabric Agent 30, 61, 361, 369 separate package 369 Fabric Agent deployment 160 Fabric Agent registration 904 Fabric fully qualified host name 77 Fabric Manager 84, 303, 321, 351, 902 install user ID rights 110 native installer setup.exe program 361 Fabric Manager Installation option 160 Fabric Manager ports 60 Fabric user ID password considerations 76 FA-MIB 771 FAStT CIM Agent 290291 defining devices 230 SLP registration 233 FAStT device 718, 995 fault isolation 817 Fault Record 817 FC_MGMT MIB 814 FCPortTxFrames 789 FE-MIB 771 FETCH clause 984 Fibre Alliance MIB 814 Fibre Channel 34 Common HBA API 35 Generic Service 34 switch 715 Fibre Channel MIB 769 files at risk 548 forbidden 611 modified since backup 629 most at risk 627 obsolete 548, 615 orphaned 548, 615 statistics 528 filesystem capacity 605 filesystems 736 firewall configuration 45 FlashCopy 829 flashsess command 889
E
element manager 723 ElementCatalog.properties file 154 enable logging 931 enabling WAS trace 932 enterprise-specific MIB 768 ESS 594 attached hosts report 656 CIM/OM 635 Logical Sub System 580 reporting 634635, 653, 656 Tivoli Storage Resource Manager Probe 635 used and free storage report 653 ESS CIM Agent 196, 291, 353 addess command 213 addessserver command 212, 214 configuring 212 install 202 log files 209 post install 211 setuser interactive tool 215 Truststore file 291 verify ESS connectivity 216 verify install 211 ESS CIM agent SLP registration 224 ESS CIMOM CIM Browser interface 220 restart 215 SLP registration 218 telnet command 220 ESS CIMOM verification 216 ESS CLI 196 install 196 verification 201 verifyconfig 218 ESS data collection 429 ESS thresholds 457 ESS user authentication 940 esscli command 201 Ethernet 812
G
gauge definition 14 Index
1001
gauge properties 454 gauges 22, 443 creating 443 delete 456 display 447, 454 exception 450 performance 443 properties 454 General Parallel File System (GPFS) 79 GPFS (General Parallel File System) 79 group 537 GUID 822
H
HBA 731 health monitoring 20 High-Performance Parallel Interface (HIPPI) 34 HIPPI (High-Performance Parallel Interface) 34 historical reporting 774 Host Centric view 714 host name 75, 343 hot link 698 HTML 592, 697 HTML output 697 HTTP 34
I
IBM Director 18 database name 121 Event Action Plans 921 Event Services 24 host name 118 user name 118 IBM Director (ID) 29, 55, 259, 993 IBM Director event logs 917 IBM Director Scheduler device discovery 253 IBM FAStT 713, 991 models for storage subsystem support 78 IBM Object REXX 981 IBM Tivoli Storage Area Network Manager 319 IBM Tivoli SAN Manager 20 IBM TotalStorage Enterprise Storage Server CIM Agent 30 managing Productivity Center for Disk 84 storage device 63 IBM TotalStorage Open Software Family 3, 5 IBM WebSphere Application Server 30 selection panel 130 IBMCDB 85 IBMCDB database 98 ICAT (Integrated Configuration Agent Technology) 47 ICMP 604 IDE 560 IDs and passwords (IP) 59 IETF 811812 IGMP (Internet Group Management Protocol) 44
IIS 85, 698 ikeyman 902 ikeyman utility 52, 135 inband agent 348 Inband discovery 32 inband discovery 12 incremental backup planning 661 Indication Record 817 inode 562 install image depot 94 installation Agent Manager 98 agentTrust.jks file 158 check list 64 Data Manager 171 Data planning 78 database considerations 75 database tuning 154 database types 62 DB2 90 DB2 username 134 Disk and Replication Base 127 DMCOSERV database 137 dmlog.txt 139 Fabric Agent deployment 160 Fabric manager 157 Fabric planning 75 Fabric port numbers 162 fully qualified host names 75 Generate Self-Signed Certificate 136 hardware 58 IBM Director 114 image depot 93 Internet Information Services 73 personal firewall 77 PMDATA database 143 prerequisite software 85 Productivity Center for Disk 140 Productivity Center for Replication 146 Replication Manager subcomponent database 152 SNMP 73 Suite Installer 110 SVC Replication Manager subcomponent database 153 TSRMsrv1 local account 177 user ID privileges 66, 131 user IDs 65 WebSphere Application Server 92 Windows Management Instrumentation 110 Windows Terminal Services 75 installation directory 187, 304 installation image depot 93 Integrated Configuration Agent Technology (ICAT) 47 Intelligent Peripheral Interface (IPI) 34 Internet Assigned Numbers Authority 44 Internet Engineering Task Force 32 Internet Group Management Protocol (IGMP) 44 Internet Information Server (IIS) 53, 295 Internet Information Services 73, 85 IP address 102, 291, 347, 366, 903, 992
1002
IP address assignment 44 IP multicasting 43 IP network 38, 248, 712, 805 iSCSI 810811 adapter 812 Auth MIB 814 discovery 733, 813 driver 812 initiators 811 iSNS MIB 814 MIB 813 NetView discovery 733, 813 SmartSet 734, 794 SNMP 733 targets 811 iSCSI Discovery 32 ISL 728, 738 iSNS 812 MIB 814
J
JDBC 979 job scheduling facility 433
K
key file for Agent Manager 69
L
Launch Device Manager 378 launcher program 910 Launchpad 28, 264, 313, 899 launchpad 906 customization 909 management application 910 leaf node 769 local subnet 45, 290, 900 multicast service requests 900 log files 917 Logical Sub System 580 Lotus 1-2-3 697 lscollection 475 lsfilter 475 lsgauge 475 lspair command 889 lsseq command 889 lssess command 886 LUN modeling 580 LUN to host port mapping 385, 392 LUNs 731
mdisks display 384, 390, 396, 403 MDS 9000 734 MIB 766767 applications 787 definitions 772 enable 772 enable in NetView 769 enterprise-specific 768, 770, 774 iSCSI 813 object ID 769, 776, 786 objects 769, 775, 784, 787, 804 performance objects 772 standard 768 subtree 773 thresholds 774 tree structure 769 MIB compiler 37 MIB-II 771, 781 Microsoft Excel 697 Internet Information Server 698 modified since backup files 629 MOSPF (Multicast Open Shortest Path First) 45 most at risk files 627 multicast 43 multicast group 44 individual hosts 44 multicast messages 39 Multicast Open Shortest Path First (MOSPF) 45 multicast request 40 multicast traffic 42 Multiple Device Manager database maintenance 977 database query 976 DB2 logging 986 device discovery 248 export PM data 979 report tools 979 SAN Manager 320 SQL commands 982 multiple machines, installing at the same time 361
N
namespace nterop 257 ootbm 257 NAS (network attached storage) 183 NAS devices 183 nestat command 207 netmon 805 netstat command 162 NetView 320, 704 acknowledge 744 Advanced Menu 769 arm threshold 781 child submap area 705 clearing the database 806 copying MIB 770 data collection 775, 791 Index
M
manage replication sessions 877 management application 30, 34 Management Information Base 37, 767 Management Information Base (MIB) 37 managers 30 mdisk group 398
1003
data collection troubleshooting 784 database 806 database maintenance 783 discovery process 805 enable MIB 769 event browser 343, 345 event logging 781 existing installation 343 explorer display 705, 731 graph 782, 803 Graph Properties 792 graphs 766, 774, 784 historical reporting 767, 774, 785 interface 704 iSCSI 813 iSCSI discovery 733, 813 iSCSI SmartSet 734, 794 loading MIB 771772 Management Page 741 maps 704 MIB applications 787, 790 MIB Browser 774, 789 MIB Data Collection 775, 802 MIB Data Collector 772, 774, 786 MIB Tool Builder 772, 786787, 793 Navigation Tree 707 netmon daemon 805 Object Properties 707, 740 performance applications 766 performance data 774, 781 polling 787, 792, 795 properties panel 730 real-time reporting 767, 786, 793 restricting discovery 805 root map 704, 725, 800 search function 822 seed file 805, 807 Server Setup 806807 SmartSets 725, 733, 794, 802, 813 SmartSets and Data Collection 802 status propagation 711, 800 submap stack 705 submap window 705 submaps 704, 725 supported MIBs 768 System Configuration view 706 Tivoli Storage Area Network Manager view 706 topology map 781, 785, 805 trap 781 trap forwarding 343, 346 trapfrwd daemon 343344 traps 342 unacknowledge 744 unmanage object 743 NetView commands nvsniffer 733, 794, 813 ovaddobj 344 ovstart 344 ovstart snmpcollect 784 ovstatus snmpcollect 784
network attached storage (NAS) 183 network bandwidth 781 network management 767 network monitoring 766, 787 network problem resolution 766 network resource allocation 766 NIC 812 Not responding status 348 Notification Record 817 NW_FAT (NF) 79
O
object ID 769 obsolete files 548, 615 offline archived logs 987 On Demand environment 25 Open Shortest Path First (OSPF) 45 Oracle archive log 588 regular administration 591 orphaned files 548, 615 OSPF (Open Shortest Path First) 45 outband agent 346 Outband discovery 32 outband discovery 11 ovaddobj command 344 ovstart command 344 own subnet 900
P
password Common Agent Services 901 Resource manager 901 Tivoli Agent Manager 901 perfcli tool 940 performance database purge task 944 performance gauge 443 Performance Manager 21 command line interface 475 customized reports 976 Enable threshold button 458 ESS data collection 429 ESS data collection task 429 exporting data 976 function 428 gauges 22, 443 remote console 270 threshold filters 459 thresholds task 457 Volume Performance Advisor 23 performance metrics 774 perftool 475 personal firewall 77 PFA 815 PIM 44 Ping 592, 594, 604 ping command 216 pmcli 475 PMDATA table 977
1004
poll interval for the SAN configuration 349 polling interval 349, 722 port 343 port number 295, 364, 911 port statistics 775 PPRC 829 Predictive Failure Analysis 815 prerequisite software 85 Probe 560, 585, 592594, 604, 634635 problem determination dmlog.txt 139 product 30 Productivity Center 16, 83, 362, 910 components 16 Productivity Center for Data 7 architecture 9 features 8 managing user IDs 68 Productivity Center for Disk functions 13 gauge definition 14 remote console 264 Volume Performance Advisor 14 Productivity Center for Fabric 9 agent 721 application launch 739 benefits 12 change device icon 738 change device label 738 Connection 738 Device Centric view 713 element managers 723 Host Centric View 732 Host Centric view 713714 Interconnect Elements 729 overview 9 parent icons 720 physical view 726 polling interval 722 SAN topology 718 SAN view 713714 Sensors/Events 737, 739 Submap Explorer button 725 switch display 737 switch port 722 topology map 781 topology view 725, 728 troubleshooting 902 unknown device 738 Productivity Center for Fabric - Agent checking inband agents 348 Productivity Center for Fabric Server outband agents 346 Productivity Center for Fabric. See Tivoli SAN Manager Productivity Center for Replication 23 overview 14, 828 Profile 617, 628, 672, 700 progressive incremental backup 633 Protocol-Independent Multicast 44 provider component 49 provisioning 320
Q
QMF 979 QMF for Windows 980 Query Management Facility (QMF) 980 Quota 594, 610 violation report 617
R
raswatch 921 RDBMS 177 RDBMS support 79 Redbooks Web site 998 Contact us xvii relational database management system 944 remote console 259 Performance Manager 270 Remote Fabric Agent Deployment option 160 remote GUI 295, 334 repcli command 894 repcli command syntax 885 repcli utility 885 Replication Manager 23 CLI 884 Continuous Synchronous Remote Copy 861, 873 copyset 833 Copyset details 870 create a group 838 create a storage group 838 Create Path wizard 850 define storage group 843 delete a storage pool 848 freeze operations 930 groups 831 Managing a storage pool 847 modify a storage group 841 overview 280, 828 Point-in-Time Copy session 852 remote CLI 280 replica sessions 23, 828 restarting 931 sequence 833 Session Properties window 876 Sessions window 873 setting up 834 start replication session 888 storage paths 850 storage pool 831 storage pool create 844 suspend a session 882 suspended status 875 synchronized state 881 tasks 830 troubleshooting 930 verifying source-target relationship 856 view group properties 842 view storage pool properties 849
Index
1005
Replication Manager (RM) 59 Replication Manager problem determination 930 Replication Manager subcomponent 150151 replication session 830 Replication subject matter expert 14 reporting assets 593, 595 availability 594, 604 backup 594 backup storage requirements 630 backups 627 batch 683, 688, 697 by userID 532, 683 capacity 594, 605 computer uptime 657 Constraint violation 614 customized 683 database assets 598 database batch 693 database capacity 607 database Quota violations 622 database space usage 685 database usage 608 disk capacity 605 filesystem capacity 605 owned by a username 686 Quota violation 617 saved reports 687 scheduling 683, 697 storage capacity 605, 684 storage subsystems 594 top 10 reports 653 uptime 657 usage 594, 607 usage violation 594, 610 wasted space 594 Web publishing 698 reporting categories 593 Tivoli Storage Resource Manager 593 reports HTML output 697 resetarchiveattribute 632 resource accounting 700 Resource Manager 30, 51, 168, 351, 993 configured user ID 901 registration 902 troubleshootng 902 user ID 76 user ID and password 67 Resource Manager ports 59 Reverse Path Forwarding 44 REXX 979 RFC 769 rmgauge 475 RNID 731 root cause analysis 814
S
SAN Cleared Record 817
discovery 805 fault isolation 817 Fault Record 817 historical reporting 767, 785 interconnects 728 management 794 monitoring 766, 787 navigation 725 Notification Record 817 performance data 774 real-time reporting 767, 786, 793 root cause analysis 814 switch port statistics 775 topology 704, 726, 728, 800 SAN (storage area network) 30 SAN component 345 SAN Manager 20, 320 SAN view 714 SAN Volume Controller thresholds 461, 466 SAN Volume Controller (SVC) 78, 84, 996 Scan 537, 563, 592, 601602, 614, 617, 675 Scan job log 602 scheduled reports 592 scheduling 697 SCSI 560 protocol 811 seed file 805, 807 Service Agent 31, 38, 900 service information 39 Service Location Protocol multicast 42 Service Location Protocol (SLP) 38, 290 service agent 38 user agent 38 service URL 40, 42 setdevice command 928 setessthresh/setsvcthresh 475 setfilter 475 setoutput 476 setoutput command 894 setuser command 216, 257 setuser interactive tool 215 showgauge 475 showsess command 886 Simple Network Management Protocol 36 Simple Network Management Protocol (SNMP) 64, 85 SLP active DA discovery 41 address 44 broadcast communication 43 CIM Agent 47 configuration recommendation 193, 900 DA configuration 900 DA considerations 192 DA discovery 41 DA functions 42 directory agent configuration 194, 900 Discovery 901 discovery requirements 242, 248
1006
environment configuration 900 ESS CIM agent 224 firewall configuration 45 multicast address and port 193 multicast communication 43 multicast group 44 multicast messages 39 multicast request 40 multicast service request 41 passive DA discovery 42 port number 44, 901 registration 39 registration persistency 243 router configuration 193, 900 service agent 38 service attributes 39 service type 39 slp.conf 221 starting 211 unicast 42, 193 unicast communication 43 user agent 39 User Datagram Protocol message 39 verify install 211 verifyconfig command 224 when to use DA 43, 193 SLP (Service Location Protocol) 38 SLP considerations 928 SLP DA 38, 290, 900 SLP discovery summary 242, 901 SLP environment 38 slp logfile 929 SLP tracing 929 slp.conf file 221, 929 slp.reg file 243 slptool command 901 SmartSet 794 SmartSets 725, 733734, 794, 813 SMI-S 32 SMI-S (Storage Management Initiative - Specification) 35 SNIA (Storage Networking Industry Association) 35 SNMP 36, 73, 110, 733, 767, 795 agents 767 collect daemon 777, 782, 784 community name 343 console 346 manager 767 port 317, 345 trap 692, 815 trap forwarding 342343 SNMP management application 37, 346 station product 37 SNMP manager 37, 320 SNMP trap 37, 342343 spreadsheets 697 SQL Assist 962 SQL command example 982 SQL scripts considerations 956
SQL-Server 600 SRMURL 742 SSL configuration 135 standard MIB 768 standard reporting 594 standards organizations 4 startcimbrowser command 220 startesscollection 475 startsvccollection 475 Stochastic 817 stopcollection 475 stopflashsess command 893 stopsess command 893 storage capacity 605, 684 utilization 548 storage area network (SAN) 30 storage device 27, 63, 248, 320, 900, 994 important information 994 inband management 34 Storage Management Initiative - Specification 32 Storage Management Initiative - Specification (SMI-S) 35 Storage Networking Industry Association (SNIA) 35 storage orchestration 5 storage subsystem 290 Monitored check box 294 Policy Management jobs 294 Storage Subsystems reporting 594, 634 Structured Query Language (SQL) 944, 980 subagent 351352 subcomponent 30 subnet 900 query group membership 44 Subsystem Device Driver (SDD) 58 Suite Installer 110 suite installer 83, 304, 319, 361, 911, 993 launchpad 907 suspendsess command 892 SVC mdisk group 398 SVC CIMOM 234 console 234 Multiple Device Manager console account 235 register to SLP DA 241 SVC console account 234 verification 241 SVC data collection error 940 SWDATA 949 swFCPortTxFrames 775, 788, 802803 swFCRxErrors 783 swFCRxFrames 783 swFCTxErrors 783 swFCTxFrames 785 switch commands snmpmibcapset 772 switch port 722 switches administrative rights 347 API 347 display 728
Index
1007
environmentals 739 login ID 347, 772 management applications 740, 767 performance data 774, 781 port connections 738 port statistics 775 sensors 739 trap forwarding 342 zone information 347, 730 SW-MIB 771, 788 system reports 685
T
Table 600601 TCP/IP 811 TCP/IP ports 59 TEC 317, 711, 815 telnet 772 telnet command 220, 241 telnet connection 911 Test Connection button 258 threshold checking 21 threshold properties 462, 467 thresholds task 457 Tivoli 342 Tivoli Agent Manager 66 password 901 Tivoli Agent Manager service 171 Tivoli Common Agent Service 29, 901 Tivoli Common Agent Services 51 Tivoli Enterprise Console 545 Tivoli Event 76, 320 Tivoli Event Console (TEC) 320 Tivoli Monitoring for Databases 591 Tivoli NetView 12, 30, 60, 331, 993 7.1.3 icon 904 General Topology map service 60 Object Collection facility socket 60 Object Database 60 Object Database event socket 60 OVs_PMD management service 60 OVs_PMD request service 60 Pager 60 PMD service 60 SAN menu 75 SnmpServer 60 Topology Manager 60 Topology Manager socket 60 trapd socket 60 Web Server socket 60 Tivoli NetView installation 346 Tivoli NetView service 70 Tivoli SAN Manager 817 agents 343, 731 application discovery 740 change device icon type 708 change device label 708, 727, 744 Clear History 744 Configure Agents 711 configure management application 740
Configure Manager 349, 711, 744 database 815 Device Centric View 726, 731 device icons 709, 736 device label 736 device properties 736 display switch connections 738 ED/FI Configuration 711 ED/FI Properties 711 event forwarding 774 fabric ports 738 filesystem display 736 historical reporting 767, 774, 785 Host Centric View 726 host display 736 icons 709, 743 indication record 817 initial poll 349 iSCSI 812 iSCSI discovery 733, 813 iSCSi discovery 733 Launch Application 711, 740 launch Tivoli Storage Resource Manager 711 logical views 731732 LUN display 731 managed hosts 732 Navigation Tree 707 NetView traps 342 object status 709, 743 outband agents 346 physical topology 725 polling 343, 349 Predictive Failure Analysis 815 propagation 711 real-time reporting 767, 786, 793 RNID 731 SAN menu 711, 735 SAN Properties 711, 727, 735, 743 Sensors/Events 735 Set Event Destination 711 status colors 709 status cycle 743 status propagation 711, 800 submap 711 switch environmentals 739 Tivoli Storage Resource Manager 711 topology map 343, 726, 743744, 800 trap forwarding 342343, 346 zone display 347, 727, 730 Tivoli Storage Manage resetarchiveattribute 632 Tivoli Storage Manager 575, 588, 627 archive bit 627 Constraint violation report 610 progressive incremental backup 633 Tivoli Storage Manager capabilities Backup-Restore 594 Tivoli Storage Resource Manager 607, 615, 697 Alert 613 Alert Disposition 316
1008
Alert log 317, 564, 616, 619 asset reporting 593, 595 at risk files 548 availability reporting 594, 604 backup reporting 594, 627, 659 backup storage requirements 630, 661 batch reports 592, 683, 688, 697 capacity reporting 594, 605 chargeback 700 Computer Alert 560 Computer Group 561 Computer Uptime 657 computer uptime 657 Constraint 610, 614 Constraint Violation report 614 CSV output 592, 697 customized reporting 683 database asset reporting 598 define Alert 559 Directory Alert 563 Directory Group 538, 598 directory monitoring 598 disk capacity 605 email notification 317 ESS reporting 634635, 653, 656 file statistics 528 Filesystem Alert 562 filesystem capacity 605 Filesystem Group 537 forbidden file 611 graphical reporting 616 Group definition 539 HTML output 592, 693, 697 interactive reporting 592 launch 711 LUN modeling 580 mail port 317 modified since backup files 629 monitored directories 598 most at risk files 627 My Reports 592, 683 NetWare reporting 595 obsolete files 548, 615 orphaned files 548 OS User Group Group 540 Ping 592, 594, 604 pre-defined reports 592 Probe 560, 592594, 604, 634635 Profile 617, 628, 672, 700 Quota 594, 610 Quota violation report 617 report scheduling 683, 697 reports on the Web 698 saved reports 687 Scan 563, 592, 614, 617, 675 Scan job log 602 scheduled reports 592 script parameters 561 standard reporting 594 Storage Subsystem Reporting 594, 653, 656
System Reports 683 system-wide view 598 TEC configuration 317 tool bar 602 top 10 reports 653 Triggered Action 560 Triggering condition 560, 562 uptime reporting 604, 657 usage reporting 594 usage violation reporting 594, 610 User Group 540 username reporting 532, 683 wasted space report 594 Tivoli Storage Resource Manager for Chargeback 592, 700 Tivoli Storage Resource Manager for Databases Alert 588 Alert log 589 asset reporting 585, 598 availability check 591 batch reports 693 capacity reporting 607 Computer Groups 584, 607 create Table Group 601 database instance report 667 database LUN reporting 669 Database Quota 591 database Scan 601 Database-Tablespace Alert 588 Instance Alert 588 Instance Quota 591, 622 My Reports 683 Network Quota 590 policy management 589 Probe 585 Profile 586 Quota 589 Quota violations 622 Scan 587, 601 script 589, 591 storage usage 665 system reports 685 Table Alert 589 Table Group 589, 600 usage reporting 608 User Group 585 user specific reports 688 top 10 reports 653 TotalStorage Productivity Center 6 communication user 67 components installed 910 Device Manager 20 distinct managers 31 environment 30 Event Action Plan 23 foundation 1 Launchpad 264, 313 launchpad 906 performance considerations 194 Performance Manager 21
Index
1009
remote console 259 SAN Manager 20 SAN view 727 server 994 universal user 66 vdisk display 384, 390, 396, 403 Version 2.1 320 TotalStorage Productivity Center (TPC) 83, 320, 991 trap 37 trap forwarding 342343, 346 trapfrwd 344 trapfrwd daemon 343 trapfrwd.conf 344 traps 781 trend analysis 766 troubleshooting enable trace logging 931 TRP-MIB 771 TSANM 405, 409, 413 TSANMDB database 75 TSRMsrv1 local account 177
W
WAS trace 930 WAS trace control 938 wasted space report 594 WBEM (Web-Based Enterprise Management) 34 WBEM browser 220 WBEM initiative 34 Web browser 698 Web-Based Enterprise Management (WBEM) 34 Webshpere Application Server trace 932 WebSphere changing user ID 76 startServer.log 927 WebSphere Application Server 29, 67, 295 ikeyman utility 135 Information panel 102 install 92 SSL communication 69 update 62 WebSphere logfile 927 WebSphere user ID password 76 Windows archive bit 627, 632 Windows Management Instrumentation 85 Windows Services 69 Windows Terminal Services 75 WWN 731, 822
U
unattended installation 93 unicode 980 Uniform Resource Locator (URL) 39 UNIX inode 562 uptime 604, 657 usage reporting 594, 607 usage violation reporting 594, 610 User Agent (UA) 31, 3840, 248, 290, 900 SLP User Agent interactions 41 User Datagram Protocol (UDP) 39 message 39 user ID 65, 85, 366, 991, 993 user name 348 user rights 125
X
XML 49 xmlCIM 34
Z
zones 730 zoning 347
V
vdisks display 384, 390, 396, 403 verifyconfig command 218, 224 Volume Performance Advisor 14, 23 authentication 488 getting started 482 multiple recommendations 481 predefined workload profiles 483 recommendation process 480 workload characteristics 479 workload profile 482 VPA overview 478 VPCCH table 926, 973 VPCLUS table 926 VPCRK table 926, 974, 983 VPVPD table 983 VTHRESHOLD table 977
1010
Back cover