SG 248006
SG 248006
SG 248006
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
Learn about the business benefits of SSI and LGR Understand important concepts and planning Step through implementation and practical approaches
Lydia Parziale Anthony Bongiorno Howard Charter Jo Johnston Volker Masen Clovis Pereira Sreehari Somasundaran Srivatsan Venkatesan
ibm.com/redbooks
International Technical Support Organization An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR) April 2012
SG24-8006-00
Note: Before using this information and the product it supports, read the information in Notices on page vii.
First Edition (April 2012) This edition applies to Version 6, Release 2, of z/VM.
Copyright International Business Machines Corporation 2012. All rights reserved. Note to U.S. Government Users Restricted Rights -- Use, duplication or disclosure restricted by GSA ADP Schedule Contract with IBM Corp.
Contents
Notices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vii Trademarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . viii Preface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . The team who wrote this book . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Now you can become a published author, too! . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Comments welcome. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Stay connected to IBM Redbooks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Chapter 1. Introduction. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.1 Business benefits . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.1.1 Simplified z/VM systems management . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.1.2 Share system resources with high resource utilization . . . . . . . . . . . . . . . . . . . . . . 1.1.3 Non-disruptive maintenance and upgrades . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.1.4 Concurrent support for disparate environments . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.2 Navigating the z/VM library . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.2.1 z/VM V6R2.0 General Information . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.2.2 z/VM V6R2.0 Installation Guide . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.2.3 z/VM V6R2.0 CP Planning and Administration . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.2.4 z/VM V6R2.0 Getting Started with Linux on System z . . . . . . . . . . . . . . . . . . . . . . 1.2.5 z/VM V6R2.0 VMSES/E Introduction and Reference . . . . . . . . . . . . . . . . . . . . . . . 1.2.6 z/VM V6R2.0 Directory Maintenance Facility Commands Reference. . . . . . . . . . . 1.2.7 z/VM V6R2.0 Connectivity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.2.8 z/VM V6R2.0 CP Commands and Utilities Reference . . . . . . . . . . . . . . . . . . . . . . 1.2.9 z/VM V6R2.0 Migration Guide. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ix ix xi xi xi 1 2 2 2 2 3 3 3 4 4 4 5 5 6 6 7
Chapter 2. Single system image (SSI) overview. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9 2.1 Introduction to the z/VM single system image (SSI) clusters . . . . . . . . . . . . . . . . . . . . 10 2.2 Major attributes of the z/VM SSI cluster . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10 2.2.1 z/VM installation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10 2.2.2 Changes to applying service . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10 2.2.3 Workload balancing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11 2.2.4 Single configuration users and multiconfiguration users. . . . . . . . . . . . . . . . . . . . 11 2.2.5 DASD requirements . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12 2.2.6 Network requirements . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14 2.2.7 Equivalency identifiers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15 2.2.8 MAC addresses . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16 2.2.9 New user IDs for z/VM 6.2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17 2.2.10 Persistent data record . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18 2.2.11 VMSSI considerations. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19 2.3 Planning considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19 2.3.1 Logical partition (LPAR) planning considerations . . . . . . . . . . . . . . . . . . . . . . . . . 19 2.3.2 DASD planning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19 2.3.3 ISFC . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20 2.3.4 Non-SSI z/VM installation planning considerations. . . . . . . . . . . . . . . . . . . . . . . . 20 2.4 z/VM SSI cluster operation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20 2.4.1 SSI member states . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21 2.4.2 SSI cluster modes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
iii
Chapter 3. Live guest relocation (LGR) overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.1 Introducing live guest relocation (LGR) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.2 Major attributes of live guest relocation. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.2.1 Factors affecting relocation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.3 Supported configurations for relocation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.4 Memory and paging considerations for live guest relocation . . . . . . . . . . . . . . . . . . . . 3.5 Relocation domains. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.5.1 Characteristics of relocation domains . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.5.2 Steps for defining relocation domains . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.5.3 Steps for placing virtual servers into relocation domains . . . . . . . . . . . . . . . . . . . 3.6 Performing a relocation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.6.1 VMRELOCATE command. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.6.2 Performing the relocation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.7 Conditions that prevent a relocation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
23 24 24 24 25 26 28 28 30 31 31 31 32 33
Chapter 4. Scenario one: Creating a single system image cluster with four new z/VM members . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35 4.1 Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36 4.2 Setting up the SSI cluster . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36 4.2.1 Planning worksheet. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36 4.2.2 Hardware definition . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37 4.2.3 Defining the software . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37 4.2.4 Configuring the DASD. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37 4.2.5 Configuring the network . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40 4.2.6 Guests. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45 4.3 Installation process . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45 4.3.1 Instplan command. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46 4.3.2 INSTALL command. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47 4.3.3 INSTSCID command. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47 4.3.4 Configuring the system . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47 4.3.5 Setting up the software . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54 4.4 Relocation examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 56 4.4.1 Live guest relocation for Linux on System z . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57 4.5 Problems encountered . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 60 4.6 Lessons learned . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 60 Chapter 5. Scenario two: Converting two existing stand-alone z/VM systems to a single system image cluster . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 61 5.1 Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 62 5.1.1 Preparing and executing scenario two . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 63 5.2 Setting up ITSOSSI5 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 64 5.2.1 Planning worksheet. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 64 5.2.2 Defining the hardware . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 64 5.2.3 Defining the software . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65 5.2.4 Installing the z/VM stand-alone client . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65 5.2.5 Installing additional software products . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66 5.3 Setting up ITSOSSI6 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66 5.3.1 Defining the hardware . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67 5.3.2 Defining the software . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67 5.3.3 Installation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67 5.4 Converting ITSOSSI5 to a member of an SSI cluster . . . . . . . . . . . . . . . . . . . . . . . . . . 69 5.4.1 Setting RACF to use as a shared database . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 69 5.4.2 Completing the conversion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 70 5.5 Creating an ITSOSSI6 system by cloning ITSOSSI5 . . . . . . . . . . . . . . . . . . . . . . . . . . 70 iv
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
5.5.1 Cloning ITSOSSI5. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.6 Suggestions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.7 Relocation examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.7.1 Live guest relocation of a z/VM guest . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.7.2 Relocating an SAP system in a typical System z setup . . . . . . . . . . . . . . . . . . . .
70 77 78 78 79
Appendix A. Frequently asked questions. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 87 A.1 Guests do not relocate . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 88 A.1.1 Linux with a DEVNO minidisk in use elsewhere . . . . . . . . . . . . . . . . . . . . . . . . . . 88 A.1.2 Linux with CMS mdisks defined . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 89 A.1.3 Using a LINUX guest with V-Disk minidisk definition . . . . . . . . . . . . . . . . . . . . . . 90 A.1.4 Moving a CMS guest . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 91 A.1.5 Using DEDICATE devices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92 A.1.6 Live guest relocation of zLinux guests outside domains. . . . . . . . . . . . . . . . . . . . 97 A.1.7 Relocation exceeds MAXQUIESCE time . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 98 A.1.8 Relocation exceeds available auxiliary paging space. . . . . . . . . . . . . . . . . . . . . . 99 A.1.9 Relocation exceeds available capacity of storage . . . . . . . . . . . . . . . . . . . . . . . 100 A.2 LPAR names are the same on separate CECs in the cluster . . . . . . . . . . . . . . . . . . . 101 A.3 Incorrect system name . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 102 A.4 Recovery messages on the Linux console during relocation . . . . . . . . . . . . . . . . . . . 102 Appendix B. New and updated commands. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . B.1 Single system image and live guest relocation commands . . . . . . . . . . . . . . . . . . . . B.2 DIRMAINT commands . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . B.2.1 Adding identities . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . B.2.2 Getting identities . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . B.2.3 Using prototypes and batch functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . B.3 VMSES commands. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Appendix C. Additional material . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Locating the web material . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Using the web material. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . System requirements for downloading the web material . . . . . . . . . . . . . . . . . . . . . . . Downloading and extracting the web material . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Related publications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Other publications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Online resources . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Help from IBM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 105 106 107 107 109 109 111 113 113 113 113 114 115 115 115 115
Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 117
Contents
vi
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
Notices
This information was developed for products and services offered in the U.S.A. IBM may not offer the products, services, or features discussed in this document in other countries. Consult your local IBM representative for information on the products and services currently available in your area. Any reference to an IBM product, program, or service is not intended to state or imply that only that IBM product, program, or service may be used. Any functionally equivalent product, program, or service that does not infringe any IBM intellectual property right may be used instead. However, it is the user's responsibility to evaluate and verify the operation of any non-IBM product, program, or service. IBM may have patents or pending patent applications covering subject matter described in this document. The furnishing of this document does not give you any license to these patents. You can send license inquiries, in writing, to: IBM Director of Licensing, IBM Corporation, North Castle Drive, Armonk, NY 10504-1785 U.S.A. The following paragraph does not apply to the United Kingdom or any other country where such provisions are inconsistent with local law: INTERNATIONAL BUSINESS MACHINES CORPORATION PROVIDES THIS PUBLICATION "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESS OR IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF NON-INFRINGEMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Some states do not allow disclaimer of express or implied warranties in certain transactions, therefore, this statement may not apply to you. This information could include technical inaccuracies or typographical errors. Changes are periodically made to the information herein; these changes will be incorporated in new editions of the publication. IBM may make improvements and/or changes in the product(s) and/or the program(s) described in this publication at any time without notice. Any references in this information to non-IBM websites are provided for convenience only and do not in any manner serve as an endorsement of those websites. The materials at those websites are not part of the materials for this IBM product and use of those websites is at your own risk. IBM may use or distribute any of the information you supply in any way it believes appropriate without incurring any obligation to you. Information concerning non-IBM products was obtained from the suppliers of those products, their published announcements or other publicly available sources. IBM has not tested those products and cannot confirm the accuracy of performance, compatibility or any other claims related to non-IBM products. Questions on the capabilities of non-IBM products should be addressed to the suppliers of those products. This information contains examples of data and reports used in daily business operations. To illustrate them as completely as possible, the examples include the names of individuals, companies, brands, and products. All of these names are fictitious and any similarity to the names and addresses used by an actual business enterprise is entirely coincidental. COPYRIGHT LICENSE: This information contains sample application programs in source language, which illustrate programming techniques on various operating platforms. You may copy, modify, and distribute these sample programs in any form without payment to IBM, for the purposes of developing, using, marketing or distributing application programs conforming to the application programming interface for the operating platform for which the sample programs are written. These examples have not been thoroughly tested under all conditions. IBM, therefore, cannot guarantee or imply reliability, serviceability, or function of these programs.
vii
Trademarks
IBM, the IBM logo, and ibm.com are trademarks or registered trademarks of International Business Machines Corporation in the United States, other countries, or both. These and other IBM trademarked terms are marked on their first occurrence in this information with the appropriate symbol ( or ), indicating US registered or common law trademarks owned by IBM at the time this information was published. Such trademarks may also be registered or common law trademarks in other countries. A current list of IBM trademarks is available on the Web at http://www.ibm.com/legal/copytrade.shtml The following terms are trademarks of the International Business Machines Corporation in the United States, other countries, or both:
AIX CICS DB2 ECKD FICON GDPS IBM IMS MVS RACF Rational Redbooks Redpapers Redbooks (logo) System z10 System z Tivoli VTAM WebSphere z/OS z/VM z/VSE z10
The following terms are trademarks of other companies: Intel, Intel logo, Intel Inside, Intel Inside logo, Intel Centrino, Intel Centrino logo, Celeron, Intel Xeon, Intel SpeedStep, Itanium, and Pentium are trademarks or registered trademarks of Intel Corporation or its subsidiaries in the United States and other countries. Linux is a trademark of Linus Torvalds in the United States, other countries, or both. Other company, product, or service names may be trademarks or service marks of others.
viii
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
Preface
IBM z/VM 6.2 introduces significant changes to z/VM in the form of multi-system clustering technology allowing up to four z/VM instances in a single system image (SSI) cluster. This technology is important, because it offers clients an attractive alternative to vertical growth by adding new z/VM systems. In the past, this capability required duplicate efforts to install, maintain, and manage each system. With SSI, these duplicate efforts are reduced or eliminated. Support for live guest relocation (LGR) allows you to move Linux virtual servers without disruption to the business, helping you to avoid planned outages. The z/VM systems are aware of each other and can take advantage of their combined resources. LGR enables clients to avoid loss of service due to planned outages by relocating guests from a system requiring maintenance to a system that remains active during the maintenance period. Together, the SSI and LGR technologies offer substantial client value, and they are a major departure from past z/VM practices. This IBM Redbooks publication gives you a broad understanding of the new SSI architecture and an overview of LGR. In 5.7.2, Relocating an SAP system in a typical System z setup on page 79, we show an LGR example that shows a typical SAP user environment. In our example, the SAP Application Server Central Instance resides on a Linux on System z guest and an IBM DB2 10 database server runs on z/OS. This book is written for IT architects, who design the systems, and IT specialists, who build the systems.
ix
Jo Johnston is a Certified IT Specialist and Chartered Engineer, who works in the IBM System z Strategic Outsourcing Team in the United Kingdom. She has worked on IBM mainframe systems as a systems programmer supporting z/VM, z/OS, IBM MVS, IBM CICS, DB2, and IBM IMS for more than 30 years. She joined IBM in 2001 working in Strategic Outsourcing Commercial Platform Support where she provided day-to-day z/OS, CICS, DB2, and IMS support for client systems that had been outsourced to IBM. She then moved to the IBM System z Technical Sales Team specializing in IBM WebSphere Business Integration products on System z with specific responsibility for WebSphere Application Server, WebSphere Process Server, and WebSphere Service Registry and Repository. She now works in the System z Database team supporting not only DB2 and Adabas on z/OS, but also WebSphere Application Server on z/OS and Tivoli Storage Productivity Center for Replication on z/OS and z/VM. Jo has co-authored two IBM Redpapers on WebSphere Process Server on System z. Volker Masen is a IT specialist in the IBM Development Lab in Boeblingen, Germany. He started his career 20 years ago in the System z environment, supporting the library and build management environment around ISPF/SCLM. After spending several years supporting development environments for the IBM Rational brand, he moved back into the System z environment three years ago as a system programmer for z/VM in Boeblingen and specialist for other virtualization environments (VMware and KVM). Clovis Pereira is a System Support Analyst in Brazil. He has about 28 years of experience in supporting IBM operating systems, the last 22 at IBM. His areas of expertise include IBM z/VM, z/OS, and z/VSE. Today, he works in the Customer Support Center (CAC) in Brazil, helping clients to solve problems related to operating systems (z/VM, z/OS, and z/VSE) and related products. He has written extensively on Chapter 4, Scenario one: Creating a single system image cluster with four new z/VM members on page 35 and Appendix A, Frequently asked questions on page 87. Sreehari Somasundaran is an IBM and Open Group Certified IT Specialist working with the Systems and Technology Group. He is a zChampion and the System z Techline Specialist for IBM India/South Asia. He holds an MBA in Marketing and Operations from the Indian Institute of Management (IIMK) and an Engineering Degree in Electronics and Communication from Kerala University. He joined IBM in 1997 and has worked in multiple roles handling System z application development, project management, infrastructure consulting, TCO/ROI Analysis, performance analysis and capacity planning studies, and has managed multiple benchmarks and proofs of concept. His areas of expertise include System z hardware, z/OS, and z/VM. He is a regular speaker at the System z Technical conferences in India. Srivatsan Venkatesan is an IT Specialist in the Systems and Technology Group in the US. He has one year of experience in the z/VM and Linux on System z fields. He holds a degree in Computer Science from the University of South Carolina. His areas of expertise include Linux and middleware on System z. Thanks to the following people for their contributions to this project: Bill Bitner of the z/VM Customer Focus and Care team The z/VM Architecture and Development team: Emily K Hugenbruch, John Franciscovich, Mark Lorenc, Michael Wilkins, Jim Switzer, Ken Davis, Susan Farrell, and Carol Everitt Roy Costa and Robert Haimowitz International Technical Support Organization, Poughkeepsie Center Oliver Petrik IBM Boeblingen x
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
Comments welcome
Your comments are important to us! We want our books to be as helpful as possible. Send us your comments about this book or other IBM Redbooks publications in one of the following ways: Use the online Contact us review Redbooks form found at: ibm.com/redbooks Send your comments in an email to: [email protected] Mail your comments to: IBM Corporation, International Technical Support Organization Dept. HYTD Mail Station P099 2455 South Road Poughkeepsie, NY 12601-5400
Preface
xi
xii
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
Chapter 1.
Introduction
We are in an era of information technology in which enterprises address the challenge of an ever-increasing demand for IT resources and tighter budgets. IBM smarter computing addresses this challenge through IT transformation.
Virtualization is a key element of this transformation. z/VM Version 6.2 provides virtualization enhancements that help reduce sprawling IT. It helps provide flexibility to meet user demands for security, availability, scalability, and the manageability of IBM System z.
This chapter provides a brief explanation of the business benefits obtained when using IBM z/VM Version 6.2 with single system image (SSI) and live guest relocation (LGR). Additionally, we provide navigation of the z/VM library with a brief abstract of the purpose of each volume, and where these books have additional information about z/VM V6.2.
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
the z/VM software or hardware maintenance. This reduced impact delivers the application continuity that clients require and is an important factor contributing to an optimized system.
Chapter 1. Introduction
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
Information in this manual describes how to configure and use z/VM functions and facilities for Linux servers running on the IBM System z platform. The document provides requirements and guidelines to implement during z/VM installation, but primarily assumes that you have installed z/VM and are ready to deploy Linux in virtual machines. Early sections acquaint you with z/VM and take you from the point after z/VM installation to the point where you are ready to install your first Linux server. You must turn to the installation documentation that is provided by your Linux distributor. Following the deployment of your first Linux server, you can replicate or clone additional servers. When you finish the early sections, you have two or more Linux servers running on z/VM with TCP/IP connections to the network. Subsequently, you can turn to vendor-supplied documentation to install applications on Linux. Later sections cover operations, administration, performance, guest relocation, and other day-to-day bare essentials. See Section Planning for SSI clusters and LGR in Chapter 2 for information about additional planning guidelines for an SSI cluster and LGR. You can access this document at this website: http://publibz.boulder.ibm.com/epubs/pdf/hcsx0c10.pdf
capabilities include support for the enhanced directory syntax for SSI, conversion of directory contents to aid transition to an SSI cluster, and assistance in adjusting the directory when adding a member to the cluster. To support SSI clusters, several DIRMAINT commands are updated and three new operands are added: SSI operand of the DIRMAINT command to prepare a source directory to use on a node in an SSI cluster. UNDOSSI operand of the DIRMAINT command to roll back BUILD statement changes made by the SSI operand and to remove the SSI operand from the DIRECTORY statement. VMRELOCATE operand to query, update, or delete the relocation capability that is associated with a user or profile entry. ADD operand is updated for cloning SUBCONFIG entries. A new SSI_VOLUMES section is added to the EXTENT CONTROL file to support the ADD operand. You can access this document at this website: http://publibz.boulder.ibm.com/epubs/pdf/hcsl4c10.pdf
Information is added about new and changed CP commands and utilities in support of SSI cluster configuration and management. The following CP commands are added or changed for this version and release (this list is not exhaustive): AT DEFINE RELODOMAIN MSG has a new AT operand. QUERY CHPIDV QUERY NAMES has a new AT operand. QUERY RELODOMAIN QUERY SSI QUERY VIRTUAL CHPID QUERY VMRELOCATE SET VMRELOCATE SET SSI SET CPTRACE changed. New trace codes for SSI cluster operations are added. VMRELOCATE WNG has a new AT operand. The following CP utilities are added or changed for this version and release: FORMSSI is new. CPSYNTAX is changed. A logical partition (LPAR) operand is added, and enhancements are added to support the new SYSTEM CONFIG file statements SSI, BEGIN, and END. See Support for z/VM Single System Image Clusters in the Summary of Changes of this manual for information about commands and utilities: Cross-system SPOOL and SCIF ISFC infrastructure enhancements LGR Multiple access ports per guest Real device mapping Shared disk enhancements SSI user identity and configuration Virtual networking support for an SSI cluster You can access this document at this website: http://publibz.boulder.ibm.com/epubs/pdf/hcse4c10.pdf
Chapter 1. Introduction
It provides guidance and procedures for certain migration tasks that you might need to perform. Topics on SSI are included in this book: System changes: The chapter is redesigned. It now lists subchapters preceded by their z/VM version and release level. It is much easier to look up a task and then isolate your specific needs by version level: Installation, migration, and service Service changes to support the SSI environment Support and exploitation of hardware and architectures Information on SSI cluster toleration within an ensemble Connectivity and networking Virtual networking support for an SSI cluster System administration and operation: SSI cluster configuration and management: A list of added or updated functions SSI cluster user identity and configuration: Descriptions of a new type of virtual machine definition, changes to the layout of the system minidisks, and changes to the MAINT user ID A description of the changes implemented to provide LGR in an SSI cluster
An important appendix (Appendix C) that explains the importance of upgrading a CSE system to z/VM V6.2 You can access this document at this website: http://publibz.boulder.ibm.com/epubs/pdf/hcsf9c10.pdf
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
Chapter 2.
10
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
Service updates can be planned over a period of time and can be applied to a single member of the cluster, and then rolled out to other members of the cluster later. Figure 2-1 shows how service is applied to an SSI cluster.
11
12
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
Shared volumes
M01RES
M01P01 Paging
M02RES
M01PV1
Private Mdisk
M01T01 T-Disk
VMCOM1
VMCOM2
620RL1
M02PV1
Private Mdisk
M02T01 T-Disk
M01S01 Spool
M02S01 Spool
M03S01 Spool
M04S01 Spool
M03RES
M04RES
M03PV1
Private Mdisk
M03T01 T-Disk
M04PV1
Private Mdisk
M04T01 T-Disk
To prevent members from allocating control program (CP) data on a volume owned by another member, CPFMTXA has been modified to mark CP-owned volumes with ownership information, SSI cluster, and owning SSI member. Example 2-1 shows the ownership information of a non-shared CP-owned volume SSI5RS on our ITSOSSI5 system.
Example 2-1 Non-shared CP-owned volume
cpfmtxa 123 ssi5rs alloc HCPCCF6209I INVOKING ICKDSF. ICK030E DEFINE INPUT DEVICE: FN FT FM, "CONSOLE", OR "READER" CONSOLE . . . ICK03000I CPVOL REPORT FOR 0123 FOLLOWS: ICK03021I 0123 IS FORMATTED FOR VM/ESA MODE CYLINDER ALLOCATION TYPE START -------PERM 0 DRCT 1 PERM 21 PARM 39 PARM 159 PARM 160 PERM 280 ICK03092I VOLUME VM SSI OWNER CURRENTLY IS AS FOLLOWS: END TOTAL ------0 1 20 20 38 18 158 120 159 1 279 120 10016 9737 = ITSOSSIB
Chapter 2. Single system image (SSI) overview
13
ICK03093I VOLUME SYSTEM OWNER = ITSOSSI5 ICK00001I FUNCTION COMPLETED, HIGHEST CONDITION CODE WAS 0 Most CP-owned volumes that are shared by the cluster have ownership of the SSI cluster and NOSYS. However, this ownership does not apply to all CP-owned volumes. For example, spool volumes have ownership of the SSI cluster and the member. Example 2-2 shows the ownership information of the common volume SSIBC1 on our ITSOSSIB cluster.
Example 2-2 CP-owned volume shared by the cluster
cpfmtxa 141 ssibc1 alloc HCPCCF6209I INVOKING ICKDSF. ICK030E DEFINE INPUT DEVICE: FN FT FM, "CONSOLE", OR "READER" CONSOLE . . . ICK03000I CPVOL REPORT FOR 0141 FOLLOWS: ICK03021I 0141 IS FORMATTED FOR VM/ESA MODE CYLINDER ALLOCATION TYPE START -------PERM 0 PARM 1 PERM 121 ICK03092I VOLUME VM SSI OWNER ICK03093I VOLUME SYSTEM OWNER ICK00001I FUNCTION COMPLETED, CURRENTLY IS AS FOLLOWS: END TOTAL ------0 1 120 120 10016 9896 = ITSOSSIB = (NO OWNER ASSIGNED) HIGHEST CONDITION CODE WAS 0
14
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
ITSOSSI5: Begin RDEVICE 3080-308F EQID OSA1 TYPE OSA DEFINE VSWITCH VSWITCH1 RDEV 3083 DEFINE VSWITCH VSW999 RDEV 3086 IP DEFINE VSWITCH VSW199 RDEV 3089 IP
2 1
15
ITSOSSI5: End
ITSOSSI6: Begin RDEVICE 2100-210F EQID OSA1 TYPE OSA DEFINE VSWITCH VSWITCH1 RDEV 2103 DEFINE VSWITCH VSW999 RDEV 2106 IP DEFINE VSWITCH VSW199 RDEV 2109 IP ITSOSSI6: End
2 1
The EQID concept also helps determine which devices are eligible for relocation to ensure data integrity.
USERPREFIX USERPREFIX specifies the 3-byte prefix that is used when user-defined locally administered MAC addresses are generated. It must be six hexadecimal digits in the range 020000 - 02FFFF. The USERPREFIX value must be identical for all members of the cluster, so the same MAC address is created when the virtual machine is logged on to any member. The USERPREFIX value for the cluster cannot be the same as the MACPREFIX value for any member. The default USERPREFIX value in an SSI cluster is 020000. MACID This MACID (three bytes) is appended to the system MACPREFIX or USERPREFIX (three bytes) to form a unique MAC address for the virtual switch. If no MACID is set for the virtual switch, CP generates a unique identifier for this virtual switch.
16
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
In many cases, you can use the default values for the attributes MACID, MACPREFIX, and USERPREFIX, because z/VM ensures that the MAC addresses are unique. If you require a predictable MAC address assignment for virtual NICs on your system, you must declare MACPREFIX and USERPREFIX statements in the SYSTEM CONFIG. Nevertheless, we suggest that you define at least a unique MACPREFIX for each system to ensure that you have unique MAC addresses across your environment. See Example 2-4 for an example of setting these attributes.
Example 2-4 Define MACPREFIX and USERPREFIX in SYSTEM CONFIG
ITSOSSI5: VMLAN MACPREFIX 025555 USERPREFIX 02AAAA ITSOSSI6: VMLAN MACPREFIX 026666 USERPREFIX 02AAAA For more information about MAC address definition, see z/VM V6R2.0 CP Planning and Administration, SC24-6178-02.
17
Cluster-wide disks One set per cluster Member 1 (PMAINT 141) VMCOM1 IPL
M01RES MAINT CF1 CPLOAD Warm start Checkpoint Object Directory MAINT 190 / 193 M01P01 Paging MAINT 19D / 19E
Member 2
PDR
PMAINT CFO System Config PMAINT 41D VMSES/E PMAINT 2CC Source Directory
IPL
MAINT CF1 CPLOAD Warm start Checkpoint Object Directory MAINT 190 / 193 MAINT 19D / 19E M02P01 Paging
620R L1 MAINT 620 490 / 493 MAINT 620 51D MAINT 620 CF2
M01S01 Spool
M02S01 Spool
formssi display 141 HCPPDF6618I Persistent Data Record on device 0141 (label SSIBC1) is for ITSOSSIB HCPPDF6619I PDR state: Unlocked HCPPDF6619I time stamp: 11/14/11 10:04:37 HCPPDF6619I cross-system timeouts: Enabled HCPPDF6619I PDR slot 1 system: ITSOSSI5 HCPPDF6619I state: Joined HCPPDF6619I time stamp: 11/14/11 10:04:37 HCPPDF6619I last change: ITSOSSI5 HCPPDF6619I PDR slot 2 system: ITSOSSI6 HCPPDF6619I state: Joined HCPPDF6619I time stamp: 11/14/11 10:04:24 HCPPDF6619I last change: ITSOSSI6
18
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
It contains information about member status and the heartbeat data. It is also used for health-checking and ensures that a stalled or stopped member can be detected.
19
volumes are allocated to each member of the cluster, the spool volumes are shared by all members in the SSI cluster. See Figure 2-3 on page 13.
2.3.3 ISFC
All members of the SSI cluster are part of the same ISFC collection. Every member of the SSI cluster must have direct ISFC connections to every other member of the SSI cluster. The number of required CTC adapters depends on the amount of data that needs to be accessed on a continuous basis, the size of the Linux guests being relocated, and the time available to relocate a Linux guest. You can improve the relocation time and quiesce time by adding additional channel paths. The preferred practice is to have multiple CTCs distributed on multiple FICON channel paths between each pair of members. This practice avoids write collisions and provides connectivity redundancy in the event of a failure. The preferred practice is to use the same real device number for the same CTC on each member. It is possible to have multiple virtual switches (VSwitches), but the definitions need to be duplicated across all the members participating in the SSI cluster. All members in the SSI cluster must have identical network connectivity and connect to the same physical LAN segments and the same SAN fabric. The VSwitch name, type, and virtual LAN (VLAN) settings must also be identical across all members of the SSI cluster if Linux guests are to be relocated successfully. The MAC address assignments must also be coordinated across the SSI cluster.
20
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
q ssi SSI Name: ITSOSSIB SSI Mode: Stable Cross-System Timeouts: Enabled SSI Persistent Data Record (PDR) device: SSIBC1 on 9D20 SLOT SYSTEMID STATE PDR HEARTBEAT RECEIVED HEARTBEAT 1 ITSOSSI5 Joined 11/11/11 11:10:37 11/11/11 11:10:37 2 ITSOSSI6 Joined 11/11/11 11:10:39 11/11/11 11:10:39 3 -------- Available 4 -------- Available
q ssi SSI Name: ITSOSSIB SSI Mode: Stable Cross-System Timeouts: Enabled SSI Persistent Data Record (PDR) device: SSIBC1 on 9D20 SSI Persistent Data Record (PDR) device: SSIBC1 on 9D20 SLOT SYSTEMID STATE PDR HEARTBEAT RECEIVED HEARTBEAT 21
1 ITSOSSI5 Down (shut down successfully) 2 ITSOSSI6 Joined 11/11/11 11:45:09 11/11/11
11:45:09
For more information about member states, see z/VM V6R2.0 CP Planning and Administration, SC24-6178-02.
22
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
Chapter 3.
23
24
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
Relocation options Relocation options can influence the relocation and quiesce time for relocation. See section 3.6.1, VMRELOCATE command on page 31 for a short description of these options. Other non-relocation activity Other non-relocation activity on source and destination systems might make relocation or quiesce time longer.
We studied the following supported configurations: A Linux guest running in an ESA or XA virtual machine in ESA/390 or IBM zArchitecture mode A Linux guest running with a disconnected virtual console A Linux guest that was defined as a single configuration virtual machine, that is, the virtual machine definition in the directory begins with a USER statement A Linux guest with virtual processor types of all CP or all Integrated Facility for Linux (IFL) A Linux guest that we IPLed from a device or IPLed from a named saved system (NSS) that meets the NSS criteria identified in the next item of this list A Linux guest with a discontiguous saved segment (DCSS) or NSS as long as the DCSS or NSS contains no shared write (SW), shared with CP (SC), or shared with no data (SN) page ranges, and the identical DCSS or NSS is available on the destination system A Linux guest with supported devices that are available on the destination member: Dedicated devices, including Open Systems Adapter (OSA) and Fibre Channel Protocol (FCP) devices, provided that same device address is available on the destination system. Private virtual disks in storage (created with the DEFINE VFB-512 command). Virtual unit record devices with no open spool files other than open console files. If the guest has an open console file, the relocation process closes the file on the source system and opens a new file on the destination system. Virtual network interface cards (NICs) coupled to active virtual switches. The following facilities are supported: Cryptographic adapter when enabled by CRYPTO APVIRTUAL statement and AP type must be the same on the source and the destination IUCV connections to *MSG and *MSGALL CP system services Application monitor record (APPLDATA) collection if guest buffer is not in a shared DCSS Single console image facility
Chapter 3. Live guest relocation (LGR) overview
25
Figure 3-1 Guest storage must fit into the available space on the destination member
You must perform additional checks before the relocation starts. Nevertheless, the following checks can be overruled by using the FORCE STORAGE operand with the VMRELOCATE command. But, use the option FORCE STORAGE with caution. See 3.6.1, VMRELOCATE command on page 31. The following checks are additional: The guest storage size must be less than the auxiliary paging capacity on the destination. See Figure 3-2.
Figure 3-2 Guest storage size must be less than the auxiliary paging capacity on the destination
z/VM 5.4 implemented a functionality that allowed you to configure reserved and standby storage before you IPL a guest operating system in the virtual machine. This functionality allows the guest to exploit dynamic storage reconfiguration for specific instructions. This 26
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
functionality is considered in the following checks. Figure 3-3 shows that the guest maximum storage size must be less than all storage and paging capacity available on the destination. Figure 3-4 shows that the guest maximum storage size must be less than the paging capacity on the destination.
Figure 3-3 Guest maximum storage size is less than destination available storage and paging capacity
Figure 3-4 Guest maximum storage size must be less than the paging capacity on the destination
Paging requirements: For certain usage scenarios of LGR, such as moving guests to another z/VM member for maintenance reasons, LGR implies increased paging requirements. Therefore, available paging space must be at least two times the total virtual memory of all guests, including guests to be relocated to this member. Avoid allocating more than 50% of available paging space. If the size of guests to be relocated increases the in-use amount of space to greater that 50%, system performance can be affected.
27
members VMSYS02 and VMSYS04. And, domain Z includes members VMSYS03 and VMSYS04.
The following architectural features are used in this example: GIEF EX-EXT FLOAT-PT E-MONITOR FACILITY X General-instructions-extension facility Execute-extensions facility Floating-point extensions facility Enhanced-monitor facility Possible future facility
Remember: The features that we illustrate are examples only. You do not need to understand these features. These features are used to explain the concepts of domains. Figure 3-5 shows the following characteristics: The default domain for single configuration virtual machines (USER) is the cluster-wide relocation domain SSI, which contains all four members and has an architecture level containing the common architectural features. The default domain for multiconfiguration virtual machines (IDENTITY) is the single member domains, for example, VMSYS3 and contains the architectural level of that member. Dependent on the cluster member, where the guest is logged on, it gets those architectural features. Such multiconfiguration virtual machines cannot be relocated. Linux guests (defined as USERs), which belong to the default domain SSI, can use the common architectural features of the SSI domain (GIEF and EX-EXT). Linux guests can be relocated between all members of the cluster. Software fencing and hardware blocking facilities enforce that only these features are available for the Linux guests in that domain.
29
As soon as this guest has to use more architectural features than provided in the default domain SSI, for example, the Linux guest has to use the additional architecture feature FACILITY X, the guest has to be moved to another relocation domain. In our example, the guest has to be moved to domain X. That means, now the guest only can be relocated between the systems VMSYS02 and VMSYS03 of domain X. The way to move guests to another relocation domain is shown in Example 3-4 on page 31. If a single configuration virtual machine is assigned to relocation domain X and logs on to member VMSYS04, which is not included in domain X, CP assigns a virtual architecture level. This virtual architectural level includes the features that are common to domain X and member VMSYS04, which are GIEF and EX-EXT. For further details, see Chapter 27 of z/VM V6R2.0 CP Planning and Administration, SC24-6178-02.
For a permanent definition, place the RELOCATION_DOMAIN statement in your SYSTEM CONFIG file. Example 3-2 shows the RELOCATION_DOMAIN statement for domainA in our cluster.
Example 3-2 Relocation_domain statement
RELOCATION_DOMAIN domainA members itsossi1 itsossi2 Your existing relocation domains can be checked with the QUERY RELODOMAIN command. Example 3-3 shows the output from the query relodomain command on our cluster.
Example 3-3 Query Relodomain command
q relodomain all DOMAIN DOMAINA MEMBERS: ITSOSSI1 ITSOSSI2 DOMAIN ITSOSSI1 MEMBERS: ITSOSSI1 DOMAIN ITSOSSI2 MEMBERS: ITSOSSI2 DOMAIN ITSOSSI3 MEMBERS: ITSOSSI3 DOMAIN ITSOSSI4 MEMBERS: ITSOSSI4 DOMAIN SSI MEMBERS: ITSOSSI1 ITSOSSI2 ITSOSSI3 ITSOSSI4 Ready; T=0.01/0.01 17:28:35
30
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
set vmrelocate itsolnx1 on domain domainA Running on member ITSOSSI1 Relocation enabled in Domain DOMAINA Ready; T=0.01/0.01 17:29:57 It is best to assign the virtual server to its appropriate relocation domain permanently in its directory entry so that its virtual architecture level is determined at logon time. Example 3-5 shows the related dirmaint command. If the logon is allowed to default its domain, because no relationship is indicated in the directory entry, and then changed dynamically at a later time, the virtual architecture level of the server can be changed during server operations. The server might not handle such a dynamic change gracefully.
Example 3-5 Permanent setting of a relocation domain for a Linux guest
31
Execute relocations one at a time (SYNCHRONOUS) so that they do not compete with each other for system resources. This competition might increase relocation times and quiesce times. IMMEDIATE option The IMMEDIATE option causes the quiesce to occur after only one pass through memory. This option usually results in shorter overall relocation times, but longer quiesce times. It might be necessary to move more pages of memory during the quiesce phase depending on the activity on the Linux guest. MAXTOTAL/MAXQUIESCE option: The MAXTOTAL time is the maximum total time (in seconds) that the command issuer is willing to wait for the entire relocation to complete. If a relocation needs more time, the relocation is canceled. The default MAXTOTAL time is NOLIMIT. The MAXQUIESCE time is the amount of time (in seconds) that a Linux guest can be stopped during a relocation attempt. If a relocation needs more quiesce time, the relocation is canceled. The default MAXQUIESCE time is 10 seconds. If the option IMMEDIATE is specified, the default MAXQUIESCE time is NOLIMIT. FORCE options: Use these options with caution: FORCE ARCHITECTURE: This option forces the relocation even if the destination cluster member does not support all architectural features of the source cluster member. The option can be used to allow out-of-domain relocation to a cluster member that does not support all the architectural features of the source cluster member. FORCE DOMAIN: This option forces the relocation even if the destination cluster member is an out-of-domain member. This option does not permit a relocation outside of the domain if the move causes any loss of architectural features. In this case, you have to use the FORCE ARCHITECTURE DOMAIN option. FORCE STORAGE: In 3.4, Memory and paging considerations for live guest relocation on page 26, we described the checks that were run during the eligibility checks of the relocation. The checks, which are listed as additional can be overruled with the FORCE STORAGE option. See Figure 3-2 on page 26, Figure 3-3 on page 27, and Figure 3-4 on page 27.
Relocation steps
Perform the following steps for the relocation: 1. Choose one class A user ID from which to issue all your relocation commands. Issuing the VMRELOCATE command from one user ID helps keep your relocations organized and helps you avoid issuing multiple relocations at one time. 2. For each virtual server that you want to relocate, determine whether there is a maximum amount of time that the server can be quiesced. Look at the applications running on the virtual server to determine whether any timeout values exist for those applications. For instance, a web server might need to reply to a request to serve a page within 5 seconds. 3. Issue the VMRELOCATE TEST command. The command tells you any reasons why the Linux guest is ineligible to relocate. If there are any issues, address them and then retry the VMRELOCATE TEST command. Example 3-6 and Example 3-7 show two examples of the VMRELOCATE TEST command.
Example 3-6 VMRELOCATE TEST command with a successful check for eligibility
vmrelocate test itsolnx1 itsossi2 User ITSOLNX1 is eligible for relocation to ITSOSSI2 Ready; T=0.01/0.01 17:31:33
Example 3-7 VMRELOCATE TEST command for a virtual server that is ineligible for relocation
vmrelocate test itsolnx1 itsossi2 HCPRLH1940E ITSOLNX1 is not relocatable for the following reason(s): HCPRLE1950I ITSOLNX1: Virtual machine is not running disconnected HCPRLI1954I ITSOLNX1: Virtual machine device 0009 is a logical device HCPRLL1813I ITSOLNX1: Maximum pageable storage use (20640M) exceeds available auxiliary paging space on destination (14423036K) by 6712324K Ready(01940); T=0.01/0.01 17:40:48 4. Issue the VMRELOCATE MOVE command with the MAXTOTAL and MAXQUIESCE times that you determined are required. The default value for the MAXTOTAL time is NOLIMIT. The default value for the MAXQUIESCE time is 10 seconds. By default, the command is issued synchronously, meaning you cannot issue any other commands while the VMRELOCATE MOVE is ongoing. Example 3-8 on page 33 shows an example of the VMRELOCATE MOVE command.
Example 3-8 VMRELOCATE MOVE command
vmrelocate move itsolnx1 itsossi2 maxtotal 10 maxquiesce 5 Relocation of ITSOLNX1 from ITSOSSI1 to ITSOSSI2 started User ITSOLNX1 has been relocated from ITSOSSI1 to ITSOSSI2 Ready; T=0.01/0.01 17:32:43
33
Guest state conditions: The user ID is logged off or is still logging on. The user ID is not disconnected. The guest is a multiconfiguration virtual machine; the directory definition of the user begins with an IDENTITY statement. Device conditions: The guest configuration includes an unsupported device. The guest has a link to a local minidisk that cannot be linked on the destination site. The guest configuration includes a real FCP, OSA, Hipersocket, or CTC device for which the destination system has no available device with the same EQID. Device state conditions: The guest has a device for which the matching device on the destination system is offline or the device is not available. The guest has a non-full pack minidisk on a device for which the matching real device on the destination system is not attached to the system. Virtual facility conditions: The guest is a coupling facility machine or the guest configuration includes access to a coupling facility service machine. That is, the CFUSER option is specified in the virtual machine definition of the user. Or, the guest has a device that is coupled to a coupling facility service machine. The guest has an inter-user communication vehicle (IUCV) connection to a system service (other than to *MSG or *MSGALL, which are allowed) or to another virtual machine. Configuration conditions: The guest has a dedicated CPU. The guest has dedicated cryptographic capability. The guest configuration includes more than one virtual processor type, which precludes the relocation of any guest with a virtual zIIP or zAAP. The guest is defined with all-type CP virtual processors and the destination is an all-type IFL environment. Resource limit conditions: The guest storage size does not fit into the available space on the destination. The guests virtual disk in storage usage violates the usage restrictions on the destination system.
34
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
Chapter 4.
Scenario one: Creating a single system image cluster with four new z/VM members
In this chapter, we discuss the planning and creation of a single system image (SSI) cluster with four members, all of which are created from the installation media. This scenario includes enabling and configuring the following products: TCP/IP (mandatory installation selection) Programmable Operator (optional installation selection) We also describe the following priced features: Dirmaint (encouraged) Remote Spooling Communications Subsystem (RSCS) (optional) Performance Toolkit (encouraged)
35
4.1 Overview
The scenario described in this chapter is the implementation of a first level four-member SSI cluster. We configured other software products to run in the environment, as well. We set up a Linux guest under one of the members and performed a live guest relocation across the members without any disruption of operations to the Linux guest. To recreate a practical scenario of setting up an SSI cluster and identifying the planning considerations for setting it up, we created a cluster with multiple members. Although an SSI cluster can be set across 1, 2, 3, or 4 members, we describe setting up a four-member SSI cluster and then demonstrate live guest relocation across all four members.
First
Record an"M" if you will load the product to a minidisk or an "F" if you will load the product to the VMSYS file pool in the Install To column
Install To M M M
Install To M M M
Install To M M M
Default system lanuguage: DASD type and model: SCSI Volume size: Common service filepool name: Installation Type: Non-SSI X SSI Number of Members
AMENG 3390-9
SSISFSPL
Because the setup is for a four-member SSI cluster, we plan to have first-level z/VM logical partitions (LPARs) running. Two LPARs will run on a z10EC machine, and two LPARs will run on a z196 machine. We plan to install all products and then customize the products that are
36
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
required for this installation. This method gives us an opportunity to identify any additional installation considerations that are specific to setting up an SSI cluster.
In the setup that we created, ITSOSSI1 and ITSOSSI2 run on the z10 EC machine. and ITSOSSI3 and ITSOSSI4 run on the z196 machine. Figure 4-2 shows the worksheet that we completed for our SSI cluster members.
SSI Member Name(s) / IPL LPAR Name(s) or UserID Name(s): Slot Num ber Member Name IPL LPAR/User ID 1 ITSOSSI1 A1A 2 ITSOSSI2 A1B 3 ITSOSSI3 A2A 4 ITSOSSI4 A2B
Figure 4-2 LPAR allocation and member names across z196 and z10
Operator console addresses that are defined to Console 1 and Console 2: ITSOSSI1 and ITSOSS12: F200 and F280 ITSOSSI3 and ITSOSS14: F200 and F280 DASD that is assigned for this setup: ITSOSSI1,ITSOSS12, ITSOSSI3, and ITSOSS14: 9E20 - 9E2F
Chapter 4. Scenario one: Creating a single system image cluster with four new z/VM members
37
extended count key data (ECKD) volumes, we can allocate fewer disks for common, release, and work volumes. Consider the following information when setting up DASD for an SSI cluster: You need one set of cluster-wide shared common volumes. This set contains shared data files for the SSI cluster, such as the shared system configuration file, shared source directory file, and cluster-wide minidisks. These files are owned by the PMAINT user ID. You need one set of shared release volumes for each z/VM release in the cluster. The minidisks on this volume are owned by the release-specific user ID MAINT620. You need one set of non-shared member-specific system volumes. This set contains member-specific data, such as standard system minidisks that are owned by MAINT, volumes for paging, and temporary disks. You need spool volumes for each member. These spool volumes are shared by the SSI cluster. Figure 4-3 shows our DASD configuration.
SSI1I2 MAINT CF1 CPLOAD Warm s tart Chec kpoint Objec t Directory MAINT 190/193 MAINT 19D / 19E
ITSOSSI1 Cluster-wide disks One set per cluster ( PMAINT 141 ) IPL
SSIAC2 PDR PMAINT CF0
System Config
ITSOSSI3
SSI3I2 MAINT CF1 CPLOAD Warm s tart Checkpoint Object Direc tory MAINT 190/193 MAINT 19D / 19E
SSI3P2 Paging
IPL
SSI1S2 Spool
SSI3S2 Spool
PMAINT 41D
VMSES/E
ITSOSSI2
ITSOSSI4
SSI2P2 Paging
W arm start Chec kpoint Objec t Directory MAINT 190/193 MAINT 19D / 19E
SSI4P2 Paging
IPL
IPL
SSI2S2 Spool
SSI4S2 Spool
Figure 4-3 Organization of DASD volumes and minidisks in the SSI cluster ITSOSSIA
To create this setup, we carefully planned the DASD volumes that we used, identifying which volumes were shared and non-shared. For this scenario, we described the available DASD volumes in detail in the planning worksheet for the SSI installation that is shown in Figure 4-4 on page 39.
38
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
Volume Type COMMON COMMON2 RELVOL RELVOL2 Volume Type RES SPOOL PAGE WORK RES SPOOL PAGE WORK Default Label M01RES M01S01 M01P01 M01W01 M03RES M03S01 M03P01 M03W01 New Label Member 1: SSI1I2 SSI1S2 SSI1P2 Member 3: SSI3I2 SSI3S2 SSI3P2
Common Volumes Default Label New Label VMCOM1 SSIAC2 VMCOM2 620RL1 SSIAR2 620RL2 Dedicated Volumes Address Volume Type 9E21 9E25 9E2A Not required for model 9 9E23 9E27 9E2C Not required for model 9 RES SPOOL PAGE WORK RES SPOOL PAGE WORK
Address 9E20 Not required for model 9 9E29 Not required for model 9 Default Label New Label Member 2: M02RES SSI2I2 M02S01 SSI2S2 M02P01 SSI2P2 M02W01 Member 4: M04RES SSI4I2 M04S01 SSI4S2 M04P01 SSI4P2 M04W01 Address 9E22 9E26 9E2B Not required for model 9 9E24 9E28 9E2D Not required for model 9
Note: You must not use any of IBM's default volume labels for a volume other than the volume for which it is orginially defined
z196
ITSOSSI3(A2A)
9E2C
SSI1I2
RES
SSI1P2
PAGE Private MDISK
Shared Volumes
9E20 9E29
SSI1I2
RES
SSI1P2
PAGE Private MDISK
Dedicated Volumes
Dedicated Volumes
SSIAC2
COMMON
SSIAR2
RELVOL
9E25
9E26
9E27
9E28
SSI1S2
SPOOL
SSI2S2
SPOOL
SSI3S2
SPOOL
SSI4S2
SPOOL
ITSOSSI2(A1B)
9E22 9E2B
Shared MDISK
ITSOSSI4(A2B)
9E2D
SSI1I2
RES
SSI1P2
PAGE Private MDISK
SSI1I2
RES
SSI1P2
PAGE Private MDISK
Dedicated Volumes
Dedicated Volumes
Chapter 4. Scenario one: Creating a single system image cluster with four new z/VM members
39
Configuring ISFC
Table 4-1 show the SSI cluster connections that are available for ISFC CTC.
Table 4-1 SSI cluster connections for ISFC ITSOSSI1 4A60-4A63 4A68-4A6B 5A60-5A63 5A68-5A6B 43A0-43A3 43A8-43AB 53A0-53A3 53A8-53AB 43B0-43B3 43B8-43BB 53B0-53B3 53B8-53BB 43A0-43A3 43A8-43AB 53A0-53A3 53A8-53AB 43B0-43B3 43B8-43BB 53B0-53B3 53B8-53BB 5A60-5A63 5A68-5A6B 4A60-4A63 4A68-4A6B 5A60-5A63 5A68-5A6B 4A60-4A63 4A68-4A6B ITSOSSI2 5A50-5A53 5A58-5A5B 4A50-4A53 4A58-4A5B 5A50-5A53 5A58-5A5B 4A50-4A53 4A58-4A5B 5A50-5A53 5A58-5A5B 4A50-4A53 4A58-4A5B ITSOSSI3 ITSOSSI4
40
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
ITSOSSI1
ITSOSSI2
ITSOSSI3
ITSOSSI4
The allocation of the ISFC connections across the members is one of the most crucial steps in installation planning. Figure 4-6 shows the allocation of the CTC addresses based on Table 4-1 on page 40.
Real addresses for the C OMMON volume on each mem ber LPAR Member 2 Member 3 Address Address 9E20 9E20
Member 1 Address 9E20 CTC Device addresses: From: To: To: To: To: Member 1 Member 1 Member 2 Member 3 Member 4
From: N/A 4A60 53A8 43B2 5A60 43A8 53B2 To: To: To: To:
Member 2 Member 1 Member 2 Member 3 Member 4 53A2 43B8 5A50 N/A 43A2 53B8 4A50
Member 3 Member 1 Member 2 Member 3 Member 4 53BA 4A58 4A62 N/A 43BA 5A58 5A62
Member 4 Member 1 Member 2 Member 3 Member 4 5A52 5A68 43AA N/A 4A52 4A68 53AA
Figure 4-7 on page 42 shows a graphical representation of the SSI CTC allocation for ISFC.
Chapter 4. Scenario one: Creating a single system image cluster with four new z/VM members
41
Common LAN
z10EC
ITSOSSI1 4A60 5A60 53A8 43A8 43B2 53B2 CTCs for ISFC 4A58 5A58 4A62 5A62
z196
ITSOSSI3 53BA 43BA
42
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
Table 4-2 SSI cluster connections for RSCS ITSOSSI1 4A90-4A93 4A98-4A9B 5A90-5A93 5A98-5A9B ITSOSSI2 ITSOSSI3 ITSOSSI4 WTSCVMXA 4A50-4A53 5A58-5A5B 4A50-4A53 4A58-4A5B
Figure 4-8 on page 44 shows a graphical representation of the SSI CTC allocation for RSCS.
Chapter 4. Scenario one: Creating a single system image cluster with four new z/VM members
43
Common LAN
z10EC
5A9B ITSOSSI1 5A6B 53AB 53BB CTCs for RSCS 4A5B 4A6B
z196
ITSOSSI3 53BB 5A9B
4A5B 5A9B
53AB
4A5B 4A63
ITSOSSI2 53B3
43BB 43AB
Configuring TCP/IP
Table 4-3 on page 45 shows the TCP/IP setup that is based on the input data that is provided for this scenario.
44
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
Table 4-3 TCP/IP setup sheet for the four members Host name Domain name Domain name server (DNS) IP Gateway IP Interface name Device number IP address Subnet mask Maximum transmission unit (MTU) discovery Interface Router type Router advertisements MTU size ITSOSSI1 itso.ibm.com 9.12.6.7 9.12.4.1 OSA1 3080-3082 9.12.4.232 255.255.252.0 Enabled ITSOSSI2 itso.ibm.com 9.12.6.7 9.12.4.1 OSA1 30A0-30A2 9.12.4.233 255.255.252.0 Enabled ITSOSSI3 itso.ibm.com 9.12.6.7 9.12.4.1 OSA1 2100-2102 9.12.4.234 255.255.252.0 Enabled ITSOSSI4 itso.ibm.com 9.12.6.7 9.12.4.1 OSA1 2120-2122 9.12.4.235 255.255.252.0 Enabled
4.2.6 Guests
You must plan for the Linux guests that participate in live guest relocation (LGR), as well. Linux guests must meet certain eligibility criteria and requirements for the destination member: The guest operating system must be supported for relocation. The source and target server environments for the relocatable members must be comparable and architecturally supported. If relocation domains are established and certain policies are set, the relocation must comply with these relocation domains and policies. The destination environment must have sufficient system resources to accommodate the relocating guest. System resources that are required by the relocating guest must be available on the target environment.
Chapter 4. Scenario one: Creating a single system image cluster with four new z/VM members
45
1: 2: 3: 4:
AFTER INTSALLATION IS COMPLETE, MEMBERS WILL BE IPLed FROM: First-Level THE DASD TYPE YOU SELECTED TO LOAD ON IS: 3390 Mod 9 THE VOLUMES NEEDED TO LOAD z/VM ARE: COMMON:VMCOM1 RELEASE:620RL1 MEMBER1:M01RES MEMBER2:M02RES MEMBER3:M03RES MEMBER4:M04RES
46
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
The DASD, as planned in Figure 4-4 on page 39. In this case, we defined 14 numbers for model 3390-9 DASD. Figure 4-6 on page 41 shows the address of the common volume for each member, which, in this case, is the same address for all members. Figure 4-6 on page 41 also shows the CTC connections for the entire cluster. At installation time, we can only define two connections between each member. We can add more connections after the installation is complete, if required. The last window from the INSTPLAN EXEC command shows that the DASD will be formatted and the connections will be defined based on the data input.
Chapter 4. Scenario one: Creating a single system image cluster with four new z/VM members
47
HCPPLM1698I The mode of the SSI cluster is STABLE USER DSC LOGOFF AS AUTOLOG1 USERS = 9 Q SSI SSI Name: ITSOSSIA SSI Mode: Stable Cross-System Timeouts: Enabled SSI Persistent Data Record (PDR) device: SSIAC2 on 9E20 SLOT SYSTEMID STATE PDR HEARTBEAT RECEIVED HEARTBEAT 1 ITSOSSI1 Joined 11/02/11 16:37:36 11/02/11 16:37:36 2 ITSOSSI2 Joined 11/02/11 16:37:46 11/02/11 16:37:46 3 ITSOSSI3 Joined 11/02/11 16:37:48 11/02/11 16:37:48 4 ITSOSSI4 Joined 11/02/11 16:37:44 11/02/11 16:37:44 LOG CONNECT= 00:03:12 VIRTCPU= 000:00.00 TOTCPU= 000:00.04 LOGOFF AT 16:40:19 EDT WEDNESDAY 11/02/11
Figure 4-11 SSI cluster stable with four members
With all members running, we logged on to MAINT620 on the Hardware Management Console (HMC) of each member. We ran the program IPWIZARD to set up TCP/IP, which enabled us to log on to the members of the cluster without having to use the HMC. We configure TCP/IP for normal use later in this section.
48
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
Now, we can define the additional CTC devices, which we did not include earlier as part of the ISFC setup. Example 4-1 shows sections of the SYSTEM CONFIG file after all the changes are made. Appendix C, Additional material on page 113 includes the full SYSTEM CONFIG file. This file resides on minidisk CF0, which is owned by the PMAINT machine.
Example 4-1 SYSTEM CONFIG with four members
/**********************************************************************/ /* SYSTEM CONFIG FILE */ /**********************************************************************/ . . /**********************************************************************/ /* System_Identifier Information */ /**********************************************************************/ System_Identifier System_Identifier System_Identifier System_Identifier LPAR LPAR LPAR LPAR A1A A1B A2A A2B ITSOSSI1 ITSOSSI2 ITSOSSI3 ITSOSSI4
/**********************************************************************/ /* SSI Statement required for VMSSI feature */ /**********************************************************************/ SSI ITSOSSIA SLOT 1 SLOT 2 SLOT 3 SLOT 4 PDR_VOLUME SSIAC2, ITSOSSI1, ITSOSSI2, ITSOSSI3, ITSOSSI4
/**********************************************************************/ /* Checkpoint and Warmstart Information */ /**********************************************************************/ ITSOSSI1: System_Residence, Checkpoint Volid Warmstart Volid System_Residence, Checkpoint Volid Warmstart Volid System_Residence, Checkpoint Volid Warmstart Volid System_Residence, Checkpoint Volid Warmstart Volid
From CYL 21 From CYL 30 From CYL 21 From CYL 30 From CYL 21 From CYL 30 From CYL 21 From CYL 30
ITSOSSI2:
ITSOSSI3:
ITSOSSI4:
Chapter 4. Scenario one: Creating a single system image cluster with four new z/VM members
49
1 1 1 1
/**********************************************************************/ /* DUMP & SPOOL VOLUMES */ . . /**********************************************************************/ CP_Owned CP_Owned CP_Owned CP_Owned Slot Slot Slot Slot 10 11 12 13 SSI1S2 SSI2S2 SSI3S2 SSI4S2
/**********************************************************************/ /* PAGE & TDISK VOLUMES */ . . /**********************************************************************/ /* Page and Tdisk volumes for Member 1 */ /**********************************************************************/ ITSOSSI1: ITSOSSI1: BEGIN CP_Owned END
Slot 255
SSI1P2
/**********************************************************************/ /* Page and Tdisk volumes for Member 2 */ /**********************************************************************/ ITSOSSI2: ITSOSSI2: BEGIN CP_Owned END
Slot 255
SSI2P2
/**********************************************************************/ /* Page and Tdisk volumes for Member 3 */ /**********************************************************************/ ITSOSSI3: ITSOSSI3: BEGIN CP_Owned END
Slot 255
SSI3P2
50
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
Slot 255
SSI4P2
/**********************************************************************/ /* Activate ISLINK statements */ /**********************************************************************/ ITSOSSI1: ITSOSSI1: ITSOSSI1: ITSOSSI2: ITSOSSI2: ITSOSSI2: ITSOSSI3: ITSOSSI3: ITSOSSI3: ITSOSSI4: ITSOSSI4: ITSOSSI4: ACTIVATE ACTIVATE ACTIVATE ACTIVATE ACTIVATE ACTIVATE ACTIVATE ACTIVATE ACTIVATE ACTIVATE ACTIVATE ACTIVATE ISLINK ISLINK ISLINK ISLINK ISLINK ISLINK ISLINK ISLINK ISLINK ISLINK ISLINK ISLINK 4A60 53A8 43B2 5A50 53A2 43B8 4A58 4A62 53BA 5A52 5A68 43AA 5A60 43A8 53B2 4A50 43A2 53B8 5A58 5A62 43BA 4A52 4A68 53AA NODE NODE NODE NODE NODE NODE NODE NODE NODE NODE NODE NODE ITSOSSI2 ITSOSSI3 ITSOSSI4 ITSOSSI1 ITSOSSI3 ITSOSSI4 ITSOSSI1 ITSOSSI2 ITSOSSI4 ITSOSSI1 ITSOSSI2 ITSOSSI3
/**********************************************************************/ . . /**********************************************************************/ /* Shared User Volumes */ /**********************************************************************/ User_Volume_List SSIAR2 User_Volume_List LXDE1C LX603D LX9B25 LX9B26 User_Volume_Include SSI* . . /**********************************************************************/ Features , Disable , Set_Privclass , Enable , Auto_Warm_IPL , Clear_TDisk , STP_TZ , Retrieve , Default 60 , Maximum 255 , MaxUsers noLimit , Passwords_on_Cmds , Autolog no , Link no , Logon no , Vdisk Userlim 144000 blocks, Disconnect_timeout off
/* /* /* /* /* /* /* /* /* /* /* /* /* /* /*
Disable the following features Disallow SET PRIVCLASS command Disable the following features No Prompt at IPL always Clear TDisks at IPL time Test for STP Support Retrieve options Default.... default is 20 Maximum.... default is 255 No limit on number of users What commands allow passwords? ... AUTOLOG does ... LINK does ... and LOGON does, too Maximum vdisk allowed per user
*/ */ */ */ */ */ */ */ */ */ */ */ */ */ */
Chapter 4. Scenario one: Creating a single system image cluster with four new z/VM members
51
/**********************************************************************/ /* Relocation Domains /**********************************************************************/ Relocation_Domain DMNZ10 Relocation_Domain DMNZ196 MEMBER ITSOSSI1 ITSOSSI2 MEMBER ITSOSSI3 ITSOSSI4
. . /**********************************************************************/ /* VSWITCH Section */ /**********************************************************************/ ITSOSSI1: Begin VMLAN MACPREFIX 021111 USERPREFIX 02AAAA RDEVICE 3080-308F EQID OSA1SET1 TYPE OSA DEFINE VSWITCH ITSOVSW1 RDEV 3083 ITSOSSI1: End ITSOSSI2: Begin VMLAN MACPREFIX 022222 USERPREFIX 02AAAA RDEVICE 30A0-30AF EQID OSA1SET1 TYPE OSA DEFINE VSWITCH ITSOVSW1 RDEV 30A3 ITSOSSI2: End ITSOSSI3: Begin VMLAN MACPREFIX 023333 USERPREFIX 02AAAA RDEVICE 2100-210F EQID OSA1SET1 TYPE OSA DEFINE VSWITCH ITSOVSW1 RDEV 2103 ITSOSSI3: End ITSOSSI4: Begin VMLAN MACPREFIX 024444 USERPREFIX 02AAAA RDEVICE 2120-212F EQID OSA1SET1 TYPE OSA DEFINE VSWITCH ITSOVSW1 RDEV 2123 ITSOSSI4: End MODIFY MODIFY MODIFY MODIFY MODIFY MODIFY MODIFY MODIFY MODIFY VSWITCH VSWITCH VSWITCH VSWITCH VSWITCH VSWITCH VSWITCH VSWITCH VSWITCH ITSOVSW1 ITSOVSW1 ITSOVSW1 ITSOVSW1 ITSOVSW1 ITSOVSW1 ITSOVSW1 ITSOVSW1 ITSOVSW1 GRANT GRANT GRANT GRANT GRANT GRANT GRANT GRANT GRANT ITSOLNX1 ITSOLNX2 ITSOLNX3 ITSOLNX4 ITSOLNX5 ITSOLNX6 ITSOLNX7 ITSOLNX8 ITSOLNX9
52
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
System_Console ITSOSSI1: End ITSOSSI2: Begin Operator_Consoles Emergency_Message_Consoles ITSOSSI2: End ITSOSSI3: Begin Operator_Consoles Emergency_Message_Consoles ITSOSSI3: End ITSOSSI4: Begin Operator_Consoles Emergency_Message_Consoles ITSOSSI4: End /**********************************************************************/ We tested this CONFIG file with the new format of the CPSYNTAX command, as shown in Example 4-2.
Example 4-2 CPSYNTAX command
cpsyntax system config z (lpar a1b CONFIGURATION FILE PROCESSING COMPLETE Ready; T=0.27/0.27 14:54:19 cpsyntax system config z (lpar a2b CONFIGURATION FILE PROCESSING COMPLETE Ready; T=0.27/0.27 14:54:23 cpsyntax system config z (lpar a2a CONFIGURATION FILE PROCESSING COMPLETE Ready; T=0.27/0.27 14:54:28 cpsyntax system config z (lpar a1a CONFIGURATION FILE PROCESSING COMPLETE Ready; T=0.27/0.27 14:54:39
-- NO ERRORS ENCOUNTERED.
-- NO ERRORS ENCOUNTERED.
-- NO ERRORS ENCOUNTERED.
-- NO ERRORS ENCOUNTERED.
Next, we restarted all the members of the cluster. Example 4-3 is the command to shut down to restart all the members. We omitted several lines to improve the clarity of this example.
Example 4-3 Shut down
q ssi SSI Name: ITSOSSIA SSI Mode: Stable Cross-System Timeouts: Enabled SSI Persistent Data Record (PDR) device: SSIAC1 on 9A20
Chapter 4. Scenario one: Creating a single system image cluster with four new z/VM members
53
SLOT SYSTEMID STATE PDR HEARTBEAT RECEIVED HEARTBEAT 1 ITSOSSI1 Joined 10/31/11 15:27:16 10/31/11 15:27:16 2 ITSOSSI2 Joined 10/31/11 15:26:50 10/31/11 15:26:50 3 ITSOSSI3 Joined 10/31/11 15:27:14 10/31/11 15:27:14 4 ITSOSSI4 Joined 10/31/11 15:27:06 10/31/11 15:27:06 Ready; T=0.01/0.01 15:27:19 at itsossi1 cmd shutdown reipl HCPSHU960I System shutdown may be delayed for up to 530 seconds Ready; T=0.01/0.01 15:27:47 at itsossi2 cmd shutdown reipl HCPSHU960I System shutdown may be delayed for up to 530 seconds Ready; T=0.01/0.01 15:28:06 at itsossi4 cmd shutdown reipl HCPSHU960I System shutdown may be delayed for up to 530 seconds Ready; T=0.01/0.01 15:28:10 q ssi SSI Name: ITSOSSIA SSI Mode: Stable Cross-System Timeouts: Enabled SSI Persistent Data Record (PDR) device: SSIAC1 on 9A20 SLOT SYSTEMID STATE PDR HEARTBEAT RECEIVED HEARTBEAT 1 ITSOSSI1 Down (shut down successfully) 2 ITSOSSI2 Down (shut down successfully) 3 ITSOSSI3 Joined 10/31/11 15:27:44 10/31/11 15:27:44 4 ITSOSSI4 Down (shut down successfully) Ready; T=0.01/0.01 15:28:13 shutdown reipl SYSTEM SHUTDOWN STARTED HCPSHU960I System shutdown may be delayed for up to 530 seconds Ready; T=0.01/0.01 15:42:13
Dirmaint
We installed DIRMAINT using the instructions from the book, z/VM V6R2.0 Getting Started with Linux on System z, SC24-6194. We defined MAINT and MAINT620 as administrators. AUTOLOG1 PROFILE EXEC note: With SSI, the PROFILE EXEC of AUTOLOG1 differs from previous z/VM versions. It was changed to start the DIRMAINT machine on the first member that is loaded after the start-up and the respective DIRMSATx in the other members.
54
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
Example 4-4 shows the main configurations changes to RSCS. We considered the following information in making these changes: SYSTEM NETID must be updated in two places: the MAINT620 490 mdisk (common to the four members) and on the MAINT 190 of each member. After the change to the 190 mdisk, update the segments with the PUT2PROD SAVECMS command in each SSI member. RSCS CONFIG file: The LOCAL definition is used to identify each member, and it does not need the LINKDEF/PARM. The same information is true for PROFILE GCS. In our tests, we do not use RSCSDNS and RSCSAUTH. We commented both RSCSDNS and RSCSAUTH out in the GCS profile. So, AUTOLOG1 will only XAUTOLOG GCS.
Example 4-4 RSCS definitions
---> SYSTEM NETID (MAINT620 490 and MAINT 190) *CPUID NODEID NETID 1ADE50 ITSOSSI1 RSCS 1BDE50 ITSOSSI2 RSCS 2A3BD5 ITSOSSI3 RSCS 2B3BD5 ITSOSSI4 RSCS ---> RSCS CONFIG * Local Nodeid * -----------LOCAL ITSOSSI4 * LINKDEFINE WTSCVMXA LINKDEFINE ITSOSSI1 LINKDEFINE ITSOSSI2 LINKDEFINE ITSOSSI3 LINKDEFINE ITSOSSI4
Application ID -------------* RSCS04 TYPE TYPE TYPE TYPE TYPE NJE NJE NJE NJE NJE LINE LINE LINE LINE LINE 5A9B 4A5B 4A6B 43AB 43AB QUEUE QUEUE QUEUE QUEUE QUEUE PRI PRI PRI PRI PRI NODE NODE NODE NODE NODE WTSCVMXA ITSOSSI1 ITSOSSI2 ITSOSSI3 ITSOSSI4
* * Linkid PARM Text * -------- --------------------------------------------------------PARM WTSCVMXA STREAMS=2 MAXU=2 MAXD=10 LISTPROC=NO TA=1 TAPARM=TH=100 PARM ITSOSSI1 STREAMS=2 MAXU=2 MAXD=10 LISTPROC=NO TA=1 TAPARM=TH=100 PARM ITSOSSI2 STREAMS=2 MAXU=2 MAXD=10 LISTPROC=NO TA=1 TAPARM=TH=100 PARM ITSOSSI3 STREAMS=2 MAXU=2 MAXD=10 LISTPROC=NO TA=1 TAPARM=TH=100 PARM ITSOSSI4 STREAMS=2 MAXU=2 MAXD=10 LISTPROC=NO TA=1 TAPARM=TH=100 * What to * Reroute Userid Nodeid Userid * -------------------- -----REROUTE NOTRCVG FOR OPERATOR TO ITSOSSI1 OP1 ---> Profile GCS RSCS START ITSOSSI1 RSCS START ITSOSSI2 RSCS START ITSOSSI3 RSCS START ITSOSSI4 RSCS START WTSCVMXA ---> Profile EXEC, into AUTOLOG1 machine 00000 * * * Top of File * * * 00001 /*********************************************************************/ 00002 /* AUTOLOG1 PROFILE EXEC */
Chapter 4. Scenario one: Creating a single system image cluster with four new z/VM members
55
/*********************************************************************/ -------------------- 10 line(s) not displayed -------------------PIPE CP XAUTOLOG GCS PIPE CP SLEEP 30 SEC
--> Changed PROFILE GCS, into GCS machine /****************************************/ /* Profile GCS for the Recovery Machine */ /****************************************/ /*XAUTOLOG RSCSDNS */ XAUTOLOG RSCS /*XAUTOLOG RSCSAUTH */
Performance Toolkit
We installed the Performance Toolkit program by using the instructions in the Program Directory for Performance Toolkit for VM, which is at this website: http://www.ibm.com/eserver/zseries/zvm/library/ We also used z/VM V6R2.0 Getting Started with Linux on System z, SC24-6194. We used the LOCALMOD command, as instructed, for the FCONX $PROFILE file. So, this file became common for all members. We performed the remainder of the configuration as described in z/VM V6R2.0 Getting Started with Linux on System z, SC24-6194.
TCP/IP
To configure TCP/IP, we copied the PROFILE STCPIP to each TCP/IP 198 mdisk with the respective name of each member. We merged the definitions that were made by IPWIZARD in PROFILE TCPIP to enable FTPSERVE.
56
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
PROFILE LINCMS COMMAND CP SPOOL CONSOLE START TO VMLOGS IPL CMS PARM AUTOCR SPOOL 000C 2540 READER * SPOOL 000D 2540 PUNCH A SPOOL 000E 1403 A CONSOLE 009 3215 T LINK MAINT 0190 0190 RR LINK MAINT 019D 019D RR LINK MAINT 019E 019E RR LINK LINUXCMS 191 191 MR Example 4-6 is the other profile entry. The LINDFLT profile does not include any minidisks that are defined on any local DASD. The LINDFLT profile adds the OPTION CHPIDV that is mandatory to relocate Linux machines. Before we put the Linux guest into production, we only needed to change the INCLUDE from LINCMS to LINDFLT to allow the relocate function to work.
Example 4-6 Profile directory entries for Linux guests
PROFILE LINDFLT CLASS G OPTION CHPIDV ONE VMRELOCATE ON CONSOLE 0009 3215 T SPOOL 000C 2540 READER * SPOOL 000D 2540 PUNCH A SPOOL 000E 1403 A
If the Linux guest needs a PROFILE EXEC, perhaps to run a CMS command before starting Linux, it must detach all local minidisks before the IPL of Linux. When the PROFILE EXEC only contains CP commands, it is not necessary to IPL CMS. You can convert the PROFILE EXEC to command entries in its directory entry. To monitor the Linux guest while it runs, we added the commands that are shown in Example 4-7 to the profile directory entries. In a non-SSI environment, these commands reside in the PROFILE EXEC file of the Linux guest.
Example 4-7 CP commands that we added to the profile directory entry
CP CP CP CP CP
SPOOL CONSOLE START TO VMLOGS TERM LINEND % SET PF12 RETRIEVE TERM MORE 1 0 SET RUN ON
Chapter 4. Scenario one: Creating a single system image cluster with four new z/VM members
57
COMMAND CP TERM HOLD OFF COMMAND CP SET OBSERVER OPERATOR Example 4-8 is the directory entry for the Linux SUSE guest ITSOLNX3.
Example 4-8 Linux guest ITSOLNX3 running SUSE
USER ITSOLNX3 ITSOSSI 6G 6G G * * SLES11 - SP1 * 0201 = swap space * 0202 = / root fs * INCLUDE LINDFLT IPL 202 MACH ESA 2 MDISK 0201 3390 0001 1000 LX9B25 MR MDISK 0202 3390 1001 9016 LX9B25 MR Example 4-9 is the directory entry for the Red Hat guest ITSOLNX4.
Example 4-9 Linux guest ITSOLNX4 running Red Hat
USER ITSOLNX4 ITSOSSI 4G 4G G * * RHEL 5.6 * 0201 = swap space * 0202 = / root fs * INCLUDE LINDFLT IPL 202 MACH ESA 2 MDISK 0201 3390 0001 1000 LX9B26 MR MDISK 0202 3390 1001 9016 LX9B26 MR When these machines are running in disconnected mode, we can issue the VMRELOCATE command to check their eligibility for relocation or move them to any member in the cluster. Example 4-10 on page 59 shows the sequence of commands to use. The first query command shows the two machines running at ITSOSSI2 on the z10. The VMRELOCATE TEST command checks whether the resources that the machines use are available on the target member ITSOSSI3. Command note: You can use the AT<member> CMD command to send the VMRELOCATE command to the member where the machine is running. After a positive response, we transferred the machines using the command VMRELOCATE MOVE. The last two queries in Example 4-10 on page 59 show the machines running on ITSOSSI3 on the z196. All of this activity is transparent to the applications that run on both Linux guests. The PING command did not lose any packets.
58
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
q user itsolnx3 at all ITSOSSI2: ITSOLNX3 - DSC Ready; T=0.01/0.01 13:59:44 q user itsolnx4 at all ITSOSSI2 : ITSOLNX4 - DSC Ready; T=0.01/0.01 13:59:50 at itsossi2 cmd vmrelocate test user itsolnx4 to ITSOSSI3 User ITSOLNX4 is eligible for relocation to ITSOSSI3 Ready; T=0.01/0.01 14:11:24 at itsossi2 cmd vmrelocate test user itsolnx3 to ITSOSSI3 User ITSOLNX3 is eligible for relocation to ITSOSSI3 Ready; T=0.01/0.01 14:11:49 at itsossi2 cmd vmrelocate move user itsolnx3 to ITSOSSI3 Relocation of ITSOLNX3 from ITSOSSI2 to ITSOSSI3 started User ITSOLNX3 has been relocated from ITSOSSI2 to ITSOSSI3 Ready; T=0.01/0.01 14:12:31 at itsossi2 cmd vmrelocate move user itsolnx4 to ITSOSSI3 Relocation of ITSOLNX4 from ITSOSSI2 to ITSOSSI3 started User ITSOLNX4 has been relocated from ITSOSSI2 to ITSOSSI3 Ready; T=0.01/0.01 14:12:58 q user itsolnx3 at all ITSOSSI3: ITSOLNX3 - DSC Ready; T=0.01/0.01 14:13:23 q user itsolnx4 at all ITSOSSI3 : ITSOLNX4 - DSC Ready; T=0.01/0.01 14:13:30 close cons See Appendix A, Frequently asked questions on page 87 for the other scenarios that we tested.
1 2
vmcp is a way to send commands to the Control Program, when running under z/VM. It is part of the s390-tools package. bonnie is a performance benchmark program that targets various aspects of UNIX file systems. It was installed from the basic SUSE software packages.
Chapter 4. Scenario one: Creating a single system image cluster with four new z/VM members
59
60
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
Chapter 5.
Scenario two: Converting two existing stand-alone z/VM systems to a single system image cluster
This chapter describes converting two existing stand-alone z/VM systems into a single system image (SSI) cluster with two members. This section describes our actions in detail, the problems that we encountered, the lessons that we learned, and our suggestions.
61
5.1 Overview
This scenario is typical for most existing clients when they migrate from stand-alone z/VM systems. In this scenario, we created two non-SSI z/VM systems and placed them into an SSI cluster. In the installation methodology, we selected a non-SSI installation in the initial installation panels, which resulted in a single z/VM system. Even though the installation is designed for a single z/VM image, use this method to migrate from a non-SSI environment to an SSI cluster more easily. In addition to the system residence, spool and work volumes, we must specify the common volumes and the release volumes even for a non-SSI installation. Unlike the SSI installation, which requires 3390 volumes, you can perform non-SSI installations on Small Computer System Interface (SCSI) volumes. However, these non-SSI installations need one extended count key data (ECKD) volume for the Persistent Data Record (PDR) and two ECKD volumes for your IBM RACF databases. The IPL packs can still be SCSI, if you are converting from non-SSI to SSI. If you want to migrate your existing stand-alone z/VM systems running z/VM 5.3 or higher, perform the following tasks: 1. Upgrade one of the existing stand-alone z/VM systems to a Version 6.2 stand-alone non-SSI z/VM system. 2. Migrate this z/VM Version 6.2 stand-alone non-SSI system to a single-member SSI cluster. 3. Clone the SSI system. 4. Move the guests from their remaining pre-Version 6.2 stand-alone z/VM systems onto the members of the SSI cluster. In our environment, each z/VM system runs on a separate central processor complex. All DASDs are visible to both logical partitions (LPARs) with common addressing. Figure 5-1 on page 63 shows our scenario.
62
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
ITSOSSIB
5
ITSOSSI5 ITSOSSI6
610zvm
610zvm
ITSOSSI6
8
lnxrh56 lnxrh56
lnxsu12
lnxsu12
610zvm
4
lnxsu11 lnxsu11 lnxrh56
TCPIP
TCPIP
TCPIP
2
RACF VMPTK RACF RSCS RACF RSCS
DIRMAINT
RSCS
DIRMAINT
VMPTK
DIRMAI NT
VMPTK
3 Create a second non-SSI system ITSOSSI6. For simplicity, we created a clone from ITSOSSI5. 4 Create guests on each non-SSI system. We created Linux guests LNXSU11 and LNXSU12 on ITSOSSI5. We created z/VM guest 610ZVM and Linux guest LNXRH56 on ITSOSSI6. 5 Convert the non-SSI system ITSOSSI5 to a single-member z/VM SSI cluster ITSOSSIB. 6 Add an additional member (ITSOSSI6) to the z/VM SSI cluster ITSOSSIB by cloning the existing member ITSOSSI5.
Chapter 5. Scenario two: Converting two existing stand-alone z/VM systems to a single system image cluster
63
7 Move the guests from the non-SSI system ITSOSSI6 (610ZVM and LNXRH56) to the SSI cluster. 8 Start the relocation tests. We describe these steps in more detail in the following sections.
Table 5-1 ITSOSSI5 definition Device type 3390-09 3390-09 3390-09 3390-09 3390-09 Device type OSA OSA OSA TPCIP RSCS Volume serial number SSIBC1 SSIBR1 SSI5RS SSI5S1 SSI5P1 Address 9D20 9D21 9D22 9D23 9D24 Address 3080 3081 3082 9.12.4.236 4A90 Use Common disk Release disk SYSRES volume Spool volume Paging volume
Optional priced features that are preinstalled on the z/VM V6.2 base product media
Chapter 5. Scenario two: Converting two existing stand-alone z/VM systems to a single system image cluster
65
For the System DASD model, we selected 3390 Mod 9. For the FBA DASD size, we used an FBA DASD size of 6.0. For the System Type, we selected SSI Install. For the number of members, we typed 4. For the SSI Cluster Name, we typed ITSOSSIA.
66
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
5.3.3 Installation
We performed the following installation steps.
Optional priced features that are preinstalled on the z/VM V6.2 base product media
Chapter 5. Scenario two: Converting two existing stand-alone z/VM systems to a single system image cluster
67
3. We started ITSOSSI6 as a second-level guest under ITSOSSI5, ensuring that the correct set of volumes was attached to the user ID from which it was started. First, we performed these steps: a. To start ITSOSSI6, we wrote a new copy of the stand-alone program loader on ITSOSSI6 sysres, because we needed to modify the common volume address. We wrote this new copy by entering the command: salipl 9d26 (extent 1 iplparms fn=SYSTEM ft=CONFIG pdnum=1 pdvol=9c25 b. We redefined the console 009 as 020 to enter the IPL parameters CLEAN NOAUTOLOG. c. We performed the IPL from 9D26, which is the ITSOSSI6 sysres. 4. We performed XAUTOLOG RACFVM to log on to MAINT620. From MAINT620, we changed the label and the owner of all the CP-owned volumes to reflect the ITSOSSI6 configuration by issuing the following commands: CPFMTXA 123 CPFMTXA 122 CPFMTXA 131 LINK PMAINT CPFMTXA 141 LINK $PAGE$ CPFMTXA A01 CPFMTXA 141 CPFMTXA 131 CPFMTXA 123 CPFMTXA 122 CPFMTXA A01 SSI6RS LABEL SSI6S1 LABEL SSI6R1 LABEL 141 141 MR SSI6C1 LABEL A01 A01 MR SSI6P1 LABEL SSI6C1 OWNER SSI6R1 OWNER SSI6RS OWNER SSI6S1 OWNER SSI6P1 OWNER
68
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
5. We executed the following steps to synchronize the online directory with the source copy that is held by DIRMAINT: a. Link DIRMAINT 1DF as MR. b. Access 1DF G. c. Delete the file USER DIRECT G. d. Copy the amended USER WITHPASS file to USER INPUT G. 6. We restarted DIRMAINT and issued the command DIRM USER WITHPASS to ensure that Directory Maintenance had the correct version of the directory.
Setting up TCP/IP
We performed the following installation steps: 1. We logged on to TCPMAINT. 2. We changed the PROFILE TCPIP and SYSTEM DTCPARMS on the TCPMAINT 198 disk to reflect the OSA address on ITSOSSI6. 3. We changed the TCPIP DATA on the TCPMAINT 592 disk to reflect the host name of ITSOSSI6.
Setting up RSCS
We performed the following installation steps: 1. We logged on to MAINT620, accessed 190 as G, and amended the SYSTEM NETID file to reflect the CPU ID and host IP address. 2. We logged on to 6VMRSC20 and amended the RSCS definitions for the links to the other z/VM systems. 3. We logged on to MAINT620 and built the CMS and GCS saved segments to include the amended SYSTEM NETID file by issuing the following commands: SERVICE CMS BLDNUC SERVICE GCS BLDNUC PUT2PROD
Starting ITSOSSI6
We shut down ITSOSSI6 as a second-level z/VM guest. We performed an IPL on ITSOSSI6 on the z196.
Chapter 5. Scenario two: Converting two existing stand-alone z/VM systems to a single system image cluster
69
2. We moved the RACF databases from minidisk to full-pack while keeping the current database allocation. See the RACF Database on Full-pack Minidisk section in the z/VM V6R2.0 RACF Security Server System Programmers Guide, SC24-6219. 3. We shared the RACF databases among the members. See the Sharing RACF Databases in a z/VM Single System Image Cluster section in the z/VM V6R2.0 RACF Security Server System Programmers Guide, SC24-6219. 4. We then restarted RACF to verify that these databases were usable.
cpfmtxa 9e2e ssi6s1 CPFMTXA: FORMAT WILL ERASE CYLINDERS 00000-nnnnn ON DISK 9E2E DO YOU WANT TO CONTINUE? (YES | NO) 70
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
yes HCPCCF6209I INVOKING ICKDSF. ... ICK00002I ICKDSF PROCESSING COMPLETE. MAXIMUM CONDITION CODE WAS 0 ENTER ALLOCATION DATA TYPE CYLINDERS spol 1-end end HCPCCF6209I INVOKING ICKDSF. ... ICK00002I ICKDSF PROCESSING COMPLETE. MAXIMUM CONDITION CODE WAS 0 cpfmtxa 9e2e ssi6s1 owner itsossib.itsossi6
cpfmtxa 9a27 ssi6rs 21.9 INVOKING ICKDSF. FORMAT WILL ERASE CYLINDERS 21-29 ON DISK vdev DO YOU WANT TO CONTINUE? (YES | NO) yes HCPCCF6209I INVOKING ICKDSF. ... ICK00002I ICKDSF PROCESSING COMPLETE. MAXIMUM CONDITION CODE WAS 0 ENTER ALLOCATION DATA TYPE CYLINDERS end HCPCCF6209I INVOKING ICKDSF. ... ICK00002I ICKDSF PROCESSING COMPLETE. MAXIMUM CONDITION CODE WAS 0 cpfmtxa 9a27 ssi6rs 30.9 INVOKING ICKDSF. FORMAT WILL ERASE CYLINDERS 30-38 ON DISK vdev DO YOU WANT TO CONTINUE? (YES | NO)
Chapter 5. Scenario two: Converting two existing stand-alone z/VM systems to a single system image cluster
71
yes HCPCCF6209I INVOKING ICKDSF. ... ICK00002I ICKDSF PROCESSING COMPLETE. MAXIMUM CONDITION CODE WAS 0 ENTER ALLOCATION DATA TYPE CYLINDERS end HCPCCF6209I INVOKING ICKDSF. ... ICK00002I ICKDSF PROCESSING COMPLETE. MAXIMUM CONDITION CODE WAS 0
USER $ALLOC$ NOLOG MDISK 0A00 3390 000 001 SSIBC1 R MDISK 0A01 3390 000 001 SSIBR1 R MDISK 0A02 3390 000 001 SSI5RS R MDISK 0A03 3390 000 001 SSI6RS R * USER $DIRECT$ NOLOG MDISK 0A01 3390 001 020 SSI5RS R MDISK 0A02 3390 001 020 SSI6RS R MDISK 0B01 3390 131 100 SSIBC1 R * USER $SYSCKP$ NOLOG MDISK 0A01 3390 021 009 SSI5RS R MDISK 0A02 3390 021 009 SSI6RS R * USER $SYSWRM$ NOLOG MDISK 0A01 3390 030 009 SSI5RS R MDISK 0A02 3390 030 009 SSI6RS R * USER $PAGE$ NOLOG MDISK A01 3390 000 END SSI5P1 R MDISK A02 3390 000 END SSI6P1 R *
72
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
USER $SPOOL$ NOLOG MDISK 0A01 3390 000 END SSI5S1 R MDISK 0A02 3390 000 END SSI6S1 R * USER $TDISK$ NOLOG * 2. We updated the directory to add the sysres for ITSOSSI6 to identify where to locate the object directory for the system: dirmaint directory change 1 ssi 123 3390 ssr5rs ssi6rs 3. We added the new, required DATAMOV2 and DIRMSAT2 entries. The supplied user directory includes commented-out entries for DATAMOV2 and DIRMSAT2, which can be used as templates for the new user IDs. Check that these entries include the minidisk definitions that are similar to the minidisk entries in DIRMSAT and DATAMOVE. 4. We also created DVHPROFA DIRMSAT2 on the DIRMAINT 191 disk by copying the existing DVHPROFA DIRMSAT file. We used the following DIRMAINT command: DIRM FILE DVHPROFA DIRMSAT2 A DVHPROFA DIRMSAT2 C 5. We updated the multiconfiguration virtual machine definitions to add the subconfiguration entries for ITSOSSI6. The DIRM SCAN BUILD ON ITSOSSI5 provides you with a list of all the subconfiguration entries that are defined for ITSOSSI5 (Example 5-4). Use this list as a reference for the new entries that are required for ITSOSSI6.
Example 5-4 Sample output from the DIRM SCAN BUILD ON ITSOSSI5 command
Userid: ======================= Qualifying Record ===================== MAINT BUILD ON ITSOSSI5 USING SUBCONFIG MAINT-1 AVSVM BUILD ON ITSOSSI5 USING SUBCONFIG AVSVM-1 TSAFVM BUILD ON ITSOSSI5 USING SUBCONFIG TSAFVM-1 GCS BUILD ON ITSOSSI5 USING SUBCONFIG GCS-1 AUDITOR BUILD ON ITSOSSI5 USING SUBCONFIG AUDITR-1 AUTOLOG1 BUILD ON ITSOSSI5 USING SUBCONFIG AUTLG1-1 CMSBATCH BUILD ON ITSOSSI5 USING SUBCONFIG CMSBAT-1 DISKACNT BUILD ON ITSOSSI5 USING SUBCONFIG DSKACT-1 . . . VSMEVSRV BUILD ON ITSOSSI5 USING SUBCONFIG VSMEVS-1 DTCSMAPI BUILD ON ITSOSSI5 USING SUBCONFIG DTCSMA-1 LOHCOST BUILD ON ITSOSSI5 USING SUBCONFIG LOHCOS-1 ZVMLXTS BUILD ON ITSOSSI5 USING SUBCONFIG ZVMLXT-1 Scan Complete. 6. For each user ID that is listed in the scan output, we entered a command that is similar to this command: dirmaint add maint-2 like maint-1 build on itsossi6 in maint Although it is possible to set up a REXX program to perform this function, we do not suggest that you process this task as a REXX exec. We learned that DIRMAINT was unable to handle our multiple requests, and the results can be unpredictable. One suggested way to process multiple DIRMAINT commands is to use the DIRM BATCH command. Create a file with the correct DIRMAINT commands and execute them with DIRM BATCH. See Example 5-5 on page 74 and Example 5-6 on page 74.
Chapter 5. Scenario two: Converting two existing stand-alone z/VM systems to a single system image cluster
73
Example 5-5 DIRM CMDS A batch file containing the DIRM statements
add maint-2 like maint-1 build on itsossi6 in maint add avsvm-2 like avsvm-1 build on itsossi6 in avsvm ...
Example 5-6 Executing the batch file
dirm batch DIRM CMDS A 7. We obtained a copy of the source directory by issuing DIRM USER WITHPASS, and we copied this directory to the PMAINT 2CC disk as USER DIRECT. 8. Next, we created an object directory on ITSOSSI5 by issuing DIRM DIRECT. 9. We then created the object directory for ITSOSSI6, as shown in Example 5-7. We detached our link to the ITSOSSI5 sysres (device 123), attached SSI6RS to our user ID as disk 123, created the object directory on SSI6RS, detached SSI6RS, and reattached the ITSOSSI5 sysres.
Example 5-7 Creating new object directory for ITSOSSI6
detach 123 attach 9a27 to * as 123 directxa user direct c detach 123 link maint 123 123 m 10.The last step in the DIRMAINT configuration was to update and reload the CONFIGSS DATADVH file with the new satellite and to datamove the user IDs for ITSOSSI6. Example 5-8 shows the contents of our CONFIGSS DATADVH file.
Example 5-8 Our CONFIGSS DATADVH
OPERATIONAL IMMED DIRMSAT ITSOSSI5 DIRMSAT2 ITSOSSI6 DATAMOVE ITSOSSI5 * DATAMOV2 ITSOSSI6 *
/**********************************************************************/ /* System_Identifier Information */ /**********************************************************************/ System_Identifier LPAR A25 ITSOSSI5 System_Identifier LPAR A19 ITSOSSI6 /**********************************************************************/ /* SSI Cluster */ /**********************************************************************/ SSI ITSOSSIB PDR_VOLUME SSIBC1, SLOT 1 ITSOSSI5,
74
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
SLOT 2 ITSOSSI6 /**********************************************************************/ /* Checkpoint and Warmstart Information */ /**********************************************************************/ ITSOSSI5: System_Residence, Checkpoint Volid SSI5RS From CYL 21 For 9 , Warmstart Volid SSI5RS From CYL 30 For 9 ITSOSSI6: System_Residence, Checkpoint Volid SSI6RS From CYL 21 For 9 , Warmstart Volid SSI6RS From CYL 30 For 9 /**********************************************************************/ /* CP_Owned Volume Statements */ /**********************************************************************/ ITSOSSI5: CP_Owned Slot 1 SSI5RS ITSOSSI6: CP_Owned Slot 1 SSI6RS /**********************************************************************/ CP_Owned Slot 5 SSIBC1 /**********************************************************************/ CP_Owned Slot 10 SSI5S1 1 CP_Owned Slot 11 SSI6S1 /**********************************************************************/ ITSOSSI5: CP_Owned Slot 255 SSI5P1 ITSOSSI6: CP_Owned Slot 255 SSI6P1 /**********************************************************************/ /* Shared User Volumes */ /**********************************************************************/ User_Volume_List SSIBR1 User_Volume_List SSI5M1 SSI6M1 LX9A28 LX9A29 /**********************************************************************/ Devices , Online_at_IPL 0000-FFFF, Offline_at_IPL 9C25-9C27 9D26 9D27, 2 Sensed 0000-FFFF /**********************************************************************/ /* ISCF Links */ /**********************************************************************/ ITSOSSI5: ACTIVATE ISLINK 4290 5290 NODE ITSOSSI6 ITSOSSI6: ACTIVATE ISLINK 5050 4050 NODE ITSOSSI5 /**********************************************************************/ /* NETWORK DEFINITIONS */ /**********************************************************************/ ITSOSSI5: DEFINE VSWITCH VSWITCH1 RDEV 3083 IP ITSOSSI6: DEFINE VSWITCH VSWITCH1 RDEV 2103 IP The following numbers refer to the numbers that are shown in Example 5-9 on page 74: 1 The spool volumes SSI5S1 and SSI6S1 do not have their member names as a qualifier, because they are shared volumes. 2 We added devices 9C25-9C27, 9D26, and 9D27 to Offline_at_IPL on the Devices statement to prevent problems with duplicate volsers at IPL time. These devices are the system volumes of the stand-alone ITSOSSI6 non-SSI system.
Chapter 5. Scenario two: Converting two existing stand-alone z/VM systems to a single system image cluster
75
To ensure that no syntax errors occurred, we used the CPSYNTAX EXEC, which is on the MAINT 193 disk, to check the SYSTEM CONFIG file.
Starting ITSOSSI6
We started ITSOSSI6 and specified CLEAN NOAUTOLOG. Then, we executed XAUTOLOG RACFVM.
25DE50 ITSOSSI5 RSCS 193BD5 ITSOSSI6 RSCS 2. We ran XAUTOLOG on the file pool servers, VMSERVS, VMSERVU, VMSERVR, and VMSERVP, and built the following CMS named saved system and the other saved segments and named saved systems: put2prod savecms put2prod segments all service gcs bldnuc put2prod
76
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
USER LNXRH56 HCHT57UI 512M 1G G INCLUDE IBMDFLT IPL CMS PARM AUTOCR IUCV ALLOW IUCV ANY MACH ESA 4 OPTION CHPIDV ONE POSIXINFO UID 59 NICDEF C200 TYPE QDIO LAN SYSTEM VSWITCH1 MDISK 0201 3390 1 1000 LX9A29 MR MDISK 0202 3390 1001 9016 LX9A29 MR MDISK 0191 3390 1 50 SSI6M1
Example 5-12 Directory entry for 610ZVM
USER 610ZVM 8D99H950 99M 999M BG INCLUDE IBMDFLT ACCOUNT ITS30000 CPU 0 BASE CPUID 616161 IPL CMS PARM AUTOCR MACH XA OPTION MAINTCCW DEVMAINT CHPIDV ONE DEDICATE D850 D850 DEDICATE D851 D851 DEDICATE D852 D852 DEDICATE D853 D853 DEDICATE D854 D854 NICDEF 3080 TYPE QDIO LAN SYSTEM VSWITCH1 SPECIAL 08E1 3270 SPECIAL 08E2 3270 SPECIAL 08E3 3270 SPECIAL 08E4 3270 SPECIAL 08E5 3270 SPECIAL 08E6 3270 MDISK 0191 3390 51 10 SSI6M1
5.6 Suggestions
In our scenario, cloning the SSI member ITSOSSI5 (see 5.5, Creating an ITSOSSI6 system by cloning ITSOSSI5 on page 70) was complicated. If possible, we suggest that you build an SSI cluster, as described in Chapter 4, Scenario one: Creating a single system image cluster with four new z/VM members on page 35, and then move your guests into the new SSI cluster.
Chapter 5. Scenario two: Converting two existing stand-alone z/VM systems to a single system image cluster
77
/**/ Address CMS Do forever 'Q T' 'PING 9.12.4.236' 'CP SLEEP 10 SEC' End
Example 5-14 VMRELOCATE command for 610ZVM guest
at itsossi5 cmd vmrelocate move 610zvm itsossi6 16:17:07 Relocation of 610ZVM from ITSOSSI5 to ITSOSSI6 started by MAINT620 16:17:07 Relocation of 610ZVM from ITSOSSI5 to ITSOSSI6 started 16:17:08 DASD D850 ATTACHED TO 610ZVM D850 BY 610ZVM WITH DEVCTL HYPERPAV BASE 16:17:08 DASD D853 ATTACHED TO 610ZVM D853 BY 610ZVM WITH DEVCTL HYPERPAV BASE 16:17:08 DASD D851 ATTACHED TO 610ZVM D851 BY 610ZVM WITH DEVCTL HYPERPAV BASE 16:17:08 DASD D854 ATTACHED TO 610ZVM D854 BY 610ZVM WITH DEVCTL HYPERPAV BASE 16:17:08 DASD D852 ATTACHED TO 610ZVM D852 BY 610ZVM WITH DEVCTL HYPERPAV BASE 610ZVM : User 610ZVM has been relocated from ITSOSSI5 to ITSOSSI6 16:17:08 User 610ZVM has been relocated from ITSOSSI5 to ITSOSSI6 16:17:08 User 610ZVM has been relocated from ITSOSSI5 to ITSOSSI6 Ready; T=0.01/0.01 16:17:08 Example 5-15 on page 78 shows the output from the ping requests that were issued during the relocation request for 610ZVM.
Example 5-15 Response to PING command on 610ZVM during relocation
Ready; T=0.03/0.04 16:17:02 test 16:17:04 TIME IS 16:17:04 EST FRIDAY 11/11/11 16:17:04 CONNECT= 00:18:44 VIRTCPU= 000:00.59 TOTCPU= 000:00.79 Ping Level 610: Pinging host 9.12.4.236. Enter #CP EXT to interrupt. PING: Ping #1 response took 0.001 seconds. Successes so far 1. 16:17:08 HCPIPQ2091I I/O priority setting available to CP has changed. 78
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
16:17:08 HCPIPQ2091I New setting is enabled with range 000 to 000. 16:17:14 TIME IS 16:17:14 EST FRIDAY 11/11/11 16:17:14 CONNECT= 00:18:54 VIRTCPU= 000:00.60 TOTCPU= 000:00.80 Ping Level 610: Pinging host 9.12.4.236. Enter #CP EXT to interrupt. PING: Ping #1 response took 0.001 seconds. Successes so far 1. 16:17:24 TIME IS 16:17:24 EST FRIDAY 11/11/11 16:17:24 CONNECT= 00:19:03 VIRTCPU= 000:00.60 TOTCPU= 000:00.80 Ping Level 610: Pinging host 9.12.4.236. Enter #CP EXT to interrupt. PING: Ping #1 response took 0.001 seconds. Successes so far 1. 16:17:34 TIME IS 16:17:34 EST FRIDAY 11/11/11 16:17:34 CONNECT= 00:19:14 VIRTCPU= 000:00.61 TOTCPU= 000:00.81 Ping Level 610: Pinging host 9.12.4.236. Enter #CP EXT to interrupt. PING: Ping #1 response took 0.001 seconds. Successes so far 1.
Chapter 5. Scenario two: Converting two existing stand-alone z/VM systems to a single system image cluster
79
SAP GUI
itsossi5
itsossi6
p5503p2
wtsc59
Figure 5-4 SAP solution on a System z with the relocation feature of z/VM 6.2
SAP is often used in a System z setup. An important requirement for SAP clients is 24x7 availability to help ensure business continuity, avoid business interruptions, and reduce lost revenue. With LGR, z/VM 6.2 provides an additional feature for 24x7 availability of SAP systems. Now, you can relocate an SAP Application Server on a Linux for System z system non-disruptively. Use LGR for applying service to the z/VM system, for workload-balancing activities, or for migrations to new System z hardware. In our example, we relocate the SAP AppServer (CI) non-disruptively between the z/VM systems ITSOSSI5 and ITSOSSI6. Example 5-16 shows the SAP AppServer (CI) guest definition in the z/VM directory.
Example 5-16 SAP guest definition
USER LNXSU11 XXXXXXXX 6G CPU 01 CPU 00 CPU 02 CPU 03 IPL 202 MACHINE ESA 4 OPTION CHPIDV ONE NICDEF C200 TYPE QDIO NICDEF 6300 TYPE QDIO NICDEF 6303 TYPE QDIO 80
16G G
LAN SYSTEM VSWITCH1 LAN SYSTEM VSW999 DEVICES 3 LAN SYSTEM VSW199 DEVICES 3
1 2
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
NICDEF 7000 TYPE QDIO LAN SYSTEM VSW199 DEVICES 3 SPOOL 000C 3505 A SPOOL 000D 3525 A SPOOL 000E 1403 A CONSOLE 0009 3215 T MDISK 0201 3390 1 10016 LX9980 MR MDISK 0202 3390 1 10016 LX9981 MR MDISK 0301 3390 1 10016 LX9982 MR MDISK 0302 3390 1 10016 LX9983 MR MDISK 0303 3390 1 10016 LX9984 MR MDISK 0304 3390 1 10016 LX9985 MR MDISK 0305 3390 1 10016 LX9986 MR MDISK 0306 3390 1 10016 LX9987 MR MDISK 0307 3390 1 10016 LX9988 MR MDISK 0308 3390 1 10016 LX9989 MR MDISK 0309 3390 1 10016 LX998A MR
3 4 5 6 7
The following numbers refer to the numbers that are shown in Example 5-16 on page 80: 1 sets CHPID virtualization to ONE 2 vSwitch definitions 3 swap space 4 /root/ filesystem 5 /usr/sap/ filesystem 6 /sapmnt/ filesystem 7 /install/ filesystem
Relocation process
Follow these steps: 1. Issue the VMRELOCATE TEST command to check the z/VM guest eligibility.
Example 5-17 VMRELOCATE TEST command
vmrelocate test lnxsu11 itsossi6 User LNXSU11 is eligible for relocation to ITSOSSI6 Ready; T=0.01/0.01 16:17:56 2. Issue the VMRELOCATE MOVE command.
Example 5-18 VMRELOCATE MOVE command
vmrelocate move lnxsu11 itsossi6 Relocation of LNXSU11 from ITSOSSI5 to ITSOSSI6 started User LNXSU11 has been relocated from ITSOSSI5 to ITSOSSI6 Ready; T=0.01/0.01 16:19:12
Monitoring results
We monitored the relocation process for two scenarios.
Scenario one
We monitored for a relocation of the SAP Application Server (CI) without any load, which means without any SAP activities. For the monitoring, we issued the VMRELOCATE STATUS each 0.3 seconds to see the states of the relocation process. Example 5-19 on page 82 and
Chapter 5. Scenario two: Converting two existing stand-alone z/VM systems to a single system image cluster
81
Example 5-20 on page 82 show the results of monitoring the relocation process without SAP activities.
Example 5-19 Performance data for SAP AppServer (CI) without load
top - 16:30:53 up 12:45, 2:45, 1 user, load average: 0.01, 0.00, 0.00 Tasks: 101 total, 1 running, 100 sleeping, 0 stopped, 0 zombie Cpu(s): 0.1%us, 0.0%sy, 0.0%ni, 99.9%id, 0.0%wa, 0.0%hi, 0.0%si, 0.0%st Mem: 6171128k total, 3947952k used, 2223176k free, 118032k buffers Swap: 7211416k total, 0k used, 7211416k free, 3500412k cached
Example 5-20 Relocation time of SAP AppServer (CI) without load
16:33:49 User LNXSU11 . . . 16:33:56 User LNXSU11 16:33:56 User LNXSU11 16:33:56 User LNXSU11 16:33:58 User LNXSU11 16:33:58 User LNXSU11 16:33:58
Elapsed 00:00:00
From To By ITSOSSI5 ITSOSSI6 ITSO From To By ITSOSSI5 ITSOSSI6 ITSO From To By ITSOSSI5 ITSOSSI6 ITSO From To By ITSOSSI5 ITSOSSI6 ITSO From To By ITSOSSI5 ITSOSSI6 ITSO
Status Moving Memory Status Moving Guest Status Final Mem Copy Status Final Mem Copy Status Cleaning Up
Elapsed 00:00:07 Elapsed 00:00:07 Elapsed 00:00:08 Elapsed 00:00:09 Elapsed 00:00:09
This scenario provided the following results: The elapsed time for the relocation process was 9 seconds. The quiesce time for the z/VM guest was about 2 seconds. See the status Final Mem Copy.
Scenario two
The second relocation of the SAP Application Server (CI) included active SAP applications. Example 5-21 and Example 5-22 on page 83 show the results of monitoring the relocation process with a full SAP load.
Example 5-21 Performance data for SAP AppServer (CI) with load
top - 09:50:07 up 23:19, 2 users, load average: 2.65, 1.84, 1.41 Tasks: 131 total, 3 running, 128 sleeping, 0 stopped, 0 zombie Cpu(s): 30.8%us, 3.2%sy, 0.0%ni, 62.2%id, 0.0%wa, 0.2%hi, 0.7%si, 2.8%st Mem: 6171128k total, 5424008k used, 747120k free, 108564k buffers Swap: 7211416k total, 0k used, 7211416k free, 4386852k cached
82
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
Example 5-22 Relocation of SAP AppServer (CI) with full SAP load
09:51:14 User LNXSU11 . . . 09:51:51 User LNXSU11 09:51:51 User LNXSU11 09:51:52 User LNXSU11 09:51:52 User LNXSU11 09:51:53 User LNXSU11 09:51:54 User LNXSU11 09:51:54 User LNXSU11 09:51:54
Elapsed 00:00:00
From To By ITSOSSI5 ITSOSSI6 ITSO From To By ITSOSSI5 ITSOSSI6 ITSO From To By ITSOSSI5 ITSOSSI6 ITSO From To By ITSOSSI5 ITSOSSI6 ITSO From To By ITSOSSI5 ITSOSSI6 ITSO From To By ITSOSSI5 ITSOSSI6 ITSO From To By ITSOSSI5 ITSOSSI6 ITSO
Status Moving Memory Status Moving Guest Status Moving Guest Status Final Mem Copy Status Final Mem Copy Status Cleaning Up Status Cleaning Up
Elapsed 00:00:37 Elapsed 00:00:37 Elapsed 00:00:38 Elapsed 00:00:38 Elapsed 00:00:39 Elapsed 00:00:40 Elapsed 00:00:40
This scenario provided the following results: The elapsed time for the relocation process was 40 seconds. The quiesce time for the z/VM guest was about 2 seconds. See the status Final Mem Copy. Figure 5-5 on page 84 and Figure 5-6 on page 85 show the results of the relocation, visible in the SAP Operation System Monitor (SAP transaction ST06).
Chapter 5. Scenario two: Converting two existing stand-alone z/VM systems to a single system image cluster
83
84
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
The host systems are shown in the field Virtualization Configuration - Host System Information.
Chapter 5. Scenario two: Converting two existing stand-alone z/VM systems to a single system image cluster
85
86
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
Appendix A.
87
USER ITSOLNX2 ITSOSSI 4G 4G G * INCLUDE LINCMS INCLUDE LINDFLT IPL 300 MACH ESA 4 NICDEF 060 Example A-3 0 TYPE QDIO LAN SYSTEM ITSOVSW1 MDISK 0300 3390 DEVNO 9A26 MR
Problem
Example A-2 shows that, in this case, the full-pack DASD was not available at the destination member.
Example A-2 VMRELOCATE TEST command
vmrelocate test user itsolnx2 to itsossi1 HCPRLH1940E ITSOLNX2 is not relocatable for the following reason(s): HCPRLI1981I ITSOLNX2: Source system DEVNO defined minidisk 0300 on real device 9 A26 cannot be created on the destination system because the equivalent real devi ce 9A26 is neither a DENVO defined minidisk nor free Ready(01940); T=0.01/0.01 15:15:09
Solution
Detaching this address from the system at the destination solved the problem. We used the command at ... cmd without needing to log on at the other member, as shown in Example A-3.
Example A-3 Use of the at command
q 9a26 DASD 9A26 CP SYSTEM DEVNO Ready; T=0.01/0.01 15:15:52 at itsossi1 cmd q 9a26
88
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
DASD 9A26 CP SYSTEM US9A26 0 Ready; T=0.01/0.01 15:16:13 at itsossi1 cmd det 9a26 system DASD 9A26 DETACHED SYSTEM Ready; T=0.01/0.01 15:16:46 at itsossi1 cmd q 9a26 DASD 9A26 US9A26 Ready; T=0.01/0.01 15:16:50 vmrelocate test user itsolnx2 to itsossi1 User ITSOLNX2 is eligible for relocation to ITSOSSI1 Ready; T=0.01/0.01 15:17:04
USER ITSOLNX2 ITSOSSI 4G 4G G INCLUDE LINCMS * INCLUDE LINDFLT IPL 300 MACH ESA 4 OPTION CHPIDV ONE NICDEF 0600 TYPE QDIO LAN SYSTEM ITSOVSW1 MDISK 0300 3390 DEVNO 9A26 MR
Problem
These CMS mdisks are defined on local non-shared volumes, which causes an error.
Example A-5 Linux with CMS mdisks
vmrelocate test user itsolnx2 to itsossi1 HCPRLH1940E ITSOLNX2 is not relocatable for the following reason(s): HCPRLI1996I ITSOLNX2: Virtual machine device 0190 is a link to a local minidisk HCPRLI1996I ITSOLNX2: Virtual machine device 019D is a link to a local minidisk HCPRLI1996I ITSOLNX2: Virtual machine device 019E is a link to a local minidisk Ready(01940); T=0.01/0.01 15:38:45
Solution
The CMS minidisks are necessary only during the installation of Linux. After the installation, the CMS minidisks are not necessary. To solve the problem, we detached these CMS minidisks from the Linux machine dynamically. Removing the INCLUDE cards makes the definitions permanent. After detaching the CMS minidisks, the VMRELOCATE TEST did not show any errors (Example A-6 on page 90). The VMRELOCATE MOVE transferred the machine to the ITSOSSI1 member.
89
send cp itsolnx2 det 190 19d 19e Ready; T=0.01/0.01 15:48:18 ITSOLNX2: 0190 019D 019E DETACHED vmrelocate test user itsolnx2 to itsossi1 User ITSOLNX2 is eligible for relocation to ITSOSSI1 Ready; T=0.01/0.01 15:48:32 vmrelocate move user itsolnx2 to itsossi1 Relocation of ITSOLNX2 from ITSOSSI4 to ITSOSSI1 started User ITSOLNX2 has been relocated from ITSOSSI4 to ITSOSSI1 ITSOLNX2: User ITSOLNX2 has been relocated from ITSOSSI4 to ITSOSSI1 ITSOLNX2: User ITSOLNX2 has been relocated from ITSOSSI4 to ITSOSSI1 ITSOLNX2: CONNECT= 00:10:37 VIRTCPU= 000:04.92 TOTCPU= 000:05.05 ITSOLNX2: LOGOFF AT 15:48:48 EST MONDAY 11/07/11 BY SYSTEM USER DSC LOGOFF AS ITSOLNX2 USERS = 19 FORCED BY SYSTEM Ready; T=0.01/0.01 15:48:48 q user itsolnx2 at all ITSOSSI1 : ITSOLNX2 - DSC Ready; T=0.01/0.01 15:49:30 close cons LOGOFF notices: We moved Linux and all its applications to ITSOSSI1. Example A-6 shows the LOGOFF notices on ITSOSSI4, because the guest is no longer running on the source member. The control program (CP) cleans up the data structures in the same way as though you logged off the guest.
Problem
Example A-7 shows a V-DISK that is defined as a minidisk to prevent the relocation.
Example A-7 V-DISk minidisk statement
type itsolnx1 direct USER ITSOLNX1 ITSOSSI 4G 4G G INCLUDE LINDFLT IPL 300 MACH ESA 4 NICDEF 0600 TYPE QDIO LAN SYSTEM ITSOVSW1 MDISK 0111 FB-512 V-DISK 128000 MDISK 0300 3390 0001 10016 LXDE1C MR MDISK 0400 3390 0001 30050 LX603D MR Ready; T=0.01/0.01 11:34:11 vmrelocate test itsolnx1 to itsossi3 HCPRLH1940E ITSOLNX1 is not relocatable for the following reason(s): HCPRLI1970I ITSOLNX1: Virtual machine device 0111 is a shareable VDISK Ready(01940); T=0.01/0.01 11:34:17
90
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
Solution
If the V-DISK is used for SWAP only and is not shared, no restrictions exist to use a V-DISK that is defined by the DEFINE command. So, changing from MDISK V-DISK to COMMAND DEFINE VFB-512 allows the relocation. Example A-8 shows the updated directory. Important: To use DASDs that are emulated in memory effectively, they must be formatted and defined as SWAP every time that Linux is started.
Example A-8 Updated directory statements
type
itsolnx1 direct
USER ITSOLNX1 ITSOSSI 4G 4G G INCLUDE LINDFLT COMMAND DEFINE VFB-512 AS 0111 BLK 128000 IPL 300 MACH ESA 4 NICDEF 0600 TYPE QDIO LAN SYSTEM ITSOVSW1 * MDISK 0111 FB-512 V-DISK 128000 MDISK 0300 3390 0001 10016 LXDE1C MR MDISK 0400 3390 0001 30050 LX603D MR Ready; T=0.01/0.01 11:53:50 vmrelocate test itsolnx1 to itsossi3 User ITSOLNX1 is eligible for relocation to ITSOSSI3 Ready; T=0.01/0.01 11:53:58
Problem
Example A-9 shows many restrictions, which cannot be solved by commands, on this guest. All minidisks are local and cannot be detached.
Example A-9 A CMS guest
USER VMLOGS ITSOSSI 64M 128M ABDEG * * Keep CONSOLE/SPOOL files on DASD for some time (90 days) * INCLUDE IBMDFLT IPL CMS PARM AUTOCR MACH ESA MDISK 0191 3390 2902 1000 SSIAC1 MR LABEL CMS191 MDISK 0193 3390 3902 0005 SSIAC1 MR LABEL CMS193
Solution
We cannot use LGR to relocate a CMS guest, but we still can use VMSSI functions (see Example A-10 on page 92). The objective of this machine is to collect and save spool files. We can stop it in one member (FORCE at ITSOSSI4) and start it again on another member
Appendix A. Frequently asked questions
91
(XAUTOLOG at ITSOSSI1). The cross-system spool functions help you avoid losing a file. See the commands in Example A-10. Linux guests: Use this same approach with Linux guests that cannot use the relocate function due to size conflicts, application characteristics, and so on. If it is a Linux guest, you must use the SIGNAL SHUTDOWN command instead of the FORCE command.
Example A-10 Manual relocation of a CMS guest
vmrelocate test user vmlogs to itsossi2 HCPRLH1940E VMLOGS is not relocatable for the following reason(s): HCPRLE1956I VMLOGS: Single path CHPID virtualization is not enabled HCPRLE1963I VMLOGS: Virtual machine is using VMCF HCPRLI1996I VMLOGS: Virtual machine device 0190 is a link to a local minidisk HCPRLI1996I VMLOGS: Virtual machine device 019D is a link to a local minidisk HCPRLI1996I VMLOGS: Virtual machine device 019E is a link to a local minidisk HCPRLI1996I VMLOGS: Virtual machine device 0401 is a link to a local minidisk HCPRLI1996I VMLOGS: Virtual machine device 0402 is a link to a local minidisk HCPRLI1996I VMLOGS: Virtual machine device 1193 is a link to a local minidisk HCPRLL1980I VMLOGS: An identical NSS or DCSS CMS does not exist on the destinati on system at itsossi4 cmd force vmlogs VMLOGS : CONNECT= 99:59:59 VIRTCPU= 000:00.85 TOTCPU= 000:02.85 VMLOGS : LOGOFF AT 16:24:44 EST MONDAY 11/07/11 BY MAINT620 USER DSC LOGOFF AS VMLOGS USERS = 18 FORCED BY MAINT620 Ready; T=0.01/0.01 16:24:44 VMLOGS : RDR FILE 0368 SENT FROM VMLOGS CON WAS 0368 RECS 0439 CPY 001 T NOH OLD NOKEEP at itsossi1 cmd xautolog vmlogs Command accepted Ready; T=0.01/0.01 16:25:00 query user vmlogs at all ITSOSSI1 : VMLOGS - DSC Ready; T=0.01/0.01 16:25:07 close cons CON FILE 0216 SENT TO VMLOGS RDR AS 0408 RECS 0150 CPY 001 T NOHOLD NOKEEP Ready; T=0.01/0.01 16:27:45 16:27:45 * MSG FROM VMLOGS : File MAINT CONSOLE received with success.
92
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
To change the characteristics of any real device, it must be offline. Example A-11 shows the commands that we used to delete the EQID.
Example A-11 Deleting EQIDs
q osa OSA 2120 ATTACHED TO TCPIP 2120 DEVTYPE OSA CHPID 0C OSD OSA 2121 ATTACHED TO TCPIP 2121 DEVTYPE OSA CHPID 0C OSD OSA 2122 ATTACHED TO TCPIP 2122 DEVTYPE OSA CHPID 0C OSD OSA 2123 ATTACHED TO DTCVSW1 0600 DEVTYPE OSA CHPID 0C OSD OSA 2124 ATTACHED TO DTCVSW1 0601 DEVTYPE OSA CHPID 0C OSD OSA 2125 ATTACHED TO DTCVSW1 0602 DEVTYPE OSA CHPID 0C OSD Ready; T=0.01/0.01 11:08:11 q 2126-2128 OSA 2126 FREE , OSA 2127 FREE , OSA 2128 FREE Ready; T=0.01/0.01 11:08:29 vary off 2126-2128 2126 varied offline 2127 varied offline 2128 varied offline 3 device(s) specified; 3 device(s) successfully varied offline Ready; T=0.01/0.01 11:08:45 set rdev 2126-2128 noeqid type osa HCPZRP6722I Characteristics of device 2126 were set as requested. HCPZRP6722I Characteristics of device 2127 were set as requested. HCPZRP6722I Characteristics of device 2128 were set as requested. 3 RDEV(s) specified; 3 RDEV(s) changed; 0 RDEV(s) created Ready; T=0.01/0.01 11:09:20 vary on 2126-2128 2126 varied online 2127 varied online 2128 varied online 3 device(s) specified; 3 device(s) successfully varied online Ready; T=0.01/0.01 11:09:31
Problem
Example A-12 shows the amended directory entry for the Linux guest with the NICDEF commented out and the respective DEDICATE commands added.
Example A-12 Amended Linux directory entry
type itsolnx2 direct USER ITSOLNX2 ITSOSSI 4G 4G G * INCLUDE LINCMS INCLUDE LINDFLT COMMAND DEFINE VFB-512 AS 0111 BLK 128000 IPL 300 MACH ESA 4 DEDICATE 0600 2126 DEDICATE 0601 2127 DEDICATE 0602 2128 * NICDEF 0600 TYPE QDIO LAN SYSTEM ITSOVSW1 11071537 11071537 11071537 11071537 11071537
11071537
Appendix A. Frequently asked questions
93
11071537
Example A-13 shows that the Linux guest was started, that the OSA was checked, and that we tried the relocation.
Example A-13 Dedicate devices without EQID
xautolog itsolnx2 Command accepted Ready; T=0.01/0.01 11:14:19 AUTO LOGON *** ITSOLNX2 USERS = 19 HCPCLS6056I XAUTOLOG information for ITSOLNX2: The IPL command is verified by the IPL command processor. q osa OSA 2120 ATTACHED TO TCPIP 2120 DEVTYPE OSA CHPID 0C OSD OSA 2121 ATTACHED TO TCPIP 2121 DEVTYPE OSA CHPID 0C OSD OSA 2122 ATTACHED TO TCPIP 2122 DEVTYPE OSA CHPID 0C OSD OSA 2123 ATTACHED TO DTCVSW1 0600 DEVTYPE OSA CHPID 0C OSD OSA 2124 ATTACHED TO DTCVSW1 0601 DEVTYPE OSA CHPID 0C OSD OSA 2125 ATTACHED TO DTCVSW1 0602 DEVTYPE OSA CHPID 0C OSD OSA 2126 ATTACHED TO ITSOLNX2 0600 DEVTYPE OSA CHPID 0C OSD OSA 2127 ATTACHED TO ITSOLNX2 0601 DEVTYPE OSA CHPID 0C OSD OSA 2128 ATTACHED TO ITSOLNX2 0602 DEVTYPE OSA CHPID 0C OSD Ready; T=0.01/0.01 11:14:24 vmrelocate test itsolnx2 to itsossi3 HCPRLH1940E ITSOLNX2 is not relocatable for the following reason(s): HCPRLI1982I ITSOLNX2: Source system real network device 2126 requires an EQID to be assigned HCPRLI1982I ITSOLNX2: Source system real network device 2127 requires an EQID to be assigned HCPRLI1982I ITSOLNX2: Source system real network device 2128 requires an EQID to be assigned Ready(01940); T=0.01/0.01 11:15:47
Solution
Note: Although we used an OSA as an example, this behavior is the same for OSA, Hipersockets, or Small Computer System Interface (SCSI) Fibre Channel Protocol (FCP) devices. All of these devices must be associated with a common EQID. The solution in this case is to associate the devices with a valid EQID or to use another device that already has a valid EQID. Example A-14 on page 95 shows how to define an EQID dynamically. The correct definition must be in the SYSTEM CONFIG file to prevent problems after the next z/VM IPL. In our environment, we defined the EQID during the installation. After the redefinition of the EQID, the Linux guest was xautologged in member ITSOSSI4 and relocated to member ITSOSSI3. The directory entry was not changed. The DEDICATE statements are now valid for ITSOSSI4, but not for ITSOSSI3, which has separate OSA addresses. z/VM identified that the OSAs use the same EQID and made the correct attachments (2126 in ITSOSSI4 was changed to 2106 in ITSOSSI3) automatically. See the last QUERY command, which is the way that VMSSI works with EQIDs.
94
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
Important: This automatic translation was performed by the VMRELOCATE command. It is not valid for the IPL command. To avoid problems during the IPL, correct the directory definition to use NICDEF or COMMAND ATTACH.
Example A-14 Defining EQID dynamically
signal shutdown itsolnx2 Ready; T=0.01/0.01 11:21:34 HCPSIG2113I User ITSOLNX2 has reported successful termination USER DSC LOGOFF AS ITSOLNX2 USERS = 18 AFTER SIGNAL vary off 2126-2128 2126 varied offline 2127 varied offline 2128 varied offline 3 device(s) specified; 3 device(s) successfully varied offline Ready; T=0.01/0.01 11:21:53 set rdev 2126-2128 eqid osa1set1 type osa HCPZRP6722I Characteristics of device 2126 were set as requested. HCPZRP6722I Characteristics of device 2127 were set as requested. HCPZRP6722I Characteristics of device 2128 were set as requested. 3 RDEV(s) specified; 3 RDEV(s) changed; 0 RDEV(s) created Ready; T=0.01/0.01 11:22:27 vary on 2126-2128 2126 varied online 2127 varied online 2128 varied online 3 device(s) specified; 3 device(s) successfully varied online Ready; T=0.01/0.01 11:22:38 xautolog itsolnx2 Command accepted Ready; T=0.01/0.01 11:23:01 AUTO LOGON *** ITSOLNX2 USERS = 19 HCPCLS6056I XAUTOLOG information for ITSOLNX2: The IPL command is verified by the IPL command processor. q osa OSA 2120 ATTACHED TO TCPIP 2120 DEVTYPE OSA CHPID 0C OSD OSA 2121 ATTACHED TO TCPIP 2121 DEVTYPE OSA CHPID 0C OSD OSA 2122 ATTACHED TO TCPIP 2122 DEVTYPE OSA CHPID 0C OSD OSA 2123 ATTACHED TO DTCVSW1 0600 DEVTYPE OSA CHPID 0C OSD OSA 2124 ATTACHED TO DTCVSW1 0601 DEVTYPE OSA CHPID 0C OSD OSA 2125 ATTACHED TO DTCVSW1 0602 DEVTYPE OSA CHPID 0C OSD OSA 2126 ATTACHED TO ITSOLNX2 0600 DEVTYPE OSA CHPID 0C OSD OSA 2127 ATTACHED TO ITSOLNX2 0601 DEVTYPE OSA CHPID 0C OSD OSA 2128 ATTACHED TO ITSOLNX2 0602 DEVTYPE OSA CHPID 0C OSD Ready; T=0.01/0.01 11:23:22 ping 9.12.4.140 Ping Level 620: Pinging host 9.12.4.140. Enter #CP EXT to interrupt. PING: Ping #1 response took 0.000 seconds. Successes so far 1. Ready; T=0.01/0.01 11:23:44 vmrelocate test itsolnx2 to itsossi3 User ITSOLNX2 is eligible for relocation to ITSOSSI3 Ready; T=0.01/0.01 11:23:52 vmrelocate move itsolnx2 to itsossi3
95
Relocation of ITSOLNX2 from ITSOSSI4 to ITSOSSI3 started User ITSOLNX2 has been relocated from ITSOSSI4 to ITSOSSI3 USER DSC LOGOFF AS ITSOLNX2 USERS = 18 FORCED BY SYSTEM Ready; T=0.01/0.01 11:24:12 type itsolnx2 direct USER ITSOLNX2 ITSOSSI 4G 4G G * INCLUDE LINCMS INCLUDE LINDFLT COMMAND DEFINE VFB-512 AS 0111 BLK 128000 IPL 300 MACH ESA 4 DEDICATE 0600 2126 DEDICATE 0601 2127 DEDICATE 0602 2128 * NICDEF 0600 TYPE QDIO LAN SYSTEM ITSOVSW1 MDISK 0300 3390 DEVNO 9A26 MR *DVHOPT LNK0 LOG1 RCM1 SMS0 NPW1 LNGAMENG PWC20111103 CRC: Ready; T=0.01/0.01 11:24:58 at itsossi3 cmd q osa OSA 2100 ATTACHED TO TCPIP 2100 DEVTYPE OSA OSA 2101 ATTACHED TO TCPIP 2101 DEVTYPE OSA OSA 2102 ATTACHED TO TCPIP 2102 DEVTYPE OSA OSA 2103 ATTACHED TO DTCVSW2 0600 DEVTYPE OSA OSA 2104 ATTACHED TO DTCVSW2 0601 DEVTYPE OSA OSA 2105 ATTACHED TO DTCVSW2 0602 DEVTYPE OSA OSA 2106 ATTACHED TO ITSOLNX2 0600 DEVTYPE OSA OSA 2107 ATTACHED TO ITSOLNX2 0601 DEVTYPE OSA OSA 2108 ATTACHED TO ITSOLNX2 0602 DEVTYPE OSA Ready; T=0.01/0.01 11:25:11 at itsossi4 cmd q osa OSA 2120 ATTACHED TO TCPIP 2120 DEVTYPE OSA OSA 2121 ATTACHED TO TCPIP 2121 DEVTYPE OSA OSA 2122 ATTACHED TO TCPIP 2122 DEVTYPE OSA OSA 2123 ATTACHED TO DTCVSW1 0600 DEVTYPE OSA OSA 2124 ATTACHED TO DTCVSW1 0601 DEVTYPE OSA OSA 2125 ATTACHED TO DTCVSW1 0602 DEVTYPE OSA Ready; T=0.01/0.01 11:25:30 at itsossi4 cmd q 2126 id OSA 2126 1732-01 CU: 1731-01 2126 EQID: OSA1SET1 Ready; T=0.01/0.01 11:25:42 at itsossi3 cmd q 2106 id OSA 2106 1732-01 CU: 1731-01 2106 EQID: OSA1SET1 Ready; T=0.01/0.01 11:26:00 ping 9.12.4.140 Ping Level 620: Pinging host 9.12.4.140. Enter #CP EXT to interrupt. PING: Ping #1 response took 0.001 seconds. Successes so Ready; T=0.01/0.01 11:27:17 close cons 11071537 11071537 11071537 11071537 11071537
00 00 00 00 00 00 00 00 00
0C 0C 0C 0C 0C 0C
far 1.
96
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
q vmrelo lnxsu12 10:32:25 Running on member ITSOSSI5 10:32:25 Relocation enabled in Domain SSI Ready; T=0.01/0.01 10:32:25 set vmrelo user lnxsu12 domain itsossi5 10:33:05 Running on member ITSOSSI5 10:33:05 Relocation enabled in Domain ITSOSSI5 Ready; T=0.01/0.01 10:33:05 We then attempted to relocate LNXSU12 to ITSOSSI6. The relocation attempt failed, as we expected. See Example A-16.
Example A-16 Attempt to relocate LNXSU12 to a system on which it is not eligible to run
vmrelocate test lnxsu12 itsossi6 10:33:30 HCPRLH1940E LNXSU12 is not relocatable for the following reason(s): 10:33:30 HCPRLE1944I LNXSU12: Architecture incompatibility Ready(01940); T=0.01/0.01 10:33:30 vmrelocate move lnxsu12 itsossi6 10:33:41 Relocation of user LNXSU12 from ITSOSSI5 to ITSOSSI6 did not complete. Guest has not been moved 10:33:42 HCPRLH1940E LNXSU12 is not relocatable for the following reason(s): 10:33:42 HCPRLE1944I LNXSU12: Architecture incompatibility Ready(01940); T=0.01/0.01 10:33:42
Problem
LNXSU12 was defined to run on ITSOSSI5 only. The responses to the relocation attempt indicate that it failed due to Architecture incompatibility. Example A-17 shows that we attempted to use a relocation command with the FORCE ARCHITECTURE operand. Important: Be careful when using the FORCE option with the VMRELOCATE command. Depending on which resource is forced, the result can be unpredictable, including a guest abend.
Example A-17 VMRELOCATE with FORCE ARCHITECTURE
vmrelocate move lnxsu12 itsossi6 force architecture 10:39:46 Relocation of user LNXSU12 from ITSOSSI5 to ITSOSSI6 did not complete. Guest has not been moved 10:39:47 HCPRLH1940E LNXSU12 is not relocatable for the following reason(s): 10:39:47 HCPRLE1944I LNXSU12: Architecture incompatibility Ready(01940); T=0.01/0.01 10:39:47 Again, the relocation failed for the same reason.
Appendix A. Frequently asked questions
97
Solution
We know that LNXSU12 failed to relocate, because its domain is set so that it can run on ITSOSSI5 only. Therefore, we ran the relocate command with the FORCE DOMAIN operand, as shown in Example A-18.
Example A-18 VMRELOCATE with FORCE DOMAIN operand
vmrelocate move lnxsu12 itsossi6 force domain 10:39:57 Relocation of LNXSU12 from ITSOSSI5 to ITSOSSI6 started with FORCE DOMA IN in effect 10:39:58 User LNXSU12 has been relocated from ITSOSSI5 to ITSOSSI6 Ready; T=0.01/0.01 10:39:58 In Example A-18, the relocation of LNXSU12 from ITSOSSI5 to ITSOSSI6 is successful. The FORCE DOMAIN option succeeds only if the architecture features of ITSOSSI6 are a super-set of the features of ITSOSSI5. If they are not a super-set, the option FORCE ARCHITECTURE DOMAIN is required.
Problem
The vmrelocate command canceled the relocation, because the maxquiesce time is exceeded. See Example A-20. The Linux guest continues to run on the source system.
Example A-20 MAXQUIESCE time is exceeded
vmrelocate move lnxsu11 itsossi6 maxquiesce 3 Relocation of LNXSU11 from ITSOSSI5 to ITSOSSI6 started Relocation of user LNXSU11 from ITSOSSI5 to ITSOSSI6 did not complete. Guest has not been moved HCPRLH1939E Relocation of LNXSU11 to ITSOSSI6 is terminated for the following re ason(s): HCPRLU1930I LNXSU11: The maximum quiesce time was exceeded Ready(01939); T=0.01/0.01 14:33:00
Solution
There are several possible solutions to this time problem, but you need to analyze the environment in more depth. Consider these ideas: Determine whether the application allows a longer quiesce time? (Simplest solution) Do any non-relocation activities make the relocation and quiesce time longer? Are parallel relocations active that increase the time? Are any changes in the Inter-System Facility for Communications (ISFC) setup necessary, such as faster CTC speeds or a larger number of CTCs?
98
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
Problem
When we attempt to relocate LNXSU12 to ITSOSSI6, we get the messages that are shown in Example A-22.
Example A-22 Relocation attempt exceeds the maximum paging space destination
vmrelocate test lnxsu12 itsossi6 HCPRLH1940E LNXSU12 is not relocatable for the following reason(s): HCPRLL1813I LNXSU12: Maximum pageable storage use (10320M) exceeds available aux iliary paging space on destination (7211520K) by 3356160K Ready(01940); T=0.01/0.01 10:54:09 vmrelocate move lnxsu12 itsossi6 Relocation of user LNXSU12 from ITSOSSI5 to ITSOSSI6 did not complete. Guest has not been moved HCPRLH1940E LNXSU12 is not relocatable for the following reason(s): HCPRLL1813I LNXSU12: Maximum pageable storage use (10320M) exceeds available aux iliary paging space on destination (7211520K) by 3356160K Ready(01940); T=0.01/0.01 10:55:32 Correction: The numbers in the HCPRLL1813I message, which are expressed in K bytes, for example, 7211520K, are only bytes. Development is correcting this message. The message states that the maximum pageable storage use exceeds the available auxiliary paging space on the destination member. In our case, we only defined one paging disk, which is insufficient for 10-GB maximum guest storage. Example A-23 shows the paging space that is allocated on ITSOSSI6.
Example A-23 Actual paging space on target member ITSOSSI6
q alloc page EXTENT EXTENT TOTAL PAGES HIGH % VOLID RDEV START END PAGES IN USE PAGE USED ------ ---- ---------- ---------- ------ ------ ------ ---SSI6P1 9E2F 1 10016 1761K 0 0 0% ------ --------SUMMARY 1761K 0 0% USABLE 1761K 0 0% Ready; T=0.01/0.01 10:56:20
99
Solution
We have two possible solutions to reactivate the relocation: Execute the relocation with the FORCE STORAGE option, if you are sure that you have enough main storage on the target member. The paging space is not used. Example A-24 shows the VMRELOCATE command with the FORCE STORAGE option specified. Important: Be careful using the FORCE STORAGE option with the VMRELOCATE command. The result can be unpredictable, including a guest abend or a z/VM abend if the CP abended due to running out of paging space.
vmrelocate move lnxsu12 itsossi6 force storage Relocation of LNXSU12 from ITSOSSI5 to ITSOSSI6 started with FORCE STORAGE in effect User LNXSU12 has been relocated from ITSOSSI5 to ITSOSSI6 Ready; T=0.01/0.01 11:00:42 Add DASD for paging space. See Example A-25.
Example A-25 Increased paging space on destination member
q alloc page EXTENT EXTENT TOTAL PAGES HIGH % RDEV START END PAGES IN USE PAGE USED ---- ---------- ---------- ------ ------ ------ ---9C21 1 10016 1761K 0 0 0% 9E2F 1 10016 1761K 2 15 1% ------ --------SUMMARY 3521K 2 1% USABLE 3521K 2 1% Ready; T=0.01/0.01 11:56:08 VOLID -----SSI6P2 SSI6P1 The check for the eligibility of the guest to relocate no longer shows any messages. See Example A-26.
Example A-26 Check for eligibility ends successfully
vmrelocate test lnxsu12 itsossi6 User LNXSU12 is eligible for relocation to ITSOSSI6 Ready; T=0.01/0.01 11:58:26 vmrelocate move lnxsu12 itsossi6 Relocation of LNXSU12 from ITSOSSI5 to ITSOSSI6 started User LNXSU12 has been relocated from ITSOSSI5 to ITSOSSI6 Ready; T=0.01/0.01 11:59:12
100
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
....
Problem
When we attempt to relocate LNXSU12 to ITSOSSI6, an additional message displays. This message is much more critical than the message in the scenario that is described in A.1.8, Relocation exceeds available auxiliary paging space on page 99. Now, the maximum storage can exceed the available capacity on the destination member. See Example A-28.
Example A-28 Attempt to relocate LNXSU12 exceeds the available capacity on the destination system
vmrelocate test lnxsu12 itsossi6 HCPRLH1940E LNXSU12 is not relocatable for the following reason(s): HCPRLL1811I LNXSU12: Maximum storage use (21136204K) exceeds available capacity on destination (19513900K) by 1622304K HCPRLL1813I LNXSU12: Maximum pageable storage use (20640M) exceeds available aux iliary paging space on destination (14423028K) by 6712613K Ready(01940); T=0.01/0.01 12:01:39
Solution
When you get this message, consider the following possible solutions to fix the storage bottleneck on your destination system: Check whether it is possible to increase the main storage of your system. Add paging space to increase the storage capacity of the system. Execute the relocation with the FORCE STORAGE option
Important: Be careful that you do not reach the capacity limit of your system. The FORCE STORAGE option with the VMRELOCATE command can lead to unpredictable results, including a guest abend or z/VM abend.
A.2 LPAR names are the same on separate CECs in the cluster
The system identifiers that are used on the SSI statement must be unique for each member. So, if your logical partition (LPAR) names are the same, you must use the machine type and serial number for the system identifier. For example, in our environment for the cluster ITSOSSIA, the installation process defined the SYSTEM CONFIG with the statements that are shown in Example A-29.
Example A-29 System identifier that is used by the installation process
/**********************************************************************/ /* System_Identifier Information */ /**********************************************************************/ System_Identifier System_Identifier System_Identifier System_Identifier LPAR LPAR LPAR LPAR A1A A1B A2A A2B ITSOSSI1 ITSOSSI2 ITSOSSI3 ITSOSSI4
To prevent problems with other LPARs with the same name, the definition that is shown in Example A-30 is better.
Appendix A. Frequently asked questions
101
/**********************************************************************/ /* System_Identifier Information */ /**********************************************************************/ System_Identifier System_Identifier System_Identifier System_Identifier 2097 2097 2817 2817 1ADE50 1BDE50 2A3BD5 2B3BD5 ITSOSSI1 ITSOSSI2 ITSOSSI3 ITSOSSI4
For more details, see z/VM V6R2.0 CP Planning and Administration, SC24-6178-02.
We then performed the IPL on z/VM and recovered the saved segments by logging on to MAINT620 and issuing the following command: PUT2PROD SEGMENTS ALL
set observer itsolnx2 * Ready; T=0.01/0.01 15:47:48 Example A-32 shows a sample of several of the messages that are displayed about the new member after the relocation request completed.
102
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
vmrelocate move itsolnx2 itsossi3 Relocation of ITSOLNX2 from ITSOSSI4 to ITSOSSI3 started User ITSOLNX2 has been relocated from ITSOSSI4 to ITSOSSI3 ITSOLNX2: User ITSOLNX2 has been relocated from ITSOSSI4 to ITSOSSI3 ITSOLNX2: User ITSOLNX2 has been relocated from ITSOSSI4 to ITSOSSI3 ITSOLNX2: qeth.3acf0c: 0.0.0600: The qeth device driver failed to recover an err or on the device qeth: irb 00000000: 00 c2 40 17 7f f6 10 38 0e 02 00 00 00 80 00 00 [email protected]... ..... qeth: irb 00000010: 01 02 00 00 00 00 00 00 00 00 00 00 00 00 00 00 ........... ..... qeth: sense data 00000000: 02 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 .... ............ qeth: sense data 00000010: 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 .... ............ . . ..... qeth.fd0b7c: 0.0.0600: A recovery process has been started for the device qdio: 0.0.0602 OSA on SC 2 using AI:1 QEBSM:0 PCI:1 TDD:1 SIGA:RW AO qeth.736dae: ITSOLNX2: CONNECT= 02:18:08 VIRTCPU= 000:04.33 TOTCPU= 000:04.61 ITSOLNX2: LOGOFF AT 13:57:00 EST THURSDAY 11/17/11 BY SYSTEM USER DSC LOGOFF AS ITSOLNX2 USERS = 17 FORCED BY SYSTEM Ready; T=0.01/0.01 13:57:00
These messages are not error messages and can be ignored. When a Linux guest relocates to a new member, it must reestablish its network connection from the new member. These messages relate to reconnecting the new member.
103
104
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
Appendix B.
105
DEFINE RELODOMAIN QUERY RELODOMAIN SET VMRELOCATE QUERY user AT ALL QUERY NAMES SET SSI
106
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
Due to the new structure of the USER DIRECT file, the IDENTITY and SUBCONFIG are treated separately. Each IDENTITY that is defined also has a SUBCONFIG definition for each member of the SSI cluster. A number of new DIRM options to process IDENTITY statements are available.
IDENTITY SRI ******** 16M 16M G ACCOUNT SYSTEMS IPL CMS MACH ESA CONSOLE 0009 3215 SPOOL 000C 2540 READER * SPOOL 000D 2540 PUNCH A SPOOL 000E 1403 A LINK MAINT 0190 0190 RR LINK MAINT 019E 019E RR LINK MAINT 019D 019D RR Create a file for each of the SUBCONFIG statements. Define the resources that are unique to each member. Include the minidisk statement that relates to that member, for example, SRI-n DIRECT A. Example B-2 on page 107 contains an example of the SUBCONFIG statements.
Example B-2 SRI-1 DIRECT A
107
Use the DIRM ADD command to add the IDENTITY statement file, for example, DIRM ADD SRI DIRECT A. Use DIRM ADD identity-1 BUILD ON member IN identity, for example, DIRM ADD SRI-n BUILD ON ITSOSSIn in SRI. Example B-3 shows the directory entry for the IDENTITY after the DIRM ADD BUILD command executes. The DIRM ADD BUILD command automatically updates the IDENTITY with the BUILD statements.
Example B-3 SRI DIRECT A after the DIRM ADD BUILD command executes
DENTITY SRI ******** 16M 16M G BUILD ON ITSOSSI1 USING SUBCONFIG SRI-1 BUILD ON ITSOSSI2 USING SUBCONFIG SRI-2 ACCOUNT SYSTEMS IPL CMS MACH ESA CONSOLE 0009 3215 SPOOL 000C 2540 READER * SPOOL 000D 2540 PUNCH A SPOOL 000E 1403 A LINK MAINT 0190 0190 RR LINK MAINT 019E 019E RR LINK MAINT 019D 019D RR Example B-4 shows the output from the DIRM FOR user REVIEW. The build statements are replaced in the output by the SUBCONFIG entries.
Example B-4 Output from DIRM FOR SRI REVIEW command
IDENTITY SRI ******** 16M 16M G DVHRXV3366I The following configurations will be used on SSI nodes. DVHRXV3366I The following configuration SRI-1 will be used on SSI DVHRXV3366I node ITSOSSI1. SUBCONFIG SRI-1 MDISK 0191 3390 6065 15 SSI1M1 MR *DVHOPT LNK0 LOG1 RCM1 SMS0 NPW1 LNGAMENG PWC20111116 CRCI DVHRXV3366I Preceeding records were included from SRI-1 configuration DVHRXV3366I for node ITSOSSI1. DVHRXV3366I The following configuration SRI-2 will be used on SSI DVHRXV3366I node ITSOSSI2. SUBCONFIG SRI-2 MDISK 0191 3390 4172 15 SSI2M1 MR *DVHOPT LNK0 LOG1 RCM1 SMS0 NPW1 LNGAMENG PWC20111116 CRCRV DVHRXV3366I Preceeding records were included from SRI-2 configuration DVHRXV3366I for node ITSOSSI2. ACCOUNT SYSTEMS IPL CMS MACH ESA CONSOLE 0009 3215 SPOOL 000C 2540 READER * SPOOL 000D 2540 PUNCH A SPOOL 000E 1403 A LINK MAINT 0190 0190 RR LINK MAINT 019E 019E RR LINK MAINT 019D 019D RR *DVHOPT LNK0 LOG1 RCM1 SMS0 NPW0 LNGAMENG PWC20111116 CRC" 108
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
DVHREV3356I The following are your user option settings: DVHREV3356I Links DISABLED Logging ON RcvMsg ON Smsg OFF NeedPW OFF Lang DVHREV3356I AMENG
IDENTITY IDENT INCLUDE IBMDFLT ACCOUNT 9999 CLUSTER1 IPL CMS PARM AUTOCR MACHINE ESA CONSOLE 001F 3215 T SPOOL 000C 2540 READER A SPOOL 000D 2540 PUNCH A SPOOL 000E 1403 A Also, we created a model for the SUBCONFIG entries, one model for each member. We named the files SUBCON-1, SUBCON-2, SUBCON-3, and SUBCON-4, all with the file type PROTODIR. Each file is a model to define an MDISK in the members respective group of DASDs, which were defined in our EXTENT CONTROL file. Example B-6 is the SUBCON-1 PROTODIR file.
Example B-6 Prototype model for a SUBCONFIG in member 1
SUBCONFIG SUBCON-1 MDISK 0191 3390 AUTOG 005 VM6201 MR READ WRITE MULTIPLE Example B-7 on page 109 is the SUBCON-2 PROTODIR file.
Example B-7 Prototype model for a SUBCONFIG in member 2
SUBCONFIG SUBCON-2 MDISK 0191 3390 AUTOG 005 VM6202 MR READ WRITE MULTIPLE
109
SUBCONFIG SUBCON-3 MDISK 0191 3390 AUTOG 005 VM6203 MR READ WRITE MULTIPLE Example B-9 is the SUBCON-4 PROTODIR file.
Example B-9 Prototype model for a SUBCONFIG in member 4
SUBCONFIG SUBCON-4 MDISK 0191 3390 AUTOG 005 VM6204 MR READ WRITE MULTIPLE Example B-10 shows the commands that we used to save the files in DIRMAINT.
Example B-10 Saving the prototype files
IDENT PROTODIR SUBCON-1 PROTODIR SUBCON-2 PROTODIR SUBCON-3 PROTODIR SUBCON-4 PROTODIR
Each time that we need one or more new multiconfiguration virtual machines, we use the prototypes. Example B-11 is a BATCH file that creates two new machines. We named the file DIRMADDU BATCH.
Example B-11 The batch file to create two new multiconfiguration virtual machines
add add add add add for add add add add add for
cmsusr3 like ident cmsus3-1 like subcon-1 cmsus3-2 like subcon-2 cmsus3-3 like subcon-3 cmsus3-4 like subcon-4 cmsusr3 setpw initpw cmsusr4 like ident cmsus4-1 like subcon-1 cmsus4-2 like subcon-2 cmsus4-3 like subcon-3 cmsus4-4 like subcon-4 cmsusr4 setpw initpw
on on on on
in in in in
on on on on
in in in in
Example B-12 shows the command that we used to send this file to be executed in DIRMAINT.
Example B-12 Executing the batch file
dirm batch dirmaddu batch After we created these user IDs, we personalize each machine. We use individual DIRMAINT commands, such as the command that is shown in Example B-11 on page 110. Or, you can use the traditional GET/REP commands for get and replace functions.
110
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
111
112
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
Appendix C.
Additional material
This book refers to additional material that can be downloaded from the Internet as described in the following sections.
113
114
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
Related publications
The publications listed in this section are considered particularly suitable for a more detailed discussion of the topics covered in this book.
Other publications
These publications are also relevant as further information sources: z/VM V6R2.0 CP Planning and Administration, SC24-6178-02 z/VM V6R2.0 Getting Started with Linux on System z, SC24-6194 z/VM V6R2.0 Installation Guide, GC24-6246 z/VM CMS File Pool Planning, Administration and Operation, SC24-6167 z/VM Migration Guide, GC24-6201 z/VM V6R2.0 General Information, GC24-6193-02 z/VM V6R2.0 VMSES/E Introduction and Reference, GC24-6243-01 z/VM V6R2.0 Directory Maintenance Facility Commands Reference, SC24-6188-02 z/VM V6R2.0 Connectivity, SC24-6174-02 z/VM V6R2.0 CP Commands and Utilities Reference, SC24-6175-02 z/VM V6R2.0 Migration Guide, GC24-6201 z/VM V6R2.0 RACF Security Server System Programmers Guide, SC24-6219
Online resources
These websites are also relevant as further information sources: z/VM V6.2 resources http://www.vm.ibm.com/zvm620/ z/VM program directories http://www.ibm.com/vm/progdir/
115
116
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
Index
A
APPC/VM VTAM Support (AVS) 6 HCPZRP6722I Characteristic 93 host 9.12.4.236 78
B
Business Class (BC) 19
I
ICK00001I Function 14 IFL environment 34 Inter-System Facility 6, 40 Inter-System Facility for Communications (ISFC) 6, 19, 40 IPL 300 88 IPL time 51, 75 Clear TDisks 51 duplicate volsers 75 ITSOSSI5 system hardware definition 64 id 67 IUCV connection 25
C
Changed PROFILE (CP) 56, 68 Channel to Channel (CTC) 15 CMS guest 91 Collaborative Memory Management Assist (CMMA) 26 COMMAND CP Spool 57 Term 57 CONFIGSS DATADVH 72 Control Program (CP) 6, 24, 59 CP Planning 17, 22 CTC connection 14, 40 CYL 21 49, 75 CYL 30 49, 75
L
Linux guest 10, 23, 36, 63, 88 amended directory entry 93 certain eligibility criteria 45 PROFILE EXEC file 57 Linux machine 57, 89 Linux view 59 Live Guest Relocation (LGR) 1, 10, 23, 36, 57, 78 local minidisk 34, 89
D
DASD 10, 37, 62 DASD 9A26 CP SYSTEM US9A26 89 Dasd Planning 19 DASD volume 4, 38 DDR copy 67 default domain 28 destination member 23, 88 available space 26 paging space 26 Device type 65 directory entry 57, 88 discontiguous saved segment (DCSS) 25 domain X 28
M
MACH ESA 2 58 4 77, 88 machines are running (MR) 57
N
named saved system (NSS) 25 ni 67, 82
E
Enterprise Class (EC) 19 EQID 15, 92, 106 EQID concept 16
O
One CTC 40 OSA address 69, 94 OSA card 92
F
FICON channel 19
P
performance data 82 Performance Toolkit 35, 63 Program Directory 56 Persistent Data Record (PDR) 18, 48, 70 PROFILE GCS 55 PROFILE TCPIP 56, 69 PUT2PROD Segment 102, 111
G
General-instructions-extension facility (GIEF) 29
H
HCPRLE1944I LNXSU12 97 Copyright IBM Corp. 2012. All rights reserved.
117
R
Redbooks website Contact us xi Relocation Domain Business reasons 28 Characteristics 28 virtual servers 31 relocation domain 17, 28, 45, 97, 106 Remote Spooling Communication Subsystem (RSCS) 63 Remote Spooling Communications Subsystem Program Directory 54 Remote Spooling Communications Subsystem (RSCS) 35
IDENTITY entries 11 user directory 2, 10, 67 User ITSOLNX1 33, 90 user itsolnx3 59 user itsolnx4 59 user LNXSU12 97
V
VIRTCPU 48, 78, 90 Virtual machine 2, 23, 34, 89 device 019D 89 device 019E 89 virtual machine concurrent support 2 uncontrolled numbers 2 vmrelocate move itsolnx1 itsossi2 maxtotal 10 33 itsolnx2 95 lnxsu11 itsossi6 81 lnxsu12 itsossi6 97 lnxsu12 itsossi6 force architecture 97 lnxsu12 itsossi6 force domain 98 user itsolnx2 90 VMRELOCATE TEST command 33, 81 command check 58
S
same EQID 15, 34, 48, 94 free device 34 same time 1011 Scenario 2 20, 62 Shared File System (SFS) 17, 36 si 82 Single System Image additional planning guidelines 5 z/VM version 6.2 1 single system image (SSI) 2, 9, 48, 106 Software Setup 54 Spool volume 37, 65 SSI Name 21, 48 st 82 sub-configuration entry 73 sy 82 SYSTEM CONFIG 47, 67, 94 System Config definition 15 file 12, 15, 30, 48, 68, 71, 102 SYSTEM CONFIG file 4849 SYSTEM NETID file 69 System z 25, 79 family 2 guest 19 hardware 80 operating environment 3 SC24-6194 56 Setup 80 System_3270 System_Console 52
W
Warmstart Information 49, 75 WORK volume 19, 62
Z
z/VM 1, 3 z/VM 6.2 15, 80 relocation feature 80 z/VM member 2, 24, 35 software maintenance 24 z/VM system 4, 10, 20, 47, 54, 62, 69, 102, 106 z/VM v6.2 1 priced feature 7 z196 machine 36 ITSOSSI4 run 37
T
TCPIP Data 69 TDISK Volume 50 test environment 3, 28 TOTCPU 48, 78, 90 Transparent Services Access Facility (TSAF) 6
U
User 610ZVM 77 User Directory
118
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
Back cover An introduction to z/VM Single System Image (SSI) and Live Guest Relocation (LGR)
Learn about the business benefits of SSI and LGR Understand important concepts and planning Step through implementation and practical approaches
IBM z/VM 6.2 introduces significant changes to z/VM in the form of multiple system clustering technology allowing up to four z/VM instances in a single system image (SSI) cluster. This technology is important, because it offers clients an attractive alternative to vertical growth by adding new z/VM systems. In the past, this capability required duplicate efforts to install, maintain, and manage each system. SSI reduces or eliminates these duplicate efforts. Support for live guest relocation (LGR) allows you to move Linux virtual servers without disruption to the business, helping you to avoid planned outages. The z/VM systems are aware of each other and can take advantage of their combined resources. LGR enables clients to avoid loss of service due to planned outages by relocating guests from a system requiring maintenance to a system that remains active during the maintenance period. Together, the SSI and LGR technologies offer substantial client value, and they are a major departure from past z/VM practices. This IBM Redpaper publication gives you a broad understanding of the new SSI architecture and an overview of LGR. We show an LGR example that shows a typical SAP user environment. In our example, the SAP Application Server Central Instance resides on a Linux on System z guest and an IBM DB2 10 database server runs on z/OS. This book is written for IT architects, who design the systems, and IT specialists, who build the systems.