You are on page 1of 82

IBM Platform HPC

Version 4.2

IBM Platform HPC, Version 4.2
Installation Guide 

SC27-6107-02

IBM Platform HPC
Version 4.2

IBM Platform HPC, Version 4.2
Installation Guide 

SC27-6107-02

Note
Before using this information and the product it supports, read the information in “Notices” on page 71.

First edition
This edition applies to version 4, release 2, of IBM Platform HPC (product number 5725-K71) and to all subsequent
releases and modifications until otherwise indicated in new editions.
© Copyright IBM Corporation 1994, 2014.
US Government Users Restricted Rights – Use, duplication or disclosure restricted by GSA ADP Schedule Contract
with IBM Corp.

. . . . . . . . . 39 Configuring your browser . . . . . . . . . . 2014 . . . Configure and test switches . . 3 Chapter 2.x prerequisites . . . . . . . . 73 . . . . . . . 37 Chapter 8. . . . . . Preparing to upgrade .1. . . . . 40 Chapter 9. . 2 . . . . . . . High availability definition Commands . . . . . Installation planning Preinstallation roadmap Installation roadmap. . 49 Upgrading to Platform HPC Version 4. . . Upgrading to Platform HPC 4. Troubleshooting upgrade problems . . . . Performing an installation Comparing installation methods Quick installation roadmap . . . . . 16 Chapter 4. 29 Completing the high availability enablement . . . pcmhatool . . . . . . . . . . Privacy policy considerations . . . . . . Troubleshooting installation problems. . . . . 44 44 45 45 46 . . . . 71 Trademarks . . . . . . . Setting up a high availability environment . . . 5 . . . . . Performing the Platform HPC upgrade . . . . . . . . 9 10 12 12 13 17 . . . Planning. . . . . .2 without OS reinstall. . . . . . . . . . . . . . . Upgrade planning . . Verifying the installation . . Upgrading LSF entitlement . . . . 29 Response file for silent installation. . . . . . . . . Applying fixes . . Troubleshooting a high availability environment enablement . . . . . . . . . . Installing and configuring the operating system on the management node . . . . . . . . . . Completing the upgrade . . . Backing up Platform HPC . . . 17 19 20 23 24 Chapter 5. . . Configure IPMI as a fencing device . . . . . . . . . . . . .2 . . . . . . . . Verifying the upgrade . . . Enable a high availability environment . . . . . 65 Chapter 7.Contents Chapter 1. . . . . . . . . . . . . . Plan your network configuration . . Quick installation . . . . . . . . 7 Chapter 3. . . . 49 49 49 50 50 50 51 52 53 55 55 55 57 57 58 60 60 61 63 63 63 Chapter 6. . . . . 13 . . . Rollback to Platform HPC 4. . . . . . . . . . . . . Verifying a high availability environment . . . . . . . . . . . . . . . . . . . Backing up Platform HPC . © Copyright IBM Corp. . . . . . . . . . . . . . . file . Upgrading entitlement. . . . . . . . .1 . . Upgrading IBM Platform HPC . . . . . . 1994. . . . . . . . . . . . Custom installation . . . . . . 41 . . . . . . . 47 Chapter 10. . Create a failover notification. . . . . . . . . . . . . Preparing to install PHPC . . . . . Upgrading checklist . 67 . . Performing a silent installation . . . . . . . . . . . . . . . . . . . . . . . . . 43 Chapter 12. . . . . . . . . . . . . . . . . Upgrading PAC entitlement . . 15 . . . . . . . . Red Hat Enterprise Linux prerequisites . Verifying the upgrade . . . 9 PHPC requirements . . . . .1. . . . 5 Planning your system configuration . . . . . . 35 Chapter 11. . . . Prepare a shared file system . . . Completing the upgrade . . . Custom installation roadmap . . . . . . . . . . . . . 73 iii . . . . Taking the first steps after installation . . 41 Preparing high availability . . 1 . . . . . . . . . . . . . . . . . . . . . . . . . . . High availability requirements . . Planning a high availability environment . . . . SUSE Linux Enterprise Server (SLES) 11. . . . . . . . Setting up SMTP mail settings . . . . . References Configuration files . . . . . . . . . . . Upgrading to Platform HPC 4. . . . . . . . . . . . . . . . .2 with OS reinstall Preparing to upgrade . . . . . . . . . . . . . . . . Performing the Platform HPC upgrade . 67 67 68 68 Notices . . . . Upgrading roadmap . .

iv Installing IBM Platform HPC Version 4.2 .

and fault-tolerant infrastructure. and monitors its progress. finds the best resources to run the job. where public-IP-address is the public IP address of your Platform HPC management node To upgrade your product entitlement for LSF refer to “Upgrading LSF entitlement” on page 63. For more information on submitting and compiling MPI jobs. more reliable workload performance. scalable. © Copyright IBM Corp. Review the preinstallation and installation roadmaps before you begin the installation process.2/docs/lsf/. and includes steps for installing Platform HPC. IBM Platform MPI is installed with IBM Platform HPC. IBM Platform LSF is an enterprise-class software that distributes work across existing heterogeneous IT resources creating a shared. For building MPI applications. delivering faster. This LSF workload management component is installed as part of Platform HPC installation. LSF balances load and allocates resources. You can find the IBM Platform LSF here: http://public-IP-address/install/kits/kit-phpc-4. LSF provides a resource management framework that takes your job requirements. 2014 1 . while providing access to those resources. Jobs always run according to host load and site policies. you must have one of the supported compilers installed.Chapter 1. Installation planning Installing and configuring IBM® Platform HPC involves several steps that you must complete in the appropriate sequence. 1994. Platform HPC includes a Platform LSF workload management component. see the IBM Platform MPI User's Guide 9. As part of the IBM Platform HPC installation. Refer to the IBM Platform MPI release notes for a list of supported compilers. and the workload management master daemon to be configured running on the node same as the Platform HPC management node. IBM Platform MPI By default. refer to the IBM Platform LSF Administration guide. the following components are installed: v IBM Platform LSF® v IBM Platform MPI Workload management with IBM Platform LSF IBM Platform HPC includes a workload management component for load balancing and resource allocation. The IBM Platform MPI release notes are in the /opt/ibm/platform_mpi/doc/ directory. The Installation Guide contains information to help you prepare for your Platform HPC installation.1 (SC27-5319-00). For more information on IBM Platform LSF.

Preinstallation roadmap
Before you begin your installation, ensure that the preinstallation tasks are
completed.
There are two cases to consider before installing Platform HPC, including:
v Installing Platform HPC on a bare metal management node.
v Installing Platform HPC on a management node that already has an operating
system installed.
If you are installing Platform HPC on a management node that already has an
operating system that is installed, you can omit preinstallation actions 4 and 5.
Table 1. Preinstallation roadmap

1.

Actions

Description

Plan your cluster

Review and plan your cluster setup.
Refer to “Planning your system
configuration” on page 5.

1.

Review Platform HPC requirements

Make sure that the minimum hardware
requirements are met, including:
v Hardware requirements
v Software requirements
Refer to “PHPC requirements” on page 9.

2.

Configure and test switches

Ensure that the necessary switches are
configured to work with Platform HPC.
Refer to “Configure and test switches” on
page 12.

3.

Plan your network configuration

Before proceeding with the installation,
plan your network configuration,
including:
v Provision network information
v Public network information
v BMC network information
Refer to “Plan your network
configuration” on page 13.

4.

Obtain a copy of your operating system

If the operating system is not installed,
you must obtain a copy of your operating
system and install it.

5.

Install and configure your operating
system

Ensure that you configure your operating
system:
v Decide on a partitioning layout
v Meet the Red Hat Enterprise Linux 6.x
prerequisites
Refer to “Installing and configuring the
operating system on the management
node” on page 13.

6.

2

Obtain a copy of IBM Platform HPC

Installing IBM Platform HPC Version 4.2

If you do not have a copy of IBM
Platform HPC, you can download it from
IBM Passport Advantage®.

Installation roadmap
This roadmap helps you navigate your way through the PHPC installation.
Table 2. Installation roadmap

1.

Actions

Description

Select an installation method

Choose an installation method from the
following:
v Installing PHPC using the installer.
Using the installer you have the
following choices:
– Quick installation
– Custom installation
v Installing PHPC using silent mode
Refer to Chapter 1, “Installation
planning,” on page 1.

2.

Perform the installation

Follow your installation method to
complete the PHPC installation.

3.

Verify your installation

Ensure that PHPC is successfully
installed.
Refer to Chapter 6, “Verifying the
installation,” on page 35.

4.

Troubleshooting problems that occurred
during installation

If an error occurs during installation, you
can troubleshoot the error.
Refer to Chapter 8, “Troubleshooting
installation problems,” on page 39.

5.

(Optional) Upgrading product
entitlement

Optionally, you can update your product
entitlement for LSF.
Refer to “Upgrading LSF entitlement” on
page 63.

6.

(Optional) Apply PHPC fixes

After you install PHPC, you can check if
there are any fixes available though the
IBM Fix Central.
Refer to Chapter 11, “Applying fixes,” on
page 65.

Chapter 1. Installation planning

3

4

Installing IBM Platform HPC Version 4.2

Chapter 2. Planning
Before you install IBM Platform HPC and deploy system, you must decide on your
network topology, and system configuration.

Planning your system configuration
Understand the role of the management node and plan your system settings and
configurations accordingly. IBM Platform HPC software is installed on the
management node after the management node meets all requirements.
The management node is responsible for the following functions:
v Administration, management, and monitoring of the system
v Installation of compute nodes
v Operating system distribution management and updates
v System configuration management
v Kit management
v
v
v
v
v

Provisioning templates
Stateless and stateful management
User logon, compilation, and submission of jobs to the system
Acting as a firewall to shield the system from external nodes and networks
Acting as a server for many important services, such as DHCP, NFS, DNS, NTP,
HTTP

The management node connects to both a provision and public network. Below,
the management node connects to the provision network through the Ethernet
interface that is mapped to eth1. It connects to the public network through the
Ethernet interface that is mapped to eth0. The public network refers to the main
network in your company or organization. A network switch connects the
installation and compute nodes together to form a provision network.

© Copyright IBM Corp. 1994, 2014

5

6 Installing IBM Platform HPC Version 4.2 . is an example of a system where compute nodes share a provisioning port. Below. the provision and BMC network can be the same. For networks where compute nodes have the same port for an Ethernet and BMC connection. Multiple compute nodes are responsible for calculations.Figure 1. They are also responsible for running batch or parallel jobs. System with a BMC network Each compute node can be connected to the provision network and the BMC network.

By default. to perform message passing over a much faster network using a high-speed interconnect such as InfiniBand. System with a combined provision and BMC network Note: For IPMI using a BMC network. Chapter 2. For more information about a particular interconnect. Alternatively. the two Ethernet interface configurations is the most common. The provision network connects the management node and compute nodes is typically a Gigabit or 100-Mbps Ethernet network. eth0 can be the public interface and eth1 the provision interface. eth0 is connected to the provision interface and eth1 is connected to the public interface. contact the appropriate interconnect vendor. Planning a high availability environment A high availability environment includes two installed PHPC management nodes locally with same software and network configuration (except the hostname and IP address). the provision network serves three purposes: v System administration v System monitoring v Message passing It is common practice. In this simple setup.Figure 2. Planning 7 . however. High availability is configured on both management nodes to control key services. Note: You can also connect compute nodes to an InfiniBand network after the installation. you must use eth0 in order for the BMC network to use the provision network. Although other system configurations are possible. A fast interconnect provides benefits such as higher throughput and lower latency.

8 Installing IBM Platform HPC Version 4.2 .

v Configure and test switches. 1994. Refer to the PHPC requirements for a list of supported operating systems. v Obtain a copy of the product. steps must be taken to ensure all prerequisite are met. Before installing PHPC. Hardware requirements Before you install PHPC. © Copyright IBM Corp. Minimum requirements for compute node for stateful package-based installations: v 1 GB of physical memory (RAM) for compute nodes v 40 GB of free disk space v One static Ethernet interface Minimum requirements for compute node for stateless image-based installations: v 4 GB of physical memory (RAM) v One static Ethernet interface Optional hardware can be configured before the installation: v Additional Ethernet interfaces for connecting to other networks v Additional BMC interfaces v Additional interconnects for high-performance message passing. such as InfiniBand Note: Platform HPC installation on an NFS server is not supported. Preparing to install PHPC Before installing PHPC. v Plan network configuration. the management node must be a node that is not in the IBM Flex Chassis. Minimum hardware requirements for the management node: v 100 GB free disk space v 4 GB of physical memory (RAM) v At least one static Ethernet configured interface Note: For IBM PureFlex™ systems.Chapter 3. 2014 9 . v Obtain a copy of the operating system. PHPC requirements You must make sure that the minimum hardware and software requirements are met. You must make sure that the minimum hardware and software requirements are met. you must complete the following steps: v Check the PHPC requirements. you must make sure that minimum hardware requirements are met. v Install an operating system for the management node.

If the secondary node has a different partition layout. Additionally.3 x86 (64-bit) High availability requirements You must make sure that these requirements are met before you set up high availability. Use df -h and fdisk -l to check the partition layout. v The management nodes must have the same partition layout. time zone.2 . issue the ifconfig command to check that the network settings are the same. you can use sequential IP addresses. all IP addresses (management nodes IP addresses and virtual IP address) are in the IP address range of your network. v The management nodes must use the same network interface to connect to the provision and public networks. v The management nodes must use the same network settings. reinstall the operating system with the same partition layout. reconfigure the network interfaces on the secondary management node according to your network plan. On each management node. Sequential IP addresses can help avoid any issues. and current date.5 x86 (64-bit) v SUSE Linux Enterprise Server (SLES) 11. Management node requirements Requirements for the primary management node and the secondary management node in a high availability environment: v The management nodes must have the same or similar hardware requirements. To ensure that all IP addresses are in the IP address range of your network. ensure that the IP address of same network interface is in the same subnet.Software requirements One of the following operating systems is required: v Red Hat Enterprise Linux (RHEL) 6. For example: 10 Installing IBM Platform HPC Version 4. After you prepare the secondary management node. v The management nodes must be configured with the same time. Collect the following high availability information: v Virtual management node name v Virtual IP address for public network v Virtual IP address for provision network v Shared directory for user home v Shared directory for system work data Note: In a high availability environment. If not. Virtual network requirements Virtual network information is needed to configure and enable high availability. Ensure that the same network interfaces are defined for the primary and secondary management nodes. you can ensure that the secondary node uses the same partition schema as the primary management node.

7. Preparing to install 11 . This ensures that no single point of failure (SPOF) errors occur. the following shared file system requirements must be met: v The shared file systems cannot be one of the management nodes. Required disk space varies based on the cluster usage.168.168.5 Network Shared file system requirements Shared file systems are required to set up a high availability environment in Platform HPC. By default.7. all shared file systems must be accessible by the provision network for both the management nodes and compute nodes. and execute permissions must be enabled for the operating system root user and the Platform HPC administrator.20. the user data is stored together with system work data.0.7.168. v The read and write permissions must be enabled for all users. use the following commands: Chapter 3.0.3 192.fsid=num where num is an integer and should be different for each shared directory.200 192.0. By default.no_root_squash. the Platform HPC administrator is phpcadmin. two shared directories are required in a high availability environment. Additionally. v If using an external NAS or NFS server to host the shared directories that are needed for high availability.3172.20. it must be added as an unmanaged device to the cluster to avoid any IP address errors. Example: Sequential IP addresses Primary management IP address range node Secondary management node Virtual IP address public 192. the following parameters must be specified in the exports entries: rw.5 provision 172.Table 3.7. to create a shared data and a shared home directory on an external NFS server.4 172.20.20. and it varies based on the disk space requirements for each user and the total user number.7.20.200 172. v The read. The following shared file systems must already be created on your shared storage server before you set up and enable a high availability environment: Shared directory for system work data v The minimum available shared disk space that is required is 40 GB. Shared directory for user data (/home) v Ensure that there is enough disk space for your data in your /home directory. v The shared file systems should be specific to and only use for the high availability environment.168. one to store user data and one to store system work data. Refer to Unmanaged devices. v If the IP address of the shared storage server is in the network IP address range that is managed by Platform HPC.0. If not provided.4 192.168. The minimum available shared disk space that is required is 4 GB.3192.0.3 172. For example.sync. In a high availability environment. write.

enable PortFast on the switch. ensure that the fsid parameter is set for each export entry. A shared file system is used in high availability to store shared work and user settings. Run the following command on both management nodes to ping the NFS server from provision network. 64 bytes from 192.255. These issues include: nodes that cannot PXE boot.0/24(rw.168.1 PING 192.1: /export/data 192.1.7.20.168. Disable the Spanning Tree on switched networks.20.168.1. run the following commands.1.036 ms 2. 12 Installing IBM Platform HPC Version 4.0 /export/home 192.1.1) from 192. modify the /etc/exports file on the external NFS server.no_root_squash.1.255.7.168.1 Created unmanaged node.0/24(rw. /export/ 172.sync.mkdir -p /export/data mkdir -p /export/home Next. nodes cannot go into interactive startup. the switch begins forwarding the packets as it begins receiving them. # ping -c 2 -I eth1 192. On the primary management node.168. If currently disabled. For example: /export/data 172.0/255.1 (192.1. Add the NFS server as an IP pool to the Platform HPC system.168.168.168. It is the forwarding scheme that the switch uses. Some installation issues can be caused by misconfigured network switches.fsid=4) Prepare a shared file system Before you enable high availability.no_root_squash.168.255.168.1.3 eth1: 56(84) bytes of data.1. This prevents the IP address of the NFS server from being allocated to a compute node and ensures that the NFS server name can be resolved consistently across the cluster. # showmount -e 192.0/255.1.fsid=0) Note: If you are using two different file systems to create the directories.1: icmp_seq=1 ttl=64 time=0. complete the following steps: 1. Configure and test switches Before installing IBM Platform HPC ensure that your Ethernet switches are configured properly. 2.1.sync.0/24(rw. #plcclient. nodes cannot download a kickstart file.1 Export list for 192.168. This speeds the PXE booting process.sh -p pcmnodeloader Loaders startup successfully.0 3. Different switch manufacturers may use different names for Portfast. To ensure that the Ethernet switches are configured correctly.no_root_squash.20.051 ms 64 bytes from 192. Confirm that the NFS server can be used for the high availability configuration and that it is accessible from the Platform HPC management nodes.sync. Enabling PortFast if it is supported by the switch is recommended.2 .7.fsid=3) /export/home 172. Procedure 1.1: icmp_seq=2 ttl=64 time=0.1. View the list of all NFS shared directories available on the NFS server. prepare a shared file system.255. #nodeaddunmged hostname=nfsserver ip=192. For best installation performance.

including if you are setting up a high availability environment. Plan your network configuration Before installing Platform HPC.domain.3. Preparing to install 13 . Information about your network is required during installation. including information about the management nodes. Run diagnostics on the switch to ensure that the switch is connected properly. collect the information for both management nodes. Note: If you are setting up a high availability environment. and network details. 4. Installing and configuring the operating system on the management node Before you can create the PHPC management node. the management node information is required for both the primary management node and the secondary management node. enable multicasting on the switch. including: v Provision network information: – Network subnet – Network domain name – Static IP address range v Public network information: – Network subnet – Network domain name – Static IP address range v BMC network information: – Network subnet – Network domain name – Static IP address range v Management node information: – Node name (use a fully qualified domain name with a public domain suffix. If currently disabled. for example: management. the primary management node and the secondary management node. Certain switches might need to be configured to allow multicast traffic on the private network. and there are no bad ports or cables in the configuration. The following information is needed to setup and configure your network.com) – Static IP address and subnet mask for public network – Static IP address and subnet mask for provision network – Default gateway address – External DNS server IP address Note: For a high availability environment. you must install an operating system on the management node. Plan your network details. ensure that you know the details of your network configuration. Chapter 3.

v The /home directory must be writable. ensure that the following conditions are met: v Decide on a partitioning layout. comment out the following line to disable IPv6: install ipv6 /bin/true For SLES: If 50-ipv6.Complete the following steps to install the operating system on the management node: 1. comment out the following line to disable IPv6: #install ipv6 /bin/true Note: After you install the operating system.d directory. you can customize the disk partitioning on compute nodes by creating a custom script to configure Logical Volume Manager (LVM) partitioning. Use the date command to check the date on the operating system. v Configure at least one static network interface. and date -s command to set the date.conf file exists in the /etc/modprobe. To enable IPv6 do the following: For RHEL: If the disable-ipv6. Do not disable IPv6 during the operating system installation. ensure that the operating system time is set to the current real time. Install and configure the operating system. If you do upgrade your operating system through yum or zypper then you must roll back your changes before proceeding with the PHPC installation. For example: date -s "20131017 04:57:00" Important: The management node does not support installing on an operating system that is upgraded through yum or zypper update. run the following command as root: # chmod u+w /home # ls -al / |grep home v The package openais-devel must be removed manually if it is already installed. make sure that shadow passwords authentication is enabled. you must first disable the autofs configuration: # chkconfig autofs off # service autofs stop To make the /home directory writable.2 . Obtain a copy of the operating system. 14 Installing IBM Platform HPC Version 4. 2. Run setup and make sure Use Shadow Passwords is checked. v Before you install PHPC on the management node. You can update the management node's operating system after installation. The suggested partitioning layout is as follows: – Ensure that the /opt partition has a least 4 GB – Ensure that the /var partition has at least 40 GB – Ensure that the /install partition has at least 40 GB Note: After you install Platform HPC. Before you install the operating system on the management node. v Use a fully qualified domain name (FQDN) for the management node. v Ensure that IPv6 is enabled for remote power and console management. Do not run a yum update (RHEL) or zypper update (SLES) before installing PHPC. If the /home directory is mounted by autofs.d directory.conf file exists in the /etc/modprobe.

If you are installing the Red Hat Enterprise Linux (RHEL) 6. Ensure that the traditional naming scheme ethN is used. If not. 4. To stop the NetworkManager service. The content of these files is distribution-specific (see /usr/share/doc/initscripts-version for details). 2. Reboot the system. b. d. Refer to the operating system documentation for how to install the operating system.x operating system. install the operating system. b. Reboot the management node. b. Log in to see the ethN names. you must revert to the traditional naming scheme ethN: a. Disable SELinux. Comment the HWADDR variable in the ifcfg-eth* files if present as it is not possible to predict here which of the network devices is named eth0. Red Hat Enterprise Linux prerequisites Before you install Platform HPC on Red Hat Enterprise Linux (RHEL) 6. Check whether the package net-snmp-perl is installed on the management node. Before installing PHPC. a. Modify it to suit the new naming scheme such as DEVICE=eth0. On the management node. ifcfg-ethN files in RHEL 6. see the additional RHEL prerequisites. edit the /etc/selinux/config file to set SELINUX=disabled. a. Reboot the management node. you must install it manually from the second RHEL 7 on POWER ISO. 1. To stop the NetworkManager service. eth1 etc. Chapter 3. you must stop the NetworkManager service.x contain a DEVICE= field which is assigned with the emN name. If you have a system that does not use the traditional naming scheme ethN. After all the conditions and prerequisites are met. you must stop the NetworkManager service. you must ensure the following: 1. c. Preparing to install 15 . run the following command: /etc/init. Disable SELinux.d/NetworkManager stop 3. The 70-persistent-net.x. edit the /etc/selinux/config file to set SELINUX=disabled. Rename all ifcfg-emN and ifcfg-p* configuration files and modify the contents of the files accordingly. For example.x operating system. 2.d/ to make the names persistent across reboots. run the following command: /etc/init. If you are installing the SUSE Linux Enterprise Server (SLES) 11.d/NetworkManager stop 3. On the management node. Before installing PHPC.rules file is created under /etc/udev/rules. see the additional SLES prerequisites.

iso b. select the System Services (Runlevel) option. Create a software repository for each ISO image using the YaST configuration and setup tool. f. e. prepare the following ISO images: v Installation ISO image: SLES-11-SP3-DVD-x86_64-GM-DVD1.SUSE Linux Enterprise Server (SLES) 11. 2. Install the createrepo and perl-DBD-Pg packages. 4) Enter the Repository Name and select a Path to ISO Image. complete the following steps: 1) Start the YaST configuration and setup tool in a terminal. you must complete the following steps. c. select the Software Repositories option and click Add. You must disable AppArmor. To save the options click OK. Select the Expert Mode option.2 . Start the YaST configuration and setup tool. complete the following steps: a. go to the Set/Reset menu and select Disable the service. Exit the YaST configuration and setup tool by clicking OK. 1. 2) From the Software menu. To disable AppArmor. b.x prerequisites Before you install Platform HPC on SUSE Linux Enterprise Server (SLES). Reboot the management node. To install the packages. 3) Select the Local ISO Image option and click Next. complete the following steps: a. run the following command: zypper install createrepo perl-DBD-Pg 3. Click Next. 16 Installing IBM Platform HPC Version 4. c. d. From the System menu. To create a software repository. If the createrepo and perl-DBD-Pg packages are not installed. You must create a software repository for both the installation ISO image and the SDK ISO image.iso v SDK ISO image: SLE-11-SP3-SDK-DVD-x86_64-GM-DVD1. Select the boot.apparmor service. 5) Click OK to save the options and exit the YaST configuration and setup tool.

Comparing installation methods IBM Platform HPC can be installed using an interactive installer in one of two methods.if no problems are found for the checked item v WARNING . and the installation cannot begin until a valid path is used.if configuration of an item does not match the requirements. Below is a complete comparison table of the two installation methods and the default values provided by the installer. Performing an installation Install PHPC using the installer. Yes Yes © Copyright IBM Corp. The installation path cannot include special characters such as a colon (:). the installation quits The installer (phpc-installer) displays the corresponding error message for problems that are detected and automatically ends the installation. Installer option comparison Options Default value Option included in the Quick installation? (Yes/No) Select a mount point for the depot (/install) directory. The installer enables you to specify your installation options. 2014 eth0 Option included in the Custom installation? (Yes/No) 17 . The custom installation method provides added installation options and enables the administrator to specify additional system configurations. the default values are used for values not specified during installation. Table 4. v A valid installation path for the installer must be used. you must resolve the identified problems then rerun the phpc-installer until all installation requirements are met. / Yes Yes Select the location that you want CD/DVD drive to install the operating system from. The quick installation method sets up quickly sets up basic options with default options.Chapter 4. 1994. exclamation point (!) or space. Usage notes v Do not use an NFS partition or a local /home partition for the depot (/install) mount point. v In the quick installation. the installer automatically checks the hardware and software configurations. the quick installation method and the custom installation method. After the installation starts. Yes Yes Specify a provision network interface. If there are errors.if the installer cannot recover from an error. installation continues despite the warnings v FAILED . The installer displays the following based on the results: v OK .

specify a subnet mask for the BMC network 255.1. public.1.3-192. specify a subnet for the BMC network N/A Yes Yes If creating a new BMC network.com Yes Yes Select a BMC network.dns.zone Yes Yes Set the domain name for the public network.255. Hardware profile. options include: IBM_System_x_M4 Yes Yes Set the domain name for the provision network. N/A specify a gateway IP address for the BMC network No Yes If creating a new BMC network.168.0 Yes Yes If creating a new BMC network.Table 4. eth1 Yes Yes Do you want to enable a public network connection? Yes Yes Yes Do you want to enable the public interface firewall? Yes No Yes Do you want to enable NAT forwarding on the management node? Yes No Yes Enable a BMC network that uses No the default provisioning template? Yes Yes Create a new network Yes Yes If creating a new BMC network.168. options include: Option included in the Custom installation? (Yes/No) v Create a new network v Public network v Provision Network v IPMI v IBM_Flex_System_x v IBM_System_x_M4 v IBM_iDataPlex_M4 v IBM_NeXtScale_M4 18 Installing IBM Platform HPC Version 4. Installer option comparison (continued) Options Default value Option included in the Quick installation? (Yes/No) Specify a public network interface.254 Yes Yes Specify the hardware profile used by your BMC network.255.2 . specify an IP address range for the BMC network 192. private.

0.org pcmdbpass Set the default root password for PASSW0RD compute nodes.1.10.168. 10. Specify a public network interface. 3.ntp.10. Preparing for PHPC quick installation Option Example values 1.0. Enable a BMC network that uses the default provisioning template? Yes Your values Chapter 4.168. You can choose to use the default example values for some or all of the options. use the following roadmap to prepare your values for each installation option.0. 10. No Yes Specify the NTP server.201-10. Installer option comparison (continued) Options Default value Option included in the Quick installation? (Yes/No) Specify the provisioning compute node IP address range. Option included in the Custom installation? (Yes/No) Quick installation roadmap Before you begin your quick installation.254 No Yes Set the IP addresses of the name 192.40. Select a mount point for the depot (/install) directory.Table 4. No Yes Do you want to export the /home Yes directory? No Yes Set the database administrator password. / 2.192. This is generated based on management node interface. No Yes No Yes pool. Table 5. Specify a provision network interface.0. eth1 5.50 servers.3-10. This is generated based on management node interface. eth0 4.10.1. Select the location that you CD/DVD drive want to install the operating system from.200 No Yes Do you want to provisioning compute nodes with the node discovery method? Yes No Yes Specify the node discovery IP address range.10. Performing an installation 19 .

you can suspend the installation (Ctrl-z). User accounts that are created after PHPC is installed are automatically synchronized across compute nodes when the compute nodes are updated. 192. If creating a new BMC network. options include: Create a new network Your values v Create a new network v Public network v Provision Network 7.255. Set a domain name for the public network? (Yes/No) Yes 12. Specify the hardware IBM_System_x_M4 profile used by your BMC network.168. you must boot into the base kernel.dns. specify a subnet for the BMC network. Set the public domain name. 255. options include: v IPMI v IBM_Flex_System_x v IBM_System_x_M4 v IBM_iDataPlex_M4 v IBM_NeXtScale_M4 10. change to another directory (cd ~). private. 20 Installing IBM Platform HPC Version 4. Set the provision network domain name. The Xen kernel is not supported. v Installing PHPC requires you to provide the OS media. Before you begin PHPC installation supports the Bash shell only. If creating a new BMC network. specify a subnet mask for the BMC network.2 .255. python mount_point/phpcinstaller). you can start the installation from another directory (for example: cd ~. v User accounts that are created before PHPC is installed are automatically synchronized across compute nodes during node provisioning. v Before you start the PHPC installation.0 8. ensure that no applications are actively using the drive (including any command shell). If you started the PHPC installation in the DVD directory. If you want to use the DVD drive.Table 5.zone 11. Preparing for PHPC quick installation (continued) 6. and then resume the installation (fg). Option Example values Select a BMC network. Alternately.com or FQDN Quick installation You can configure the management node by using the quick installation option. Hardware profile. public.0 9.1. v You must be a root user to install.

OS Distribution installation from an ISO image or mount point: Enter the path for the OS Distribution or mount point.x64. mount the ISO into a directory such as /mnt. OS Distribution installation from the DVD drive: Insert the correct OS DVD disk into the DVD drive. Chapter 4. Ensure that you have the correct permissions to add new users the /home mount point. 6. The following steps summarize the installation of PHPC on your management node: 1. issue the following command: # /mnt/phpc-installer 4. v Insert the PHPC DVD into the management node. you must add python in front of the command (python /media/PHPC-program_number/phpc-installer). make sure to insert the OS disk after you copy the PHPC core packages. 7. the DVD media is automatically mounted to /media/PHPC-program_number. Choose the Quick Installation option as your installation method. The operating system version that you select must be the same as the operating system version on the management node. Accept the license agreement and continue. License Agreement Management node pre-check Specify installation settings Installation Complete the following installation steps: Procedure 1. 5. To start the installer. Select the location that you want to install the operating system from. run: /media/PHPCprogram_number/phpc-installer. for example:/iso/rhel/6. For example: # mount -o loop phpc-4. 2. 3. Management node pre-checking automatically starts. mount to a directory such as /mnt.2. 3.v The /home mount point must have writable permission. The depot (/install) directory stores installation files for PHPC. Tip: Normally.x/x86_64/rhel-server-6.iso /mnt v If you install PHPC from DVD media.x-x86_64-dvd. Select a mount point for the depot (/install) directory. and version. Mount the PHPC installation media: v If you install PHPC from ISO file. About this task The installer completes pre-checking processes and prompts you to answer questions to complete the management node configuration. If the PHPC disk is already inserted. Performing an installation 21 . The PHPC management node checks for the required disk space. 4. If the DVD is mounted without execute permission. Choose one of the following installation methods: v Download the PHPC ISO to the management node. The disk is verified and added to the depot (/install) directory after you confirm the installation. Start the PHPC installer. architecture. 8.iso The PHPC management node verifies that the operating system is a supported distribution. 2.

13. 9. To get started with PHPC. You can find the installation log here: /opt/pcm/log/phpc-installer. press ‘99’ to reselect the settings or press '1' to begin the installation. 11.Note: If the OS distribution is found on more than one ISO image. b. – A subnet mask for the BMC network. Select a BMC network. run the following command: source /opt/pcm/bin/pcmenv. Enable a BMC network that uses the default provisioning template. After the PHPC installation is completed. 22 Installing IBM Platform HPC Version 4. use the first ISO image during the installation. Enter a domain name for the provisioning network. select: It is not connected to the public network. Options include: v Public network v Provision Network v Create a new network. If you create a new BMC network. If you choose to install from an ISO image or mount point. Select a network interface for the provisioning network. 10. What to do next After you complete the installation. you can add the next ISO image from the Web Portal. you must enter the ISO image or mount point path. you can access the Web Portal at http://hostname:8080 or http://IPaddress:8080. Select how the management node is connected to the public network. Results You successfully completed the PHPC installation. Enter a domain name for the public network. To change any of these settings. If the management node is not connected to the public network. 14. Select hardware profile for the BMC network.2 . Configuration is not required for new login sessions. If you choose to enable a BMC network. using your web browser. Set a domain name for the public network. 15. To configure PHPC environment variables.sh. 12. verify that your PHPC environment is setup correctly. A summary of your selected installation settings is displayed.log. Log in with the user account root and default password Cluster on the management node. specify the following options: – A subnet for the BMC network. you must specify the following options: a.

Set a domain name for the public network? (Yes/No) Yes 12.1 iv.10.com or FQDN 13. Select a mount point for the depot (/install) directory. You can choose to use the default example values for some or all of the options. Your values Create a new network and specify the Yes following options: i. Use the provision Network N/A Specify the hardware profile used by your BMC network.168. Subnet mask 255. Subnet 192.168.dns. Select one of the following options for creating your BMC networks: a.3-10.zone 11. Do you want to provisioning compute nodes with the node discovery method? (Yes/No) Yes Chapter 4.3192. use the following roadmap to prepare your values for each installation option. Specify the provisioning compute node IP 10. Specify a provision network interface. 3.10. Set the provision network domain name. Set the public domain name. eth1 5. This is generated based on management node interface.Custom installation roadmap Before you begin your custom installation.168. options include: IBM_System_x_M4 v IPMI v IBM_Flex_System_x v IBM_System_x_M4 v IBM_iDataPlex_M4 v IBM_NeXtScale_M4 10.254 b. Table 6. private. Enable a BMC network that uses the default provisioning template? Yes 8. / 2. eth0 4. Hardware profile.0 ii.0 iii. Do you want to enable NAT forwarding on the management node? (Yes/No) Yes 7. Specify a public network interface. 14. Use the public network N/A c. 9.1. public. Performing an installation 23 . Do you want to enable the public interface firewall (Yes/No) Yes 6. Preparing for PHPC custom installation Options Example values 1.255.0.1.0.200 address range.1. IP address range 192. Select the location that you want to install CD/DVD drive the operating system from.1.168. Gateway IP address 192.255.

192. pcmdbadm 20. If you started the PHPC installation in the DVD directory.org 18. The Xen kernel is not supported.0. Cluster Your values Custom installation You can configure the management node by using the custom installation option. v You must be a root user to install.1.ntp. and then resume the installation (fg).1. v User accounts that are created before PHPC is installed are automatically synchronized across compute nodes during node provisioning. Management node pre-check 3.0. change to another directory (cd ~). Preparing for PHPC custom installation (continued) Options Example values 15. About this task The installer completes pre-checking processes and prompts you to answer questions to complete the management node configuration. Set the IP addresses of the name servers. Installation Complete the following installation steps: 24 Installing IBM Platform HPC Version 4. v Installing PHPC requires you to provide the OS media. v Before you start the PHPC installation.Table 6.40.10. The following steps summarize the installation of PHPC on your management node: 1. Alternately. Set the default root password for compute nodes. User accounts that are created after PHPC is installed are automatically synchronized across compute nodes when the compute nodes are updated. you can start the installation from another directory (for example: cd ~.10.168. Specify installation settings 4.2 . Set the database administrator password.168. License Agreement 2. pool. you must boot into the base kernel. Do you want to export the /home directory? (Yes/No) Yes 19. This is generated based on management node interface. Before you begin Note: PHPC installation supports the Bash shell only.201-10. Ensure that you have the correct permissions to add new users the /home mount point. 10. you can suspend the installation (Ctrl-z). python mount_point/phpcinstaller).50 17. ensure that no applications are actively using the drive (including any command shell).192. Specify the NTP server.254 16. Specify the node discovery IP address range. If you want to use the DVD drive. v The /home mount point must have writable permission.

Select the location that you want to install the operating system from. Note: If the OS distribution is found on more than one ISO image. Start the PHPC installer. Accept the license agreement and continue. Enter the IP address range that is used for provisioning compute nodes. After the PHPC installation is completed. you can add the next ISO image from the Web Portal. mount to a directory such as /mnt. OS Distribution installation from the DVD drive: Insert the correct OS DVD disk into the DVD drive. mount the ISO into a directory such as /mnt.x/x86_64/rhel-server-6. Select a mount point for the depot (/install) directory. make sure to insert the OS disk after you copy the PHPC core packages. you must add python in front of the command (python /media/PHPC-program_number/phpc-installer). Choose one of the following installation methods: v Download the PHPC ISO to the management node. 5. 10. Choose whether to provision compute nodes automatically with the node discovery method. 12.x-x86_64-dvd. you must enter the ISO image or mount point path. If you choose to install from an ISO image or mount point. 8. The PHPC management node checks for the required disk space. Tip: Normally. 7. The depot (/install) directory stores installation files for PHPC. Performing an installation 25 . use the first ISO image during the installation. The operating system version that you select must be the same as the operating system version on the management node.x64. 11. architecture. If the PHPC disk is already inserted.Procedure 1. 3. The disk is verified and added to the depot (/install) directory after you confirm the installation. Management node pre-checking automatically starts.iso The PHPC management node verifies that the operating system is a supported distribution.2. issue the following command: # /mnt/phpc-installer 4. v Insert the PHPC DVD into the management node. Mount the PHPC installation media: v If you install PHPC from ISO file. Select the Custom Installation option. For example: # mount -o loop phpc-4. 9. and version. for example:/iso/rhel/6. the DVD media is automatically mounted to /media/PHPC-program_number. 2. run: /media/PHPCprogram_number/phpc-installer. If the DVD is mounted without execute permission. OS Distribution installation from an ISO image or mount point: Enter the path for the OS Distribution or mount point. Select a network interface for the provisioning network.iso /mnt v If you install PHPC from DVD media. Enter a node discovery IP address range to be used for provisioning compute nodes by node discovery. To start the installer. The node discovery IP address range is a temporary IP address range that is used to automatically provision nodes by using the Chapter 4. 6.

Enter the IP addresses of your name servers that are separated by commas 19. 20. press ‘99’ to reselect the settings or press '1' to begin the installation. select: It is not connected to the public network. you can enable the following settings: a. v Public network v Provision Network v Create a new network. 14. x3650 M4. specify the following options: – A subnet for the BMC network. optionally. Export the home directory on the management node and use it for all compute nodes. If the management node is not connected to the public network. verify that your PHPC environment is setup correctly. Enter a domain name for the provisioning network. 17. Enter the PHPC database administrator password. If you create a new BMC network. To change any of these settings. – A gateway IP address for the BMC network. Select a BMC network. 26 Installing IBM Platform HPC Version 4. 18. Set the NTP server. Select how the management node is connected to the public network. If you choose to enable a BMC network. Enter the root account password for all compute nodes. you must specify the following options: a. and x440 IBM_Flex_System_x IBM System x3550 M4.auto node discovery method. Enable a BMC network that uses the default provisioning template.2 . A summary of your selected installation settings is displayed. If your management node is connected to a public network. x3750 M4 IBM_System_x_M4 IBM System dx360 M4 IBM_iDataPlex_M4 IBM NeXtScale nx360 M4 IBM_NeXtScale_M4 15. Table 7. Enable NAT forwarding on the management node for all compute nodes. – An IP address range for the BMC network. 16. 23. b. Available hardware profiles based on hardware type Hardware Hardware profile Any IPMI-based hardware ® IPMI IBM Flex System x220. This range cannot overlap the range that is specified for the provisioning compute nodes. Enter a domain name for the public network. 22. Set a domain name for the public network. Enable PHPC specific rules for the management node firewall that is connected to the public interface. 13. b. x240. Specify a hardware profile for the BMC network. What to do next After you complete the installation. – A subnet mask for the BMC network. 21.

Chapter 4. Log in with the root user account and password on the management node. using your web browser. you can access the Web Portal at http://hostname:8080 or http://IPaddress:8080. Performing an installation 27 .To get started with PHPC.

2 .28 Installing IBM Platform HPC Version 4.

To complete the silent installation. Note: If the OS distribution is found on more than one ISO image. and the installation cannot begin until a valid path is used. use the first ISO image during the installation.conf is the location of your silent install file. only the last value is used # # by the silent installation. you must add python in front of the command (python /media/PHPC-program_number/phpc-installer).2. complete the following steps: 1. you can add the next ISO image from the Web Portal.example is located in the /docs directory in the Platform HPC ISO. # © Copyright IBM Corp. 1994. 3. # IBM Platform HPC 4. mount to a directory such as /mnt. Before you complete the installation using silent mode.x64. # ******************************************************************** # # NOTE: For any duplicated options. v Ensure that you have the correct permissions to add new users the /home mount point. To start the installer. complete the following actions: v Install the operating system on the management node. The installation path cannot include special characters such as a colon (:).conf. Tip: Normally. Performing a silent installation Silent installation installs IBM Platform HPC software using a silent response file. Response file for silent installation Response file for IBM Platform HPC silent installation. Mount the PHPC installation media: v If you install PHPC from ISO file. The silent response file phpc-autoinstall. mount the ISO into a directory such as /mnt. run: /media/PHPCprogram_number/phpc-installer.Chapter 5.iso /mnt v If you install PHPC from DVD media.conf where mnt is your mount point and path_to_phpc-autoinstall. exclamation point (!) or space. For example: # mount -o loop phpc-4. You can specify all of your installation options in the silent installation file before installation. 2014 29 .2 Silent Installation Response File # The silent installation response file includes all of the options that can # be set during a Platform HPC silent installation. Prepare the response file with installation options. If the DVD is mounted without execute permission. the DVD media is automatically mounted to /media/PHPC-program_number. After the PHPC installation is completed. 2. Run the silent installation: mnt/phpc-installer -f path_to_phpc-autoinstall. Usage notes v A valid installation path must be used.

If you have an existing /install mount point. # # 4. An accepted value for this option is a valid NIC name that exists on the management node. This is a mandatory option. Values must use alphanumeric characters and cannot use quotations ("). The value ’lo’ is not supported. provisioning_network_interface = eth0 # # public_network_interface # # The public_network_interface option sets a network device on the Platform HPC # management node that is used for accessing networks outside of the cluster. # # 3.2 . # # 2.dns. no public network interface is defined. The domain must be a fully qualified domain name. If you specify depot_path = /install or depot_path = /. # # ******************************************************************** # [General] # # depot_path # # # The depot_path option sets the path of the Platform HPC depot (/install) directory. #public_network_interface = eth1 [Media] # # os_path # # The os_path option specifies the disc. the installer places all Platform # HPC installation contents in the /usr/local/pcm/install directory and creates # a symbol link named /install that points to the /usr/local/pcm/install directory.zone # # # # # # # # provisioning_network_interface The provisioning_network_interface option sets one network device on the Platform HPC management node to be used for provisioning compute nodes. This is a mandatory option. If you specify depot_path = /usr/local/pcm/. depot_path = / # # # # # # private_cluster_domain The private_cluster_domain option sets the provisioning network’s domain name for the cluster. # The value cannot be ’lo’ and cannot include quotations ("). private_cluster_domain = private. or path of the first OS distribution 30 Installing IBM Platform HPC Version 4. ISO. The Platform HPC installation requires a minimum available disk space of 40 GB. # #Usage notes: # # 1.# NOTE: Configuration options cannot start with a space or tab. The value cannot be # the same as the value specified for the provisioning_network_interface option. The value # must be a valid NIC name that exists on the management node. # If this option is not defined. the installer places # all Platform HPC installation content into the /install directory. the installation # program places all installation contents into the /install directory regardless # of the depot_path value. by default.

0.10. for example: /root/rhel-server-6. If this option is not defined.10.iso [Advanced] # NOTE: By default.0.used to # install the # # The os_path # .10.iso to a directory where an ISO is mounted. If this option is not defined. If this option is not defined.3-10.full path # . If this option is not defined. by default.10. # # # # # # # excluded_kits The excluded_kits option lists specific kits that do not get installed. the value is automatically based on the provision network.full path # Platform HPC node. #enable_firewall = yes # # # # # # # enable_nat_forward The enable_nat_forward option enables NAT forwarding on the management node for all compute nodes.0.254 # # # # # # # enable_firewall The enable_firewall option enables Platform HPC specific rules for the management node firewall to the public interface. for example: /mnt/basekit os_path = /root/rhel-server-<version>-x86_64-dvd. The os_path is a mandatory option. This option is only available if the public_network_interface is set to yes. #static_ip_range = 10.201-10. advanced options use a default value if no value is specified. by default. the default value is set to none. This is a comma-separated list. #excluded_kits = kit1. the value is set to yes. all kits are installed. The kit name should be same with the name defined in the kit configuration file.4-x86_64-dvd. If this option is not defined. by default. This IP address range cannot overlap with the IP range used for provisioning compute nodes as specified by the static_ip_range option. for example: /dev/cdrom to an ISO file. #enable_nat_forward = yes # # enable_bmcfsp # Chapter 5. This option is only available if the enable_firewall option is set to yes.200 # # # # # # # # discovery_ip_range The discovery_ip_range option sets the IP address range that is used for provisioning compute nodes by node discovery. by default. option must use one of the following options: to CD-ROM device.kit2 # # # # # # static_ip_range The static_ip_range options sets the IP address range used for provisioning compute nodes. Performing a silent installation 31 .full path # . the value is set to yes. #discovery_ip_range = 10.0. You can set the discovery_ip_range value to ’none’ if you do not want to use node discovery.

# # bmcfsp_hwprofile options: # For x86-based systems.168. This option is required if enable_bmcfsp = new_network #bmcfsp_iprange = 192. x3650 M4.1. # provision option: Creates a BMC or FSP network that uses the provision network.# The enable_bmcfsp option enables a BMC or FSP network with the default provisioning template. If this option is not defined. x3750 M4 # IBM_Flex_System_x: IBM System x220.This is a # mandatory option.168. #bmcfsp_gateway = 192. # Options include: new_network. x440 # IBM_iDataPlex_M4: IBM System dx360 M4 # IPMI: Any IPMI-based hardware # 32 Installing IBM Platform HPC Version 4. This option is # required if enable_bmcfsp = new_network.1.3-192. Otherwise. This option is required if enable_bmcfsp = new_network.168. #bmcfsp_subnet = 192.1. the BMC or FSP network set up fails.1 # # bmcfsp_iprange # # Specify the IP address range for the BMC or FSP network. # This option indicates which network is associated with BMC or FSP network.0 # # bmcfsp_gateway # # Specify the gateway IP address for the BMC or FSP network. This value must be different than the value used by # the public and provision networks.1. #enable_bmcfsp = new_network # # bmcfsp_subnet # # Specify the subnet for the BMC or FSP network.254 # # bmcfsp_hwprofile # # Specify a hardware profile to associate with the BMC or FSP network.0 # # bmcfsp_subnet_mask # # Specify the subnet mask for the BMC netwrok. #bmcfsp_subnet_mask = 255.255. a BMC or FSP network is # not enabled.255.168. public. provision # new_network option: Creates a new BMC or FSP network by specifyingi the following options for the # the new network # [bmcfsp_subnet] # [bmcfsp_subnet_mask] # [bmcfsp_gateway] # [bmcfsp_iprange] # will be applied to create a new network # public option: Creates a BMC or FSP network that uses the public network. by default. This option is required if enable_bmcfsp = new_network. the following are supported hardware profile options: # # IBM_System_x_M4: IBM System x3550 M4.This option is available if enable_bmcfsp = new_network.2 . x240.

the value is set to none. or a comma-separated list of several user account list. by default. If this option is not defined. 0-9 or underscore (_). # by default.168. by default. The first user account name in the list is the primary LSF administrator and it cannot be the root user account.# For POWER systems. The export home directory is used on all all compute nodes. this value is set to Cluster. this value is set to pool.1. #cluster_admin = phpcadmin Chapter 5. this value is set to pcmdbadm.user_name2. p460 #bmcfsp_hwprofile = IBM_System_x_M4 # # # # # # nameservers The nameservers option lists the IP addresses of your external name servers using a comma-separated list. this value is set to yes. by default. # If this option is not defined. the following are supported hardware profile options: # IBM_Flex_System_p: IBM System p260. The cluster name must be a string containing any of the following characters: a-z. This can be a single user account name.If this option is not define.1.168. #cluster_name = phpc_cluster # # # # # # # # cluster_admin The cluster_admin specifies the Platform HPC workload manager administrator.If this option is not defined. #nameservers = 192. #ntp_server = pool. The string length cannot exceed 39 characters. If this option is not defined. this value is set to phpc_cluster.40. by default. A-Z.. by default. #enable_export_home = yes # # db_admin_password # # The db_admin_password option sets the Platform HPC database administrator password..org # # # # # # enable_export_home The enable_export_home option specifies if the /home mount point exports to the management node.htp. Performing a silent installation 33 . #compute_root_password = Cluster # # # # # # # cluster_name The cluster_name option sets the cluster name for the Platform HPC workload manager. # If this option is not defined. For example: cluster_admin=user_name1.50 # # ntp_server # # The ntp_server option sets the NTP server.192. #db_admin_password = pcmdbadm # # compute_root_password # # The compute_root_password option sets the root account password for all compute nodes. this value is set to phpcadmin. by default. If this option is not defined.org.ntp.

2 .34 Installing IBM Platform HPC Version 4.

A user account is not able to manage cluster resources but can manage jobs. © Copyright IBM Corp. c. Log in to the Web Portal. PHPC creates a default administrative account where the username and password is phpcadmin and phpcadmin. a. Refer to the Release Notes for a list of supported web browsers. res (pid 27663) is running. you can also navigate to http://mgtnode-hostname:8080. To verify that your installation is working correctly. where mgtnode-hostname is the real management node hostname. 2014 35 .. # service phpc status Show status of the LSF subsystem lim (pid 31774) is running. sbatchd (pid 27667) is running. the Resource Dashboard is displayed in the Web Portal. d. Go to http://mgtnode-IP:8080.... If you are connected to a public network. Check that the Platform HPC services are running. log in to the management node as a root user and complete the following tasks: 1. Source PHPC environment variables. # service postgresql status (pid 13269) is running. SERVICE WEBGUI STATUS STARTED WSM_PID 16550 PORT 8080 SERVICE jobdt plc plc_group2 purger vdatam STATUS STARTED STARTED STARTED STARTED STARTED WSM_PID 5836 5877 5917 5962 6018 HOST_NAME hjc-ip200 hjc-ip200 hjc-ip200 hjc-ip200 hjc-ip200 HOST_NAME hjc-ip200 4. After you log in. Log in as an administrator or a user. 3. Open a supported web browser. Verifying the installation Ensure that you have successfully installed PHPC. Note: You can find the installation log file phpc-installer.sh 2. 1994. This log file includes details and results about your PHPC installation. b.log in the /opt/pcm/log directory.. By default.Chapter 6. This default phpcadmin administrator account has all administrative privileges. Check that the PostgreSQL database server is running.. # ... /opt/pcm/bin/pcmenv. where mgtnode-IP is the real management node IP address. An administrator has administrative privileges that include managing cluster resources.

2 .36 Installing IBM Platform HPC Version 4.

as an administrator you can get started with managing your clusters. 2014 37 . The following tasks can be completed to get started with Platform HPC: v Enabling LDAP support for user authentication v Provision your nodes by adding the nodes to your cluster v Modify your provisioning template settings – Manage image profiles – Manage network profiles v Set up the HTTPS connection v Submit jobs v Create resource reports v Create application templates For more information about IBM Platform HPC.com/support/knowledgecenter/ SSDV85_4. For the latest release information about Platform HPC 4. © Copyright IBM Corp.Chapter 7.ibm.0. 1994. see the Administering IBM Platform HPC guide.2. see Platform HPC on IBM Knowledge Center at http://www. Taking the first steps after installation After your installation is complete.2.

38 Installing IBM Platform HPC Version 4.2 .

and any warnings and errors that occurred during the installation. To resolve this issue. the services that are required to run the Web Portal start automatically. If rpcbind is stopped. To help troubleshoot your installation. 1994. run the installer in DEBUG mode when troubleshooting the installation. To run the installer in debug mode. 2014 39 . Do not use it for installing PHPC using silent install. Platform HPC is already installed. Common installation issues include the following issues: v The Platform HPC installer fails with the error message “Cannot reinstall Platform HPC. set the PCM_INSTALLER_DEBUG environment variable. The DEBUG mode also generates extra log messages that can be used to trace the installer's execution. When running in DEBUG mode. However.rc. In most cases. if the Web Portal goes down. you need to manually change this setting. you must first uninstall the installed product. you must restart it and run the S03_base_nfs. v Setting up shared NFS export fails during installation. # service rpcbind start # cd /opt/pcm/rc. – Configure your web browser. Note: Only use the PCM_INSTALLER_DEBUG environment variable. # service rpcbind status 2. v During management node pre-checking.py script. Ensure that all Platform HPC requirements are met and rerun the installer. Set the PCM_INSTALLER_DEBUG environment variable to run the installer in DEBUG mode: # PCM_INSTALLER_DEBUG=1 hpc-ISO-mount/phpc-installer where hpc-ISO-mount is the mount point. one of the checks fails.py v Cannot log in to the Web Portal.rc.Chapter 8.log file that is found in the /opt/pcm/log directory. Your web browser must be configured to accept first-party and third-party cookies. – Restart the Web Portal. the installation progress is logged in a temporary directory that is found here: /tmp/phpc-installer. In some cases. Note: During the installation. complete the following steps: 1. your browser default settings can block these cookies.” To install a new Platform HPC product. the installer does not clean up all the files when an error occurs. you can view the phpc-installer. Troubleshooting installation problems Troubleshooting problems that occurred during the IBM Platform HPC installation. or view the Resource Dashboard in the Web Portal. To view detailed error messages.pcm. you © Copyright IBM Corp. to troubleshoot a PHPC installation using the interactive installer.d/ # pcmconfig -i . In this case. For more information about Platform HPC see the Release Notes®. Check the rpcbind status. This file logs the installation steps./S03_base_nfs.

7.can restart services and daemons manually. Start Firefox and enter about:plugins into the Firefox address field.com/technetwork/java/javase/downloads/index. issue the following command:# pmcadmin stop .adobe. Find dom. Restart Firefox. run the command: rpm -ivh jre-7u2-linux-x64. 2. you must install the JRE plug-in manually. Java(TM) Plug-in 1. Check that the Flash plug-in is enabled.rpm) is recommended. you must have the necessary plug-ins installed.7. 5.html. 4.so 9. About this task If you are using Firefox as your browser. In the package folder. Check that the JRE plug-in was installed correctly.2 . Exit Firefox. From the command line. If it is set to false. 40 Installing IBM Platform HPC Version 4. Enter about:config into the Firefox address field.oracle.0_02 is displayed in the list.0_02/lib/amd64/libnpjp2.plugins.com/flashplayer).oracle. double-click it to enable. 7.rpm 8. To run Java™ applets within the browser.ipc. To install the Flash and JRE plug-ins. 6. Check that the Flash plug-in is installed. When the installation is finished.com/javase/7/docs/webnotes/install/linux/linux-plugininstall.html). go to http://docs. Enter about:plugins into the Firefox address field. 3.enabled in the list and ensure that it has a value of true. Install the appropriate Adobe Flash Player plug-in from the Adobe website (http://get. Download the appropriate JRE plug-in installer from the Oracle website (http://www. complete the following steps: Procedure 1. Shockwave Flash appears in the list. pmcadmin start Configuring your browser To properly configure your browser. The 64-bit rpm installer (jre-7u2-linux-x64. enter the following commands: cd /usr/lib64/mozilla/plugins ln -s /usr/java/jre1. you are required to have the Flash and JRE plug-ins installed. For more information about installing the JRE plug-in manually.

Preparing high availability Set up the secondary management node with an operating system and Platform HPC installation. To setup a high availability (HA) environment in Platform HPC. Complete the high availability enablement After high availability is enabled setup up the compute nodes. Refer to “Installing and configuring the operating system on the management node” on page 13.Chapter 9. © Copyright IBM Corp. Verify Platform HPC high availability Ensure that Platform HPC high availability is running correctly on the primary and secondary management nodes. The secondary management node must use the same operating system and version as used on the primary management node. Enable a Platform HPC high availability environment Set up Platform HPC high availability on the primary and secondary management nodes. After the secondary management node is set up. Before you begin Ensure that all high availability requirements are met and a shared file system is created on a shared storage server. High availability environment roadmap Actions Description Ensure that the high availability requirements are met Requirements for setting up a shared storage device and a secondary management node must be met. Preparing high availability Preparing an IBM Platform HPC high availability environment. 2014 41 . About this task To prepare a high availability environment. Table 8. 1994. Procedure 1. Troubleshooting enablement problems Troubleshooting problems that occurred during a Platform HPC high availability environment setup. Both management nodes must use the same network and must be connected to the same network interface. Setting up a high availability environment Setup an IBM Platform HPC high availability environment. set the secondary management node with the same operating system and PHPC version as on the primary management node. complete the following steps. Install the operating system on the secondary node. the necessary SSH connections and configuration must be made between the primary management node and the secondary management node.

2 . for example: ln –s /usr/share/zoneinfo/US/Eastern /etc/localtime d. b. “Performing an installation. set the correct time zone in the /etc/sysconfig/clock file. For example: For RHEL: ZONE=”US/Eastern” For SLES: TIMEZONE=”America/New_York” c. Backup and configure a passwordless SSH connection between the primary management node and the secondary node. a.2. Ensure that the time and time zone is the same on the primary and secondary management nodes. Set the date on both management nodes. 5. “Performing a silent installation. If you use the silent installation to install PHPC. Verify that the management nodes can access the shared file systems. If the time zone is incorrect. Set the local time in the /etc/localtime file. you can use the same response file for both management nodes. You can complete the installation using the installer or the silent installation. run the cat /etc/sysconfig/clock command. Installation options for the primary management node are found in the installation log file (/opt/pcm/log/phpc-installer.168. lsdef -t site -o clustersite -i timezone If the system time zones are different. If the management nodes already have PHPC installed. You must use the same PHPC ISO file as you used for the management node. Run the command below on the primary management node. Ensure that the failover node name can be resolved to the secondary management node provision IP address. For example: #echo "192. Install PHPC on the secondary node. date -s current_time e. where nfs-server-ip is the IP address of the NFS server that connects to the provision network. To verify the current time zone.” on page 17. make sure to use the same installation options as the primary management node. To update the time zone.1. Issue the following command on both management nodes. update the system time zone on the secondary node. Add the secondary management node entry to the /etc/hosts file on the primary management node.4 backupmn" >> /etc/hosts 6. run the following command: chdef -t site -o clustersite timezone=US/Eastern 3. echo "secondary-node-provision-ip secondary-node-name" >> /etc/hosts #ping secondary-node-name where secondary-node-provision-ip is the provision IP address of the secondary node and secondary-node-name is the name of the secondary node. update the time zone. Refer to Chapter 4. refer to the information found in the /usr/share/zoneinfo directory.log) on the primary management node. Refer to Chapter 5. The installer includes an interactive display where you can specify your installation options. issue the showmount -e nfs-server-ip command. 42 Installing IBM Platform HPC Version 4. run the following command on both management node to get the system time zone.” on page 29 4. To determine the correct time zone.

ssh secondary-node-name cp –rf /root/. These steps are used for provisioned compute nodes that you do not want to reprovision.ssh where secondary-node-provision-ip is the provision IP address of the secondary node and secondary-node-name is the name of the secondary node. Chapter 9. The tool defines and sets up a high availability environment between the management nodes using a predefined high availability definition file. Procedure 1. Note: The high availability management tool (pcmhatool) supports Bash shell only. # xdsh __Managed ’service lsf stop’ b. including: virtual name.ssh/id_rsa. Set up a high availability environment. The high availability definition file example ha. # updatenode __Managed ’mountnfs del’ # xdsh __Managed 'umount /home' # xdsh __Managed 'umount /shared' Enable a high availability environment Enable an IBM Platform HPC high availability environment. Issue the following command on the primary management node. pcmhatool config -i ha-definition-file -s secondary-management-node where ha-definition-file is the high availability definition file that you created in step 1. Ensure that the shared storage server is always available. 7.ssh/authorized_keys scp –r /root/.PCMHA # Configure passwordless SSH between the management node and the secondary node. About this task You can set up the high availability environment using the high availability management tool (pcmhatool).pub > /root/. virtual IP address. and secondary-management-node is the name of the secondary management node. Unmount and remove the /home and /share mount points on the compute nodes.# Back up the SSH key on the secondary node.example is in the /opt/pcm/share/examples/HA directory.ssh /root/.info. Shutdown the LSF services on the compute nodes. 2. Setting up high availability 43 .ssh/* secondary-node-name:/root/. cat /root/. Define a high availability definition file according to your high availability settings. a. Before you begin Ensure that the secondary management node is installed and setup correctly. and shared storage. Setup can take several minutes to synchronize data to shared storage. Refer to “High availability definition file” on page 67. Ensure that SSH connections are configured and network settings are correct between the primary management node and the secondary management node.ssh. Prepare the compute nodes.

Configure IPMI as a fencing device In a high availability cluster that has only two management nodes. After a few minutes. The fencing process locates the malfunctioning node and disables it. the pcmadmin command cannot be used to restart the PERF loader. Before you begin This fencing method requires both management nodes to be controlled remotely using IPMI. 4. You can find the enablement log file (pcmhatool. rerun the pcmhatool command and specify the same options. During a high availability enablement. you can use the example fencing script (fencing_ipmi.sh) that is found in the /opt/pcm/share/examples/HA directory. use the following commands to restart the PERF loader: pcm-ha-support pcm-ha-support pcm-ha-support pcm-ha-support pcm-ha-support start start start start start --service --service --service --service --service PLC PLC2 JOBDT PTC PURGER What to do next After the high availability enablement is complete. you can set up and configure additional options. Ensure that you modify fencing_ipmi. You can also set up email notification when a failover is triggered. they switch to the active management node. For example. Create an executable fencing script on the shared file system. In a high availability. Use remote hardware control to configure fencing on an IPMI device. Run the following commands to create the script on a shared file system. such as configuring an IPMI device as a fencing device to protect your high availability cluster from malfunctioning nodes and services. verify that the Platform HPC high availability environment is set up correctly. Running this command again cleans up the incomplete environment and starts the high availability enablement again.log) in the /opt/pcm/log directory.Usage notes 1. If you enable high availability. it is important to configure fencing on an IPMI device. This log file includes details and results about the high availability environment setup. Completing the high availability enablement After high availability is enabled. Procedure 1.sh to your real environment settings. 44 Installing IBM Platform HPC Version 4. If the management node crashes during the high availability environment setup. you must create the corresponding fencing script accordingly. some of the services start on the standby management node instead of the active management node. If your management nodes are on a power system or using a different remote power control method. 3.2 . 2. Fencing is the process of isolating a node or protecting shared resources from a malfunctioning node within a high availability environment.

This field is only required by some servers. click General Settings. 2. Procedure 1. for a triggered failover. Create an executable script on the shared file system. Edit the HA controller service agent configuration file (ha_wsm. Expand the Mail Settings heading. Enter the user account password. you can use an executable script that sends an email when a failover is triggered. In the[__Failover__] section.sh /install/failover 2. Restart the PCMHA service agent. an SMTP server must already be installed and configured. Log in to the Web Portal as the system administrator. Enter the mail server (SMTP) host. Restart the high availability environment. a. Run the following commands to create the script on a shared file system.sh to your real environment settings. Edit the high availability controller configuration file (ha_wsm. mkdir -p /install/failover cp /opt/pcm/share/examples/HA/send_mail. For example.sh /install/failover 2. Chapter 9. In the [__Failover__] section.cfg) on the management node in the /opt/pcm/etc/failover directory. d. Enter the user account.” Procedure 1. b. Enter the mail server (SMTP) port.mkdir -p /install/failover cp /opt/pcm/share/examples/HA/fencing_ipmi. set the failover_action parameter to the absolute path of your custom script. An example send email script (send_mail. pcm-ha-support start --service PCMHA Setting up SMTP mail settings Specify SMTP mail settings in IBM Platform HPC.sh 3. For example: failover_action=/install/failover/send_mail. Refer to “Setting up SMTP mail settings. Before you begin To send email from Platform HPC. you must configure your mail parameters. such as an email notification. For example: fencing_action =/install/failover/fencing_ipmi. In the System & Settings tab. Ensure that you modify send_mail. Setting up high availability 45 .sh) is in the /opt/pcm/share/examples/HA directory.cfg) in the /opt/pcm/etc/failover directory on the active management node. c. 3. Before you begin Note: Before you can send email for a triggered failover.sh 3. pcm-ha-support start --service PCMHA Create a failover notification Create a notification. This field is only required by some servers. set the value for fencing_action parameter to the absolute path of your custom script.

2.168. # pcmhatool info Configuring status: OK ================================================================ HA group members: master.. Open a supported web browser. where mgtnode-virtual-hostname is the virtual management node hostname.200:/export/home 4. Procedure 1.200:/export/data Shared home directory on: 172. Check that Platform HPC services are running. If you are connected to a public network.2 .20.20... you can also navigate to http://mgtnode-virtual-hostname:8080. for example: # service phpc status Show status of the LSF subsystem lim (pid 29003) is running.log) in the /opt/pcm/log directory. Click Apply.. Check that Platform HPC high availability is configured. Platform HPC uses the configured SMTP server to send email.7. b. Go to http://mgtnode-virtual-IP:8080. Log on to the management node as a root user. /opt/pcm/bin/pcmenv.20. sbatchd (pid 29008) is running. Verifying a high availability environment Verify an IBM Platform HPC high availability environment.4. Refer to the Release Notes for a list of supported web browsers. However. Before you begin You can find the enablement log file (pcmhatool.7.0. Source Platform HPC environment variables..sh 3. This log file includes details and results about your PHPC enablement.100 Virtual IP for <eth1:0>: 172. The account from which the mail is sent is the user email account.. Log in to the Web Portal. if the user email account is not specified then the email account uses the management node name as the email address. SERVICE STATE ALLOC CONSUMER RGROUP RESOURCE PLC STARTED 32 /Manage* Manag* master PTC STARTED 34 /Manage* Manag* master PURGER STARTED 35 /Manage* Manag* master WEBGUI STARTED 31 /Manage* Manag* master JOBDT STARTED 36 /Manage* Manag* master PLC2 STARTED 33 /Manage* Manag* master PCMHA STARTED 28 /Manage* Manag* master PCMDB STARTED 29 /Manage* Manag* master XCAT STARTED 30 /Manage* Manag* master SLOTS SEQ_NO INST_STATE ACTI 1 1 RUN 9 1 1 RUN 8 1 1 RUN 7 1 1 RUN 4 1 1 RUN 6 1 1 RUN 5 1 1 RUN 1 1 1 RUN 2 1 1 RUN 3 5. res (pid 29006) is running. Results SMTP server settings are configured. # . where mgtnode-virtual-IP is the management node virtual IP address. a. failover Virtual node name: virtualmn Virtual IP for <eth0:0>: 192.7. 46 Installing IBM Platform HPC Version 4. All services must be in state STARTED.100 Shared work directory on: 172.

Platform HPC high availability environment uses the provision network for heartbeat communication. v During the pre-checking. the correct virtual name. If the service daemon can be started manually. Chapter 9. Under the Cluster Health option. conflicts with existing managed node. ensure that the high availability definition file does not have any formatting errors. To resolve this issue. v Syncing data to shared directory fails. check the service daemon log for more detail on how to resolve this problem. ensure that the xCAT daemon is running by issuing the command tabdump site. An administrator has administrative privileges that include managing cluster resources. v Services are running on the standby management node after an automatic failover occurs due to a provision network failure. and any warnings and errors that occurred during the high availability enablement. and fencing to stop working.log. To help troubleshoot your high availability enablement. the Resource Dashboard is displayed in the Web Portal. If it continually fails. you can view the log file is found here /opt/pcm/log/pcmhatool. log in the management node with a new session. To resolve this issue. Ensure that the external NFS server is available and check the network connection to the NFS server is available. v When you check the Platform HPC service status. go to https://mgtnode-virtual-IP:8443 or https://mgtnode-virtual-hostname:8443 to log in to the web portal. issue the following command: pcm-ha-support start --service service_name where service_name is the name of the service that is experiencing the problem. To resolve this issue. try to reboot it. d. This file logs the high availability enablement steps. Common high availability enablement issues include the following issues: v When you run a command on the management node. Also. and that the IP address does not conflict with an existing node. Troubleshooting a high availability environment enablement Troubleshooting an IBM Platform HPC high availability environment. one of the checks fails. c. To resolve this issue. issue the following command: pcm-ha-support stop --service all v Parsing high availability settings fails.If HTTPS is enabled. the service agent is set to ERROR. After you log in. both management nodes are listed. the command stops responding. Log in as an administrator or user. A user account is not able to manage cluster resources but can manage jobs. restart the service agent again. If you cannot log in the management node. ensure that all Platform HPC high availability requirements are met and rerun the high availability enablement tool. To resolve this issue. Setting up high availability 47 . the service agent attempts to restart it several times. one of the service agent statuses is set to ERROR. When the monitored service daemon is down. The provision network failure causes the management nodes to lose the communication. stop the service agents manually.

#pcmhatool check 48 Installing IBM Platform HPC Version 4. rerun the tool by setting PCMHA_NO_CLEAN environment variable. ensure that the network connection to the external shared storage is stable during the high availability enablement. or view the Resource Dashboard in the Web Portal. If a timeout occurs during data synchronization. If the issue persists. Wait a few minutes and try again. All Platform HPC services are started a few minutes after the high availability enablement.2 . #PCMHA_NO_CLEAN=1 pcmhatool config –i ha-definition-file –s secondary-management-node where ha-definition-file is the high availability definition file and secondary-management-node is the name of the secondary management node.To resolve this issue. run the high availability diagnostic tool to check the running status. This environment variable ensures that existing data on the NFS server is unchanged. v Cannot log in to the Web Portal.

refer to the Release notes for Platform HPC 4.1 to 4. 1994. Additionally. Make sure that the external storage is larger than the size of your backup files. External storage device Obtain an external storage to store the necessary backup files.1.1. The upgrade procedure ensures that the necessary files are backed up and necessary files are restored. Refer to “PHPC requirements” on page 9.2 with OS reinstall If any errors occur during the upgrade process.2 without OS reinstall v Upgrading from Platform HPC 4.1. The following upgrade paths are available: v Upgrading from Platform HPC 4.2 Upgrade from Platform HPC Version 4. Table 9. Software requirements Ensure that you meet the software requirements for Platform HPC. you can upgrade your operating system to the latest supported version. ensure you meet the following criteria before proceeding with the upgrade. 2014 49 .Chapter 10. Review the upgrade checklist and upgrade roadmap before you begin the upgrade process. Upgrading checklist Use the following checklist to review the necessary requirements before upgrading. Upgrading to Platform HPC Version 4.1 to Version 4.2 guide. Requirements Description Hardware requirements Ensure that you meet the hardware requirements for Platform HPC. Refer to “PHPC requirements” on page 9.2 ISO Get a copy of Platform HPC 4.1. Upgrading IBM Platform HPC Upgrade IBM Platform HPC from Version 4.2 (Optional) Obtain a copy of the latest supported version operating system Optionally.1 to Version 4.2. In order to upgrade to the newest release of IBM Platform HPC.1.1.1 to 4. © Copyright IBM Corp. Obtain a copy of the Platform HPC 4.1.1. you can upgrade the product entitlement files for Platform Application Center or LSF. For a list of all supported upgrade procedures. you can roll back to an earlier version of Platform HPC. Upgrade planning Upgrading IBM Platform HPC involves several steps that you must complete in the appropriate sequence.2.

2 without reinstalling the operating system.1 backup. you must complete the following steps: 50 Installing IBM Platform HPC Version 4. Before you begin To prepare for your upgrade. Completing the upgrade Ensure that data is restored and services are restarted. Preparing to upgrade Before you can upgrade to the newest release of Platform HPC you must complete specific tasks. 6. 3. Upgrading checklist Ensure that you meet all of the requirements before upgrading Platform HPC. Creating a Platform HPC 4. This backup is used to restore your existing settings to the newer version of Platform HPC. Preparing to upgrade Before upgrading your IBM Platform HPC installation.1.1.1 settings and database. Upgrading to Platform HPC 4.2 ISO file. 7. there are some steps you should follow to ensure your upgrade is successful. Verifying the upgrade Ensure that PHPC is successfully upgraded.1. 2.2 . Table 10.1.1.2 without OS reinstall v Upgrading to Platform HPC 4.1 backup Create a backup of your current Platform HPC 4. For additional requirements refer to “Upgrading checklist” on page 49. Note that if you are upgrading Platform HPC to Version 4.1. the PMPI kit version is not upgraded.2 without OS reinstall Upgrade your existing installation of IBM Platform HPC to the most recent version without reinstalling the operating system on the management node. (Optional) Applying fixes After you upgrade PHPC. make sure that you have the RHEL ISO file. v If you are upgrading the operating system. you can check if there are any fixes available though the IBM Fix Central. and that you have a corresponding OS distribution created. Perform the Platform HPC upgrade Perform the upgrade using your chosen path: v Upgrading to Platform HPC 4. 4.2 with OS reinstall 5. v The Platform HPC 4. ensure that you have the following items: v You must have an external backup to store the contents of your 4.Upgrading roadmap Overview of the upgrade procedure. About this task Before you upgrade to the next release of Platform HPC. Upgrading Platform HPC Actions Description 1.

Procedure 1. Prepare an external storage. Backing up Platform HPC Create a backup of your current Platform HPC installation that includes a backup of the database and settings before you upgrade to a newer version of Platform HPC. export PATH=${PATH}:/opt/pcm/libexec/ 4. Note: The backup procedure does not back up any custom configurations. and can manually be re-created after the upgrade is completed.x64. After the upgrade procedure is completed. the following custom configurations can be manually re-created: v Customization to the PERF loader. For RHEL: rpm -Uvh /mnt/packages/repos/kit-phpc-4. Upgrade the pcm-upgrade-tool package. On the external storage.2. 6. including internal data collection and the purger configuration files v Customization to the Web Portal Help menu navigation v Addition of custom metrics v Alert polices Chapter 10. Upgrading 51 . /opt/xcat/bin/updatenode mn-host-name -F where mn-host-name is the name of your management node.2-rhels-6-x86_64/pcm-upgrade-tool-*. b.rpm 3. If you created any new users after Platform HPC was installed. 5. Ensure that the external storage has enough space for the backup files. mkdir /external-storage-mnt/config-backup where the external-storage-mnt is the backup location on your external storage. Set up the upgrade environment. Determine which custom metrics you are using.iso /mnt 2. a. run the following commands: # du -sh /var/lib/pgsql/data # du -sh /install/ Note: It is recommended that the size of your external storage is greater than the combined size of the database and the /install directory. create a directory for the database backup.rpm For SLES: rpm -Uvh /mnt/packages/repos/kit-phpc-4. mkdir /external-storage-mnt/db-backup where the external-storage-mnt is the backup location on your external storage. you must include these new users in your backup. To check how much space you require for the back. c.2-sles-11-x86_64/pcm-upgrade-tool-*. The custom metrics are lost in the upgrade process. Create a directory for the configuration file backup. if any. Mount the Platform HPC installation media: mount -o loop phpc-4.

2 .d/xcatd stop b.conf. Restore database data from a previous backup. # service xcatd start e. pcm-upgrade-tool. pcm-upgrade-tool. pcm-upgrade-tool. If you created custom metrics in Platform HPC 4.v LDAP packages and configurations Before you begin Platform HPC does not back up or restore LSF configuration files or data.x64.1.1. Upgrade the database schema. Start the xCAT daemon. c. Upgrade Platform HPC from 4.py upgrade --packages -p /root/phpc-4. 2.py services --stop 2.py backup --files -d /external-storage-mnt/config-backup/ Performing the Platform HPC upgrade Perform the upgrade without reinstalling the operating system and restore your settings. 52 Installing IBM Platform HPC Version 4. Restore settings and database data.py upgrade --schema b. Create a database backup on the external storage. Start the HTTP daemon (HTTPd). Procedure 1.iso f. complete the following steps: a. The backup includes database files and the backup configuration file pcm. Upgrade Platform HPC. complete the following steps: a. Stop Platform HPC services: pcm-upgrade-tool.1. For RHEL: # service httpd start For SLES: # service apache2 start d. See more about Defining metrics in Platform HPC. you can manually re-create them. pcm-upgrade-tool.1 to 4. Copy the Platform HPC entitlement file to the /opt/pcm/entitlement directory. Before you begin Ensure that a backup of your previous settings was created before you proceed with the upgrade.py backup --database -d /external-storage-mnt/db-backup/ where external-storage-mnt is the backup location on your external storage.2. Stop the xCAT daemon. Procedure 1. 3.1. you can apply your backed up configuration files and data. /etc/init. Create a configuration file backup on the external storage.2. Before you upgrade.1. The database backup backs up the database data and schema. make sure to back up your LSF configuration files and data. After the upgrade is complete.

2/other_files/lsf9.2//other_files/lsf.config in the /install/kits/kit-phpc-4. b.py services --reconfig 2.2-1.py upgrade --postupdate 3.install.gz where config-backup is the location of the configuration file backup. and update the compute nodes.config .1.2/other_files/ For example: tar xvzf /install/kits/kit-phpc-4.2/ repos/kit-phpc-4.2-rhels-6-x86_64 directory.1. Uninstall the GMF package and the monitoring packages.tar.install. rpm rpm rpm rpm -e -e -e -e --nodeps --nodeps --nodeps --nodeps pcm-chassis-monitoring-1./ 3) Run the LSF installer. you must manually reinstall these packages.2/other_files/lsf.py restore --database -d /external-storage-mnt/db-backup/ where external-storage-mnt is the backup location on your external storage and db-backup is the location of the database backup. .1/other_files directory and modify the parameters as needed.pcm-upgrade-tool. Run the LSF installation. run the following commands: rpm rpm rpm rpm -qa -qa -qa -qa | | | | grep grep grep grep chassis-monitoring switch-monitoring gpfs-monitoring gmf a.2/other_files. Install the GMF package that is found in the /install/kits/kit-pcm-4. Refer to the lsf. cp /install/kits/kit-phpc-4.1 to LSF 9.2.3_lsfinstall_linux_x86_64. Refresh the database and configurations: pcm-upgrade-tool. cd /tmp/lsf 2) Copy the lsf.config Completing the upgrade To complete the upgrade to the next release of IBM Platform HPC.1-1. pcm-upgrade-tool.1-1. you must restore your system settings.config configuration file from /install/kits/kit-phpc-4.x86_64 pcm-switch-monitoring-1. Create an LSF installer configuration file (lsf.install. To check which monitoring packages are installed.x86_64 b. c. The LSF installer package is placed at /install/kits/kit-phpc-4.Z -C /tmp/lsf d.1. Restart Platform HPC services.1-1.2/other_files/KIT_phpc_lsf_setup /install/postscripts/ cp /install/kits/kit-phpc-4.3.install. 1) Navigate to the LSF installer directory. Procedure 1.2/other_files/KIT_phpc_lsf_config /install/postscripts/ cp /install/kits/kit-phpc-4. cp /install/kits/kit-phpc-4.install. Upgrade the LSF component from Version 9. Upgrading 53 .py restore --files -f /external-storage-mnt/config-backup/ 20130708-134535.2/other_files directory.x86_64 pcm-gmf-1. a.config) and add it to the /install/kits/kit-phpc-4. database settings.2.1.config /install/postscripts/phpc c./lsfinstall -f lsf.tar.2. If you previously installed GMF and the related monitoring packages with Platform HPC. Chapter 10.x86_64 pcm-gpfs-monitoring-1.install.1. pcm-upgrade-tool. Replace LSF postscripts to in directory /install/postscripts/. Extract LSF installer package to a temp directory. Restore configuration files from a previous backup. 3.

Install the chassis monitoring package that is found in the /install/kits/kit-pcm-4.rpm d. use the rpower command to reboot the node: rpower noderange reset where noderange is a comma-separated list of nodes or node groups b.rpm e. Restart Platform HPC services.2 .2-1.x86_64. An SSL V3 security issue exists within the Tomcat server when HTTPS is enabled. rpm -ivh pcm-chassis-monitoring-1.xml file. pcmadmin service stop --service WEBGUI pcmadmin service start --service WEBGUI 54 Installing IBM Platform HPC Version 4. and save the file. rpm –ivh pcm-gpfs-monitoring-1.2/repos/kit-phpc-4. rpm -ivh pcm-switch-monitoring-1. updatenode noderange -S where noderange is a comma-separated list of nodes or node groups. If you have not previously taken steps to fix this issue. run the following command to install the GPFS monitoring package. egosh ego start -f" where noderange is a comma-separated list of nodes or node groups.1-1.text/plain" connectionTimeout="20000" URIEncoding="UTF-8"/> b. a. Check if the compute nodes are reachable.2-rhels-6-x86_64 directory.2. if you have HTTPS enabled. Run the following command on the management node.1-1. xdsh noderange "source /shared/ibm/platform_lsf/conf/ego/phpc_cluster/kernel/profile. Upgrade compute nodes. For example: <Connector port="${CATALINA_HTTPS_START_PORT}" maxHttpHeaderSize="8192” maxThreads="${CATALINA_MAX_THREADS}" minSpareThreads="25" maxSpareThreads="75" enableLookups="false" disableUploadTimeout="true" acceptCount="100" scheme="https" secure="true" clientAuth="want" sslProtocol="TLS" algorithm="ibmX509" compression="on" compressionMinSize="2000" compressableMimeType="text/html. Otherwise. c. egosh ego shutdown -f. Restart the Web Portal service. complete the following steps to fix this issue. Edit the $GUI_CONFDIR/server.ego. you can skip this step. The GPFS monitoring package is available in the /install/kits/kit-pcm-4.2-rhels-6-x86_64 directory. Install the switch monitoring package that is found in the /install/kits/kit-pcm-4. a. Restart the LSF cluster.2-rhels-6-x86_64 directory.1-1. lsfrestart -f 6. 5.rpm f.rpm c. ping the compute nodes to ensure that they are connected to the management node: xdsh noderange "/bin/ls" For any compute nodes that have lost connection and cannot be reached.text/javascript. Restart monitoring services. Compute node connections can get lost during the upgrade process.2/repos/kit-phpc-4.2.2/repos/kit-phpc-4. If you have GPFS installed.rpm -ivh pcm-gmf-1.x86_64.text/xml.2.x86_64. # pcmadmin service restart --group ALL 4. set the sslProtocol value from SSL to TLS. Update compute nodes to include the Platform HPC package.text/css. In the connector XML tag.x86_64.

c. If you are connected to a public network..log file in the /opt/pcm/log directory.2 with OS reinstall Upgrade your existing installation of IBM Platform HPC to the most recent version. a. 4.. Go to http://mgtnode-IP:8080. Procedure 1. res (pid 15873) is running. Log in to the Web Portal.. # cat /etc/phpc-release 6. # service postgresql status (pid 13269) is running.. Log in to the management node as a root user..sh 3. Open a supported web browser. Note: A detailed log of the upgrade process can be found in the upgrade. # . Upgrading 55 . Source Platform HPC environment variables. where mgtnode-IP is the real management node IP address. Check that the correct version of Platform HPC is running. Check that the PostgreSQL database server is running. /opt/pcm/bin/pcmenv. Log in as a root user. d. The root user has administrative privileges and maps to the operating system root user. there are some steps you should follow to ensure your upgrade is successful. sbatchd (pid 15881) is running. Upgrading to Platform HPC 4. SERVICE RULE-EN* PCMD JOBDT PLC PURGER PTC PLC2 WEBGUI ACTIVEMQ STATE STARTED STARTED STARTED STARTED STARTED STARTED STARTED STARTED STARTED ALLOC 18 17 12 13 11 14 15 19 16 CONSUMER /Manage* /Manage* /Manage* /Manage* /Manage* /Manage* /Manage* /Manage* /Manage* RGROUP RESOURCE SLOTS Manag* * 1 Manag* * 1 Manag* * 1 Manag* * 1 Manag* * 1 Manag* * 1 Manag* * 1 Manag* * 1 Manag* * 1 [ SEQ_NO 1 1 1 1 1 1 1 1 1 OK ] INST_STATE RUN RUN RUN RUN RUN RUN RUN RUN RUN ACTI 17 16 11 12 10 13 14 18 15 5. After you log in. # service xcatd status xCAT service is running # service phpc status Show status of the LSF subsystem lim (pid 15858) is running. Chapter 10. where mgtnode-hostname is the real management node hostname.. and reinstall or upgrade the operating system on the management node. you can also navigate to http://mgtnode-hostname:8080..Verifying the upgrade Ensure that the upgrade procedure is successful and that Platform HPC is working correctly. Preparing to upgrade Before upgrading your IBM Platform HPC installation. Refer to the Release Notes for a list of supported web browsers. the Resource Dashboard is displayed in the Web Portal.. b. Check that the Platform HPC services are running. 2.

iso /mnt 2.2-sles-11-x86_64/pcm-upgrade-tool-*. run the following commands: # du -sh /var/lib/pgsql/data # du -sh /install/ Note: It is recommended that the size of your external storage is greater than the combined size of the database and the /install directory.1. Set up the upgrade environment.2. b. mkdir /external-storage-mnt/db-backup where the external-storage-mnt is the backup location on your external storage. ensure that you have the following items: v You must have an external backup to store the contents of your 4. export PATH=${PATH}:/opt/pcm/libexec/ 4. you must complete the following steps: Procedure 1. if any. Upgrade the pcm-upgrade-tool package. v The Platform HPC 4. mkdir /external-storage-mnt/config-backup where the external-storage-mnt is the backup location on your external storage. Ensure that the external storage has enough space for the backup files. a. Create a directory for the configuration file backup. 5. 56 Installing IBM Platform HPC Version 4. About this task Before you upgrade to the next release of Platform HPC. 6. For RHEL: rpm -Uvh /mnt/packages/repos/kit-phpc-4.rpm 3.rpm For SLES: rpm -Uvh /mnt/packages/repos/kit-phpc-4.1 backup. Mount the Platform HPC installation media: mount -o loop phpc-4. To check how much space you require for the back. create a directory for the database backup. v If you are upgrading the operating system. Determine which custom metrics you are using.2-rhels-6-x86_64/pcm-upgrade-tool-*.2 ISO file. Prepare an external storage. make sure that you have the RHEL ISO file. /opt/xcat/bin/updatenode mn-host-name -F where mn-host-name is the name of your management node. On the external storage.1. c.x64. and that you have a corresponding OS distribution created. you must include these new users in your backup. and can manually be re-created after the upgrade is completed. For additional requirements refer to “Upgrading checklist” on page 49. The custom metrics are lost in the upgrade process.Before you begin To prepare for your upgrade.2 . If you created any new users after Platform HPC was installed.

The database backup backs up the database data and schema. Stop Platform HPC services: pcm-upgrade-tool. b. the following custom configurations can be manually re-created: v Customization to the PERF loader. you can apply your backed up configuration files and data. 3. After the upgrade procedure is completed. including internal data collection and the purger configuration files v Customization to the Web Portal Help menu navigation v Addition of custom metrics v Alert polices v LDAP packages and configurations Before you begin Platform HPC does not back up or restore LSF configuration files or data.py services --stop 2. make sure to back up your LSF configuration files and data. Reinstall the RHEL 6. Create a configuration file backup on the external storage. The backup includes database files and the backup configuration file pcm.py backup --files -d /external-storage-mnt/config-backup/ Performing the Platform HPC upgrade Perform the upgrade with reinstalling the operating system and restore your settings. and default gateway. If you are upgrading to a new machine.py backup --database -d /external-storage-mnt/db-backup/ where external-storage-mnt is the backup location on your external storage. Create a database backup on the external storage. After the upgrade is complete. you must power off the old management node before you power on the new management node. Chapter 10. IP address. Ensure you use the same network settings as the old management node. IP address. complete the following steps: a. pcm-upgrade-tool. netmask. including: hostname. and default gateway. Record the following management node network settings: hostname. netmask.conf.5 operating system on the management node. pcm-upgrade-tool. Before you upgrade. Procedure 1.Backing up Platform HPC Create a backup of your current Platform HPC installation that includes a backup of the database and settings before you upgrade to a newer version of Platform HPC. Before you begin Ensure that you have prepared for the upgrade and have an existing backup of your previous settings. Upgrading 57 . Reinstall the management node. c. Procedure 1. Note: The backup procedure does not back up any custom configurations.

Restore configuration files from a previous backup. Install Platform HPC 4.conf 3.conf c.iso /mnt 58 Installing IBM Platform HPC Version 4. Edit the silent installation template and set the os_kit parameter to the absolute path of the operating system ISO.example .4-x86_64. database settings.conf b. Locate the default silent installation template autoinstall.1.1.gz where config-backup is the location of the configuration file backup. c. Procedure 1.conf.rhel. By default. run the following commands: a. pcm-upgrade-tool.1. vi .x64./phpc-autoinstall.tar./phpc-autoinstall.2. Restore database data from a previous backup. The OS distribution files can be manually created after the management node upgrade is complete and before upgrading the compute nodes.2 . you can manually re-create them. pcm-upgrade-tool. you must restore your system settings.py services --reconfig 2.py services -–stop b. In this step. a. d./phpc-autoinstall. To recreate an OS distribution.example in the docs directory in the installation ISO. Restart Platform HPC services. and update the compute nodes. Set up your environment. /mnt/phpc-installer -f . If you are using a different operating system. Start the installation by running the silent installation. Stop Platform HPC services. mount -o loop phpc-4. Restore settings and database data.Refer to “Installing and configuring the operating system on the management node” on page 13 on more information on installing an RHEL operating system.2. export PATH=${PATH}:/opt/pcm/libexec/ 4. If you created custom metrics in Platform HPC 4.py restore --database -d /external-storage-mnt/db-backup/ where external-storage-mnt is the backup location on your external storage and db-backup is the location of the database backup. Mount the operating system.conf. the OS distribution files are not backed up or restored. pcm-upgrade-tool. the RHEL operating system is specified.iso /mnt cp /mnt/docs/phpc-autoinstall. pcm-upgrade-tool. specify the operating system accordingly. # mount -o loop rhel-6. complete the following steps: a. 2. Refer to the "Defining metrics in Platform HPC" section in the Administering Platform HPC guide for more information. Related information: “Installing and configuring the operating system on the management node” on page 13 Completing the upgrade To complete the upgrade to the next release of IBM Platform HPC and complete the operating system reinstallation.py restore --files -f /external-storage-mnt/config-backup/ 20130708-134535.

text/css. if you have HTTPS enabled.xml file. use the lsdef -t osdistro rhels6.4/x86_64 c. set the sslProtocol value from SSL to TLS. c.platform.4-x86_64.text/javascript. complete this step. xdsh noderange "source /opt/pcm/ego/profile. # rsync -a /mnt/* /install/rhels6. the LDAP configurations are not backed up or restored.iso is the name of the OS distribution. d. If you want to upgrade the compute nodes to a higher OS version. Otherwise.4-x86_64 command to get the OS distribution path. egosh ego shutdown -f. you can skip this step. ping the compute nodes to ensure that they are connected to the management node: xdsh noderange "/bin/ls" For any compute nodes that have lost connection and cannot be reached. The backup directory must be the same as the OS distribution path. Compute node connections can get lost during the upgrade process. Update compute nodes.text/plain" connectionTimeout="20000" URIEncoding="UTF-8"/> Chapter 10. Create a new backup directory. In the connector XML tag. To determine the OS distribution path. 5.4/x86_64 3.text/xml. Restart monitoring services. b. a.py upgrade --postupdate 4. refer to "LDAP user authentication" section in the Administering Platform HPC guide.where rhel-6.2 package. Edit the $GUI_CONFDIR/server. An SSL V3 security issue exists within the Tomcat server when HTTPS is enabled. you must reprovision them. Upgrading 59 . Synchronize the new directory. By default. 6. For example: <Connector port="${CATALINA_HTTPS_START_PORT}" maxHttpHeaderSize="8192” maxThreads="${CATALINA_MAX_THREADS}" minSpareThreads="25" maxSpareThreads="75" enableLookups="false" disableUploadTimeout="true" acceptCount="100" scheme="https" secure="true" clientAuth="want" sslProtocol="TLS" algorithm="ibmX509" compression="on" compressionMinSize="2000" compressableMimeType="text/html. and save the file. Recover the SSH connection to the compute nodes. complete the following steps to fix this issue. # mkdir /install/rhels6. Check if the compute nodes are reachable. If you want to enable LDAP. xdsh noderange -K where noderange is a comma-separated list of nodes or node groups. Refresh the database and configurations: pcm-upgrade-tool. If you have not previously taken steps to fix this issue. use the rpower command to reboot the node: rpower noderange reset where noderange is a comma-separated list of nodes or node groups b. egosh ego start -f" where noderange is a comma-separated list of nodes or node groups. a. updatenode noderange -S where noderange is a comma-separated list of nodes or node groups. Update compute nodes to include the Platform HPC 4. Otherwise.

. The root user has administrative privileges and maps to the operating system root user. Note: A detailed log of the upgrade process can be found in the upgrade. Procedure 1..2 . Log in as a root user.log file in the /opt/pcm/log directory. you can also navigate to http://mgtnode-hostname:8080.. res (pid 15873) is running. SERVICE RULE-EN* PCMD JOBDT PLC PURGER PTC PLC2 WEBGUI ACTIVEMQ STATE STARTED STARTED STARTED STARTED STARTED STARTED STARTED STARTED STARTED ALLOC 18 17 12 13 11 14 15 19 16 CONSUMER /Manage* /Manage* /Manage* /Manage* /Manage* /Manage* /Manage* /Manage* /Manage* RGROUP RESOURCE SLOTS Manag* * 1 Manag* * 1 Manag* * 1 Manag* * 1 Manag* * 1 Manag* * 1 Manag* * 1 Manag* * 1 Manag* * 1 [ SEQ_NO 1 1 1 1 1 1 1 1 1 OK ] INST_STATE RUN RUN RUN RUN RUN RUN RUN RUN RUN ACTI 17 16 11 12 10 13 14 18 15 5.. # service xcatd status xCAT service is running # service phpc status Show status of the LSF subsystem lim (pid 15858) is running. Troubleshooting upgrade problems Troubleshooting problems that occur when upgrading to the new release of IBM Platform HPC. Refer to the Release Notes for a list of supported web browsers. /opt/pcm/bin/pcmenv. Log in to the management node as a root user. a.. Open a supported web browser.sh 3. where mgtnode-IP is the real management node IP address. Restart the Web Portal service. where mgtnode-hostname is the real management node hostname. 4. Log in to the Web Portal. Check that the PostgreSQL database server is running.b. d.. If you are connected to a public network. # . 2.. Check that the Platform HPC services are running. # service postgresql status (pid 13269) is running. 60 Installing IBM Platform HPC Version 4. Go to http://mgtnode-IP:8080. pcmadmin service stop --service WEBGUI pcmadmin service start --service WEBGUI Verifying the upgrade Ensure that the upgrade procedure is successful and that Platform HPC is working correctly. b. c. # cat /etc/phpc-release 6.. After you log in. Source Platform HPC environment variables. Check that the correct version of Platform HPC is running. sbatchd (pid 15881) is running. the Resource Dashboard is displayed in the Web Portal.

psql:/opt/pcm/etc/upgrade/postupdate/4.2. In most cases.xcat/keystore_pcm v If any of the following errors are found in the upgrade. and logs any warnings or errors that occur during the upgrade process Common upgrade problems include the following issues: v Cannot log in to the Web Portal after upgrading to Platform HPC Version 4. Before you begin Before you rollback to Platform HPC 4.1. psql:/opt/pcm/etc/upgrade/postupdate/4.2/update-pcmgui-records. some pages in the Web Portal do not display or display old data.xcat/keystore_pcm After upgrading to Platform HPC Version 4. To resolve this issue try the following resolutions: – Restart the Web Portal. psql:/opt/pcm/etc/upgrade/postupdate/4.2/update-pcmgui-records. psql:/opt/pcm/etc/upgrade/postupdate/4.2/update-pcmgui-records. However. v After upgrading to Platform HPC Version 4. Chapter 10. if the Web Portal goes down. they can be ignored and no further actions need to be taken.1.2.sql:29: ERROR: permission denied: "RI_ConstraintTrigger_17314" is a system trigger psql:/opt/pcm/etc/upgrade/postupdate/4.data.sql:13: ERROR: duplicate key value violates unique constraint "pmc_role_pkey" DETAIL: Key (role_id)=(10005) already exists.2/update-pcmgui-records.1 Revert to the earlier version of Platform HPC. issue the following command: # pcmadmin service restart --service WEBGUI Then run the following command from the management node to resolve this issue: /opt/pcm/libexec/pcmmkcert. Upgrading 61 .1. Run the following command from the management node to resolve this issue: v /opt/pcm/libexec/pcmmkcert.1 ISO and the original operating system ISO.sql:18: ERROR: duplicate key value violates unique constraint "pmc_role_permission_pkey" DETAIL: Key (role_permission_id)=(10009) already exists. This file logs informational information about the upgrade procedure.sql:15: ERROR: duplicate key value violates unique constraint "pmc_resource_permission_pkey" DETAIL: Key (resperm_id)=(11001-5) already exists. To resolve this issue.2/update-pcmgui-records. ensure that you have both the Platform HPC 4.1.data. you can restart services and daemons manually. the services that are required to run the Web Portal start automatically.sh /root/. clear your web browser cache and relogin to the Web Portal.log file that is found in the /opt/pcm/log directory. From the command line.1.sql:11: ERROR: duplicate key value violates unique constraint "ci_purge_register_pkey" DETAIL: Key (table_name)=(lim_host_config_history) already exists.sql:25: ERROR: permission denied: "RI_ConstraintTrigger_17314" is a system trigger psql:/external-storage-mnt/db-backup/pmc_group_role. Rollback to Platform HPC 4.To help troubleshoot your upgrade process. some pages in the Web Portal do not display. psql:/external-storage-mnt/db-backup/pmc_group_role.log file that is found in the /opt/pcm/log directory.sql:7: ERROR: duplicate key value violates unique constraint "ci_purge_register_pkey" DETAIL: Key (table_name)=(pcm_node_status_history) already exists.2.sh /root/. you can view the upgrade.1.1.

Refer to the "Defining metrics in Platform HPC" section in the Administering Platform HPC guide for more information.1.py services --reconfig 5.1. Restore database data from a previous backup. if needed. Reinstall the management node.1. pcm-upgrade-tool. Set up the environment: export PATH=${PATH}:/opt/pcm/libexec/ b. vi . Ensure you use the same network settings as the old management node.1. Reinstall the original operating system on the management node. Start the installation by running the installation program specifying the silent installation file.1. IP address. /mnt/phpc-installer -f ./phpc-autoinstall. including: hostname.1 or 4. Restore configuration files from a previous backup.py services –-stop c./phpc-autoinstall. netmask. v If the compute nodes have Platform HPC 4. recover the SSH connection for all compute nodes: xdsh noderange -K where noderange is a comma-separated list of nodes or node groups.conf b.tar. Locate the default silent installation template autoinstall.py restore --database -d /external-storage-mnt/db-backup/ where external-storage-mnt is the backup location on your external storage and db-backup is the location of the database backup.1.x86_64. complete the following steps: a.1 installed. v If the compute nodes have Platform HPC 4.1. Stop Platform HPC services: pcm-upgrade-tool./phpc-autoinstall. Edit the silent installation template and set the os_kit parameter to the absolute path for the operating system ISO. 62 Installing IBM Platform HPC Version 4.2 .1. pcm-upgrade-tool.example . Refer to “Installing and configuring the operating system on the management node” on page 13 on more information on installing an operating system.conf.example in the docs directory in the installation ISO.gz where config-backup is the location of the configuration file backup.1.1.Procedure 1.py restore --files -f /external-storage-mnt/config-backup/20130708-134535.iso /mnt cp /mnt/docs/phpc-autoinstall. and default gateway. and default gateway. netmask.2 installed. you can manually re-create them. Restart Platform HPC services: pcm-upgrade-tool. Reinstall compute nodes. b.1.2. 4. 2.1. complete the following steps: a. Restore settings and database data. complete the following steps: a. e. IP address. If you created custom metrics in Platform HPC 4.1. d.conf 3.conf. Install Platform HPC 4.conf c. mount -o loop phpc-4. Record the following management node network settings: hostname. they must be reprovisioned to use Platform HPC 4.

map configuration file in the $GUI_CONFDIR/application/vnc directory. you can upgrade your LSF or PAC entitlement file from Express to Standard.map file must specify the IP address that is mapped to the host name. Copy the new entitlement file to the unified entitlement path (/opt/pcm/entitlement/phpc. Before you begin To upgrade your product entitlement for PAC. Log in to the Web Portal. Chapter 10. About this task After you upgrade to PAC Standard. Upgrading LSF entitlement In IBM Platform HPC. ensure that you are able to connect to the remote jobs console. 2. contact IBM client services for more details and to obtain the entitlement file. Restart LSF. 2.entitlement). update the vnc_host_ip.Upgrading entitlement In IBM Platform HPC. Upgrading 63 . Procedure 1. About this task To upgrade your entitlement. pmcadmin stop pmcadmin start Results Your LSF entitlement is upgraded to the standard version. complete the following steps to connect to the remote jobs console. you can upgrade your LSF entitlement file from Express to Standard. Restart the Web Portal. Before you begin To upgrade your product entitlement for LSF. From the command line. complete the following steps on the Platform HPC management node: Procedure 1. The vnc_host_ip. after upgrading your Platform Application Center (PAC) entitlement file from Express Edition to Standard Edition. Upgrading PAC entitlement In IBM Platform HPC. as a root user. contact IBM client services for more details and to obtain the entitlement file. lsfrestart 3.

then you are missing the VncViewer. copy the VncViewer.jar Results Using PAC Standard Edition. and click Remote Job Consoles. see "Remote Console". Go to the Jobs tab.map # This file defines which IP will be use for the host.251.2 . b. for example #hostname1=192. Issue the following command: #cp /opt/pcm/web-portal/gui/3. Verify that the remote job console is running. #vncserver :1 . 8. If you get the following error. Go to the /opt/pcm/web-portal/gui/work/.111. exist.jar /opt/pcm/web-portal/gui/3. 5.0/tomcat/webapps/platform/viewgui/common/applet/VncViewer.console and vnc. Kill any VNC server sessions if they exist. 7.jar.0/tomcat/webapps/platform/pac/vnc/lib/VncViewer. Cannot find the required VNC jar file: /opt/pcm/web-portal/gui/3. To resolve this error.2 system3750=9.0/tomcat/webapps/platform/pac/vnc/lib directory.session.jar file. vncserver -kill :${session_id} 4. Restart the Web Portal.141 3. Restart the VNC server.168. you are able to connect to the remote jobs console. c. then delete them.vncserver :2 6.jar file to the /opt/pcm/web-portal/gui/3.1. 64 Installing IBM Platform HPC Version 4. vnc. Click Open My Console. If the VNC sessions files. Stop the iptables service on the management node.vnc/${USER}/ directory.0/tomcat/webapps/platform/pac/vnc/lib/VncViewer. For details about configuring remote consoles. a.# cat vnc_host_ip.

Extract the fix tar file. Go to IBM Fix Central. c. ensure that the same fixes are applied on the primary management node and the failover node. Select Platform HPC as the product name. Note: In a high availability environment. Note: In a high availability environment. Select 4. From the directory where the fix files are extracted to. b. Apply the fixes from the command line. d. run the installation script to install the fix. Procedure 1. b. Locate the product fixes. Applying fixes Check for any new fixes that can be applied to your Platform HPC installation. Select your platform. 2014 65 . Select Platform Computing as the product group. ensure that the same fixes are applied on the primary management node and the failover node About this task Fixes are available for download from IBM Fix Central website. Download each individual fix. 1994. a. 4. © Copyright IBM Corp. 3.Chapter 11.2 as the installed version. by selecting the following options: a. 2.

2 .66 Installing IBM Platform HPC Version 4.

20. For example: nicips.5 Note: A virtual NIC does not need to be created and the IP address does not need to be configured. 1994.work=work-directory Specifies the shared storage location for system work data.1.Chapter 12. For example: 172. The pcmhatool command automatically creates the needed configurations.eth0:0=eth0-IP-address Specifies the virtual IP address of a virtual NIC connected to the management node.work=work-directory sharefs_mntp.eth1:0=eth1-IP-address sharefs_mntp. nicips. If the same shared directory is used for both user home data and system work data.5 Note: A virtual NIC does not need to be created and the IP address does not need to be configured.20.home=home-directory virualmn-name: Specifies virtual node name of the active management node. Only NFS is supported.eth0:0=172. The virtual node name must be a valid node name. virtualmn-name: nicips.7. © Copyright IBM Corp. For example: nicips.eth0:0=eth0-IP-address nicips. This line must end with a colon (:).eth1:0=192. 2014 67 . where eth1-IP-address is an IP address. The pcmhatool command automatically creates the needed configurations.7. nicips. where virualmn-name is the name of the virtual node.200:/export/data. High availability definition file The high availability definition file specifies values to configure a high availability environment. References Configuration files High availability definition file High availability definition file specifies values to configure high availability. it must be the short name without the domain name. It cannot be a fully qualified domain name. where eth0-IP-address is an IP address. where work-directory is the shared storage location. sharefs_mntp.eth1:0=eth1-IP-address Specifies the virtual IP address of a virtual NIC connected to the management node.168. specify this parameter as the single shared directory.

20. is used as the location for both user home data and system work data.200:/export/home Commands pcmhatool an administrative command interface to manage a high availability environment Synopsis pcmhatool [-h | --help] | [-v | --version] pcmhatool subcommand [options] Subcommand List pcmhatool config -i | --import HAINFO_FILENAME -s | --secondary SMN_NAME [-q | --quiet] [-h | --help] pcmhatool reconfig -s|--standby SMN_NAME [-q|--quiet] [-h|--help] pcmhatool info [-h|--help] pcmhatool failto -t|--target SMN_NAME [-q|--quiet] [-h|--help] pcmhatool failmode -m|--mode FAILOVER_MODE [-h|--help] pcmhatool status [-h|--help] pcmhatool check [-h|--help] 68 Installing IBM Platform HPC Version 4.100 nicips. do not specify this parameter.20.200:/export/data # Shared storage for user home data sharefs_mntp. Only NFS is supported.168.200:/export/home.7. nicips.eth1:0=172.20.7.sharefs_mntp.100 # Shared storage for system work data sharefs_mntp.7.0. The specified sharefs_mntp.home=172.7.work=172.20. If the same shared directory is used for both user home data and system work data.work parameter.2 .eth0:0=192. Example The following is an example of a high availability definition file: # A virtual node name virtualmn: # Virtual IP address of a virtual NIC connected to the management node. For example: 172. where home-directory is the shared storage location.home=home-directory Specifies the shared storage location for user home data.

info Displays high availability settings. where SMN_NAME is the name of the secondary management node. including: the virtual IP address. Nodes that are in unavail state are unavailable and indicate a node failure or lost network connection. failmode -m|--mode FAILOVER_MODE Sets the failover mode. failure and correction data. and show high availability data and running status. the standby node only takes over the cluster if the pcmhatool failto command is issued. Subcommand Options config -i HAINFO_FILENAME -s SMN_NAME Specifies high availability settings to be used to enable high availability between the primary management node and the secondary management node. where HAINFO_FILENAME is the high availability definition file and SMN_NAME is the name of the secondary management node. References 69 . Chapter 12. It is used to enable high availability. where SMN_NAME is the current standby management node. where SMN_NAME is the name of the standby management node.Description The pcmhatool command manages a high availability environment. where HAINFO_FILENAME is the name of the high availability definition file. the standby node takes over the cluster when it detects the active node has failed. reconfig -s|--standby SMN_NAME Enables high availability on the standby management node after the management node is reinstalled. display settings. the management node name. and a list of shared directories. In manual mode. trigger a failover. -v | --version Displays the pcmhatool command version information. where FAILOVER_MODE is set to auto for automatic failover or manual for manual failover. Options -h | --help Displays the pcmhatool command help information. including: state of the nodes. -i|--import HAINFO_FILENAME Specifies the import file name of the high availability definition file. In automatic mode. status Displays the current high availability status. failover mode and status of running services. check Displays high availability diagnostic information related to the high availability environment. including current status data. failto -t|--target SMN_NAME Sets the specified standby management node to an active management node. -s|--secondary SMN_NAME Specifies the secondary management node name. set the failover mode.

70 Installing IBM Platform HPC Version 4.2 .

to: Intellectual Property Licensing Legal and Intellectual Property Law IBM Japan Ltd. or service is not intended to state or imply that only that IBM product. However. INCLUDING.S. Any functionally equivalent product. IBM may have patents or pending patent applications covering subject matter described in this document.S. 2014 71 . services. in writing. Some states do not allow disclaimer of express or implied warranties in certain transactions. THE IMPLIED WARRANTIES OF NON-INFRINGEMENT. program. program.A. Shimotsuruma. NY 10504-1785 U.Notices This information was developed for products and services offered in the U. Any references in this information to non-IBM Web sites are provided for convenience only and do not in any manner serve as an endorsement of those Web sites. this statement may not apply to you.A. Any reference to an IBM product. Yamato-shi Kanagawa 242-8502 Japan The following paragraph does not apply to the United Kingdom or any other country where such provisions are inconsistent with local law: INTERNATIONAL BUSINESS MACHINES CORPORATION PROVIDES THIS PUBLICATION “AS IS” WITHOUT WARRANTY OF ANY KIND. IBM may make improvements and/or changes in the product(s) and/or the program(s) described in this publication at any time without notice. therefore. or features discussed in this document in other countries. You can send license inquiries. or service that does not infringe any IBM intellectual property right may be used instead. program. BUT NOT LIMITED TO. to: IBM Director of Licensing IBM Corporation North Castle Drive Armonk. or service may be used. contact the IBM Intellectual Property Department in your country or send inquiries. these changes will be incorporated in new editions of the publication. The furnishing of this document does not grant you any license to these patents. 1994. © Copyright IBM Corp. Changes are periodically made to the information herein. it is the user's responsibility to evaluate and verify the operation of any non-IBM product. program. or service. 1623-14. in writing. EITHER EXPRESS OR IMPLIED. Consult your local IBM representative for information on the products and services currently available in your area. For license inquiries regarding double-byte character set (DBCS) information. MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. This information could include technical inaccuracies or typographical errors. IBM may not offer the products. The materials at those Web sites are not part of the materials for this IBM product and use of those Web sites is at your own risk.

2 . All of these names are fictitious and any similarity to the names and addresses used by an actual business enterprise is entirely coincidental. including in some cases. NY 12601-5400 USA Such information may be available. To illustrate them as completely as possible. the results obtained in other operating environments may vary significantly. companies. their published announcements or other publicly available sources. Information concerning non-IBM products was obtained from the suppliers of those products. Questions on the capabilities of non-IBM products should be addressed to the suppliers of those products. This information contains examples of data and reports used in daily business operations. These examples have not been thoroughly tested under all conditions. IBM International Program License Agreement or any equivalent agreement between us. for the purposes of developing. therefore. Furthermore. modify. which illustrates programming techniques on various operating platforms. cannot guarantee or 72 Installing IBM Platform HPC Version 4. subject to appropriate terms and conditions. Poughkeepsie. Users of this document should verify the applicable data for their specific environment. Therefore. brands. The licensed program described in this document and all licensed material available for it are provided by IBM under terms of the IBM Customer Agreement. Actual results may vary. COPYRIGHT LICENSE: This information contains sample application programs in source language. IBM.IBM may use or distribute any of the information you supply in any way it believes appropriate without incurring any obligation to you. IBM has not tested those products and cannot confirm the accuracy of performance. Any performance data contained herein was determined in a controlled environment. You may copy. marketing or distributing application programs conforming to the application programming interface for the operating platform for which the sample programs are written. some measurement may have been estimated through extrapolation. and distribute these sample programs in any form without payment to IBM. the examples include the names of individuals. payment of a fee. compatibility or any other claims related to non-IBM products. Some measurements may have been made on development-level systems and there is no guarantee that these measurements will be the same on generally available systems. and represent goals and objectives only. Licensees of this program who wish to have information about it for the purpose of enabling: (i) the exchange of information between independently created programs and other programs (including this one) and (ii) the mutual use of the information which has been exchanged. should contact: IBM Corporation Intellectual Property Law Mail Station P300 2455 South Road. and products. All statements regarding IBM's future direction or intent are subject to change or withdrawal without notice. using.

other countries. Privacy policy considerations IBM Software products. Depending upon the configurations deployed. Portions of this code are derived from IBM Corp. IBM shall not be liable for any damages arising out of your use of the sample programs. Windows. Trademarks IBM. and the Windows logo are trademarks of Microsoft Corporation in the United States. In many cases no personally identifiable information is collected by the Software Offerings.imply reliability. to help improve the end user experience. this Software Offering may use session and persistent cookies that collect each user’s user name. Intel Inside logo. Intel Inside. Celeron. the photographs and color illustrations may not appear. or both. and Platform Computing are trademarks or registered trademarks of International Business Machines Corp. must include a copyright notice as follows: © (your company name) (year). (“Software Offerings”) may use cookies or other technologies to collect product usage information. for purposes of session management. Platform. and ibm. The sample programs are provided "AS IS". Intel logo. registered in many jurisdictions worldwide.shtml. to tailor interactions with the end user. or for other purposes..com® are trademarks of International Business Machines Corp. specific information about this offering’s use of cookies is set forth below. LSF. If this Software Offering uses cookies to collect personally identifiable information. or service names may be trademarks or service marks of others. Linux is a trademark of Linus Torvalds in the United States. serviceability. Other company. Each copy or any portion of these sample programs or any derivative work.com/legal/copytrade. _enter the year or years_. product. or both. If you are viewing this information softcopy. Intel Xeon. Intel SpeedStep. including software as a service solutions. Intel Centrino logo. without warranty of any kind. the IBM logo. Sample Programs. or function of these programs. Other product and service names might be trademarks of IBM or other companies. Itanium. Java and all Java-based trademarks and logos are trademarks or registered trademarks of Oracle and/or its affiliates. Notices 73 . Some of our Software Offerings can help enable you to collect personally identifiable information. other countries. Microsoft. © Copyright IBM Corp. A current list of IBM trademarks is available on the Web at "Copyright and trademark information" at http://www. These cookies cannot be disabled. and Pentium are trademarks or registered trademarks of Intel Corporation or its subsidiaries in the United States and other countries. Windows NT..ibm. Intel Centrino. Intel. registered in many jurisdictions worldwide.

74 Installing IBM Platform HPC Version 4.2 . see IBM’s Privacy Policy at http://www.com/software/info/product-privacy. including any requirements for notice and consent.ibm.com/privacy/details the section entitled “Cookies.com/privacy and IBM’s Online Privacy Statement at http://www.If the configurations deployed for this Software Offering provide you as customer the ability to collect personally identifiable information from end users via cookies and other technologies. For more information about the use of various technologies. including cookies.ibm. for these purposes. you should seek your own legal advice about any laws applicable to such data collection. Web Beacons and Other Technologies” and the “IBM Software Products and Software-as-a-Service Privacy Statement” at http://www.ibm.

.

 Printed in USA SC27-6107-02 .