Index Index for
Section 5
Index Alphabetical
listing for O
Bottom of page Bottom of
page

OLAR_intro(5)

NAME

OLAR_intro, olar_intro - Introduction to Online Addition and Removal (OLAR) Management

DESCRIPTION

Introduction to Online Addition and Removal (OLAR) Management Online addition and removal management is used to expand capacity, upgrade components, and replace failed components, while the operating system services and applications continues to run. This functionality, sometimes referred to as "hot-swap", provides the benefits of increased system uptime and availability during both scheduled and unscheduled maintenance. Starting with Tru64 UNIX Version 5.1B, CPU OLAR is supported. Additional OLAR capabilities are planned to be added for subsequent releases of the operating system. OLAR management is integrated with the SysMan suite of system management applications, which provides the ability to manage all aspects of the system from a centralized location. You must be a privileged user to perform OLAR management operations. Or, you may configure privileges for selective authorized user or group access using Division of Privileges (DOP), as described below. Note that only one administrator at a time can initiate OLAR operations; other administrators will be prevented from initiating OLAR operations while another operation completes. CPU OLAR Overview Tru64 UNIX supports the ability to add, replace, and/or remove individual CPU modules on supported AlphaServer systems while the operating system and applications continue to run. Newly inserted CPUs are automatically recognized by the operating system, but will not start scheduling and executing processes until the CPU module is powered on and placed online through any of the supported management applications as described below. Conversely, before a CPU can be physically removed from the system, it must be placed offline and then powered off. Processes queued for execution on a CPU that is to be placed offline are simply migrated to run-queues of other running (online) processors. By default, CPUs that are placed offline will persist across reboot and system initialization, until the CPU is explicitly placed online. This behavior differs from the default behavior of previous versions of Tru64 UNIX, where a CPU that was placed offline would return to service automatically after reboot or system restart. Note that for backward compatibility, the psradm(8) and offline(8) commands still provide the non-persistent offline behavior. While the psradm(8) and offline(8) commands are still provided, they are not recommended for performing OLAR operations. On platforms supporting this functionality, any CPU can participate in an OLAR operation, including the primary CPU and/or I/O interrupt handling CPUs. These roles will be delegated to other running CPUs in the event that a currently running primary or I/O interrupt handler needs to be placed offline or removed. Currently, the platforms which support CPU OLAR are the AlphaServer GS160, and GS320 series systems. The GS80 does not support the physical removal of CPU modules, due to cabinet packaging design. Why Perform OLAR on CPUs OLAR of CPUs may be performed for the following reasons: Computational Capacity Expansion A system manager wants to provide additional computational capacity to the system without having to bring the system down. As an example, an AlphaServer GS320 with available CPU slots can have it's CPU capacity expanded by adding additional CPU modules to the system while the operating system and applications continue to run. Maintenance Upgrade A system manager wants to upgrade specific system components to the latest model without having to bring the system down. As an example, a GS160 with earlier model Alpha CPU modules can be upgraded to later model CPUs with higher clock rates, while the operating system continues to run. Failed Component Replacement A system component is indicating a high incidence of correctable errors and the system manager wants to perform a proactive replacement of the failing component before it results in a hard failure. As an example, the Component Indictment facility (described below) has indicated excessive correctable errors in a CPU module and has therefore recommended its replacement. Once the CPU module has been placed offline and powered off, either through the Automatic Deallocation Facility (also described below) or through manual intervention, the CPU module can be replaced while the operating system continues to run. Cautions Before Performing OLAR on CPUs Before performing an OLAR operation, be aware of the following cautions: · When offlining or removing one or more CPUs, processes scheduled to run on the affected CPUs will be scheduled to execute on other running CPUs, thus redistributing the processing capacity among the remaining CPUs. In general, this will result in a system performance degradation, proportional to the number of CPUs taken out of service and the current system load, for the period of the OLAR operation. Multi-threaded applications that are written to take advantage of known CPU concurrencies can expect to encounter significant performance degradation during the period of the OLAR operation. · The OLAR management utilities do not presently operate with processor sets. Processor sets are groups of processors that are dedicated for use by selected processes (see processor_sets(4)). If a process has been specifically bound to run on a processor set (see runon(1), assign_pid_to_pset(3) ), and an OLAR operation is attempted on the last running CPU in the processor set, you will not be notified by the OLAR utilities that you are effectively shutting down the entire processor set. Offlining the last CPU in a processor set will cause all processes bound to that processor set to suspend until the processor set has at least one running CPU. Therefore, use caution when performing CPU OLAR operations on systems that have been configured with processor sets. · If a process has been specifically bound to execute on a CPU (see runon(1), bind_to_cpu(3), and bind_to_cpu_id(3) for more information), and an OLAR operation is attempted on that CPU, you will be notified by the OLAR utilities that processes have been bound to the CPU prior to any operation being performed. You may choose to continue or cancel the OLAR operation. By choosing to continue, processes bound to a CPU will suspend their execution until such time that the process is un- bound, or the CPU is placed back online. Note that choosing to offline a CPU that has processes bound may cause detrimental consequences to the application, depending upon the characteristics of the application. · If a process has been specifically bound to execute on a Resource Affinity Domain (RAD) (see runon(1) and rad_bind_pid(3) for more information), and an OLAR operation is attempted on the last running CPU in the RAD, you will be notified by the OLAR utilities that processes have been bound to the RAD and that the last CPU in the RAD has been requested to be placed offline. By choosing to continue, processes bound to the RAD will suspend their execution until such time that the process is un-bound, or at least one CPU in the RAD is placed online. Note that choosing to offline the last CPU in a RAD with processes bound may cause detrimental consequences to the application, depending upon the characteristics of the application. · If you are using program profiling utilities such as dcpi, kprofile, or uprofile, that are aware of the system's CPU configuration, unpredictable results may occur when performing OLAR operations. It is therefore recommended that these profiling utilities be disabled prior to performing an OLAR operation. Ensure that all the processes including any associated daemons that are related to these utilities have been stopped before performing OLAR operations on system CPUs. The device drivers used by these profiling utilities are usually configured into the kernel dynamically, so the tools can be disabled before each OLAR operation with the following commands: # sysconfig -u pfm # sysconfig -u pcount The appropriate driver can be re-enabled with one of the following: # sysconfig -c pfm # sysconfig -c pcount The automatic deallocation of CPUs, enabled through the Automatic Deallocation Facility, should be disabled whenever the pfm or pcount device drivers are configured into the kernel, or vice versa. Refer to the documentation and reference pages for these utilities for additional information. General Procedures for Online Addition and Removal of CPUs Caution Pay attention to the system safety notes as outlined in the GS80/160/320 Service Manual. · Removing a CPU Module To perform an online removal of a CPU module, follow these steps using your preferred management application, described in the section "Tools for Managing OLAR". 1. Off-line the CPU. The operating system will stop scheduling and executing tasks on this CPU. Using your preferred OLAR management application, make note of the quad building block (QBB) number where this CPU is inserted. This is the "hard" (or physical) QBB number, and does not change if the system is partitioned. 2. Power the CPU module off. The LED on the CPU module will illuminate yellow, indicating that the CPU module is un-powered, and safe to be removed. 3. Physically remove the CPU module. Note that the operating system automatically recognizes that the CPU module has been physically removed. There is no need to perform a scan operation to update the hardware configuration. · Adding a CPU module To perform an online addition of a CPU module, follow these steps using your preferred management application, described in the section "Tools for Managing OLAR". 1. Select an available CPU slot in one of the configured quad building blocks (QBB). If there are available slots in several QBBs, it is typically best to equally distribute the number of CPUs among the configured QBBs. 2. Insert the CPU module into the CPU slot. Ensure that you align the color-coded decal on the CPU module with the color-code decal on the CPU slot. The LED on the CPU module will illuminate yellow, indicating that the CPU module is un-powered. Note that the CPU will be automatically recognized by the operating system, even though it is un-powered. There is no need to perform a scan operation for the operating system to identify the CPU module. 3. Power the CPU module on. The CPU module will undergo a short self-test (7-10 secs), after which the LED will illuminate green, indicating the module is powered-on and has passed its self-test. 4. On-line the CPU. Once the CPU is on-line, the operating system will automatically begin to schedule and execute tasks on this CPU. Tools for Managing OLAR When it is necessary to perform an OLAR operation, use the following tools which are provided as part of the SysMan suite of system management utilities. Manage CPUs "Manage CPUs" is a task-oriented application that provides the following functions: · Change the state of a CPU to online or offline · Power on or power off a CPU · Determine the status of each inserted CPU The "Manage CPUs" application can be run equivalently from an X Windows display, a terminal with curses capability, or locally on a PC (as described below), thus providing a great deal of flexibility when performing OLAR operations. Note You must be a privileged user to run the "Manage CPUs" application. Non-root users may also run the "Manage CPUs" application if they are assigned the "HardwareManagement" privilege. To assign a user the "HardwareManagement" privilege, issue the following command to launch the "Configure DOP" application: # sysman dopconfig [-display <hostname>] Please refer to the dop(8) reference page and the on-line help in the 'dopconfig' application for further information. Additionally, the Manage CPUs application provides online help capabilities that describe the operation of this application. The "Manage CPUs" application can be invoked using one of the following methods: · SysMan Menu 1. At the command prompt in a terminal window, enter the following command: [Note that the "DISPLAY" shell environment variable must be set, or the "-display" command line option must be used, in order to launch the X Windows version of SysMan Menu. If there is no indication of which graphics display to use, or if invoking from a character cell terminal, then the curses version of SysMan Menu will be launched.] # sysman [-display <hostname>] 2. Highlight the "Hardware" entry and press "Select" 3. Highlight the "Manage CPUs" entry and press "Select" · SysMan command line accelerator To launch the Manage CPUs application directly via the command prompt in a terminal window, enter the following command: # sysman hw_manage_cpus [-display hostname] [Note that the "DISPLAY" shell environment variable must be set, or the "-display" command line option must be used, in order to launch the X Windows version of Manage CPUs. If there is no indication of which graphics display to use, or if invoking from a character cell terminal, then the curses version of Manage CPUs will be launched.] · System Management Station To launch the Manage CPUs application from the System Management Station, do the following: 1. At the command prompt in a terminal window from a system that supports graphical display, enter the following command: # sysman -station [-display hostname] When the System Management Station launches, two separate windows will appear. One window is the Status Monitor view, and the other window is the Hardware view, providing a graphical depiction of the hardware connected to your system. 2. Select the Hardware view window. 3. Select the CPU for an OLAR operation by left-clicking once with the mouse. 4. Select Tools from the menu bar, or right-click once with the mouse. A list of menu options will appear. 5. Select Daily Administration from the list. 6. Select the Manage CPUs application. · Manage CPUs from a PC or Web Browser You can also perform OLAR management from your PC desktop or from within a web browser. Specifically, you can run Manage CPUs via the System Management Station client installed on your desktop, or by launching the System Management Station client from within a browser pointed to the Tru64 UNIX System Management home page. For a detailed description of options and requirements, visit the Tru64 UNIX System Management home page, available from any Tru64 UNIX system running V5.1A (or higher), at the following URL: http://hostname:2301/SysMan_Home_Page where "hostname" is the name of a Tru64 UNIX Version 5.1B, (or higher) system. hwmgr Command Line Interface (CLI) In addition to its set of generic hardware management capabilities, the hwmgr(8) command line interface incorporates the same level of OLAR management functionality as the Manage CPUs application. You must be root to run the hwmgr command; this command does not currently operate with DOP. The following describes the OLAR specific commands supported by hwmgr. To obtain general help on the use of hwmgr, issue the command: # hwmgr -help To obtain help on a specific option, issue the command: # hwmgr -help "option" where option is the name of the option you want help on. 1. To obtain the status and state information of all hardware components the operating system is aware of, issue the following command: # hwmgr -status comp STATUS ACCESS HEALTH INDICT HWID: HOSTNAME SUMMARY STATE STATE LEVEL NAME ------------------------------------------------------------- 3: wild-one online available dmapi 49: wild-one online available dsk2 50: wild-one online available dsk3 51: wild-one online available dsk4 52: wild-one online available dsk5 56: wild-one online available Compaq Alpha Server GS160 6/731 57: wild-one online available CPU0 58: wild-one online available CPU2 59: wild-one online available CPU4 60: wild-one online available CPU6 or, to obtain status on an individual component, use the hardware id (HWID) of the component and issue the command: # hwmgr -status comp -id 58 STATUS ACCESS HEALTH INDICT HWID: HOSTNAME SUMMARY STATE STATE LEVEL NAME ------------------------------------------------------------- 58: wild-one online available CPU2 To see the complete list of options for "-status", issue the command: # hwmgr -help status 2. To view a hierarchical listing of all hardware components the operating system is aware of, issue the command: # hwmgr -view hier HWID: hardware hierarchy (!)warning (X)critical (-)inactive (see -status) ------------------------------------------------------------------------- 1: platform Compaq AlphaServer GS160 6/731 9: bus wfqbb0 10: connection wfqbb0slot0 11: bus wfiop0 12: connection wfiop0slot0 13: bus pci0 14: connection pci0slot1 o o o 57: cpu qbb-0 CPU0 58: cpu qbb-0 CPU2 This example shows that CPU0 and CPU2 are children of bus name "wfqbb0", and that their physical location is (hard) qbb-0. Note that hard QBB numbers do not change as the system partitioning changes. To quickly identify which QBB a CPU is associated with, issue the command: # hwmgr -view hier -id 58 HWID: hardware hierarchy ----------------------------------------------------- 58: cpu CPU0 qbb-0 3. To offline a CPU that is currently in the online state, issue the command # hwmgr -offline -id 58 or # hwmgr -offline -name CPU2 Note that device names are case sensitive. In this example, CPU2 must be upper case. To verify the new status of CPU2, issue the command: # hwmgr -status comp -id 58 STATUS ACCESS HEALTH INDICT HWID: HOSTNAME SUMMARY STATE STATE LEVEL NAME -------------------------------------------------------------- 58: wild-one critical offline available CPU2 Note that the offline state will be saved across future reboots of the operating system, including power cycling the system. If you want the component to return to the online state the next time the operating system is booted, use the "-nosave" switch. # hwmgr -offline -nosave -id 58 or # hwmgr -offline -nosave -name CPU2 Once again, to verify the status of CPU2, issue the command: # hwmgr -status comp -id 58 STATUS ACCESS HEALTH INDICT HWID: HOSTNAME SUMMARY STATE STATE LEVEL NAME ---------------------------------------------------------------------- 58: wild-one critical offline(nosave) available CPU2 4. To power off a CPU that is currently in the offline state, issue the command: # hwmgr -power off -id 58 or # hwmgr -power off -name CPU2 Note that a component must be in the offline state before power can be removed using hwmgr. Once power has been removed from a component, is it safe to remove that component from the system. 5. To power on a CPU that is currently powered off, issue the command: # hwmgr -power on -id 58 or # hwmgr -power on -name CPU2 6. To place a CPU online so that the operating system can start scheduling processes to run on that CPU, issue the command: # hwmgr -online -id 58 or # hwmgr -online -name CPU2 Refer to the hwmgr(8) reference page for additional information on the use of hwmgr. Component Indictment Overview Component indictment is a proactive notification from a fault analysis utility, indicating that a component is experiencing high incidence of correctable errors, and therefore should be serviced and/or replaced. Component indictment involves the process of analyzing specific failure patterns from error log entries, either immediately or over a given time interval, and recommending a component's removal. The fault analysis utility signals the running operating system that a given component is suspect, causing the operating system to distribute this information via an EVM indictment event such that interested applications, including the System Management Station, Insight Manager, and the Automatic Deallocation Facility can update their state information, as well as take appropriate action if so configured (see the discussion on Automatic Deallocation Facility below). It is possible for more than one component to be indicted simultaneously if the exact source of error cannot be pinpointed. In these cases, the most likely suspect will be indicted with a `high` probability. The next likely suspect will be indicted with a `medium` probability, and the least likely suspect will be indicted with a `low` probability. When this situation arises, the indictment events can be tied together by examining the "report_handle" variable within the indictment events. Indictment events for the same error will contain the same "report_handle" value. The indicted state of a component will persist across reboot and system initialization if no action is taken to remedy the suspect component, such as an online repair operation. Once an indictment has occurred for a given component, another indictment event will not be generated for that component unless the utility has determined, through additional analysis, that the original indictment probably should be updated. In this case, the component will be re-indicted with the new probability. Once the indicted component has been serviced, it is necessary to manually clear the indicted component state with the following hwmgr command: # hwmgr -unindict -id <hwid> where <id> is the hardware id (HWID) of the component Allowing the operator to manually clear the indicted problem state, ensures positive identification of when a replaced component is operating properly. All component indictment EVM events have an event prefix of sys.unix.hw.state_change.indicted. You may view the complete list of all possible component indictment events that may be posted, including a description of each event, by issuing the command: # evmwatch -i -f '[name sys.unix.hw.state_change.indicted]' | evmshow -t "@name" -x | more You may view the list of indictment events that have occurred by issuing the command: # evmget -f '[name sys.unix.hw.state_change.indicted]' | evmshow -t "@name" CPU modules and memory pages are currently supported for component indictment. Compaq Analyze, included as part of the Web-Based Enterprise Services (WEBES) 4.0 product (or higher), is the fault analysis utility that supports component indictment on a Tru64 UNIX (V5.1A or higher) system. The WEBES product is included as part of the Tru64 UNIX operating system distribution, and must be installed after installation of the base operating system. Please refer to the Compaq Analyze documentation, distributed with the WEBES product, for a list of AlphaServer platforms that support the component indictment feature. Automatic Deallocation Facility Overview The Automatic Deallocation Facility provides the ability to automatically take an indicted component out of service, thus providing the automated ability for the system to heal itself while furthering the reliability and availability of the system. The Automatic Deallocation Facility currently supports the ability to stop using CPUs and memory pages that have been indicted. The ability to tailor the behavior of the automatic deallocation facility can be user-controlled on both single and clustered systems, through the use of the text-based OLAR Policy Configuration files. When operating in a clustered environment, automatic deallocation policy applies to all members in a cluster by default. This is specified through the cluster-wide file /etc/olar.config.common. However, individual cluster-wide policy variables can be overridden using the member-specific configuration file /etc/olar.config. The OLAR Policy Configuration files contain configuration variables that control specific behaviors of the Automatic Deallocation Facility. Behaviors such as whether or not to enable automatic deallocation, and what times of the day automatic deallocation should be enabled can be defined. Additionally, the ability to specify a user-supplied script or executable that provides the gating factor as to whether an automatic deallocation operation should occur, can be provided as well. Automatic deallocation is supported for those platforms that support the component indictment feature, as described in the Component Indictment Overview section above. Refer to the olar.config(4) reference page for additional information about the OLAR Policy Configuration files.

SEE ALSO

Commands: sysman(8), sysman_menu(8), sysman_station(8), hwmgr(8), codconfig(8), dop(8) Files: olar.config.common(4) System Administration Configuring and Managing Systems for Increased Availability Guide

Index Index for
Section 5
Index Alphabetical
listing for O
Top of page Top of
page