This chapter explains how to operate your new system in the following sections:
Before operating your system, familiarize yourself with the safety information in the following sections:
![]() | Caution: Observe all ESD precautions. Failure to do so can result in damage to the equipment. |
Wear an SGI-approved wrist strap when you handle an ESD-sensitive device to eliminate possible ESD damage to equipment. Connect the wrist strap cord directly to earth ground.
![]() | Warning: Before operating or servicing any part of this product, read the “Safety Information” in Appendix B. |
![]() | Warning: Keep fingers and conductive tools away from high-voltage areas. Failure to follow these precautions will result in serious injury or death. The high-voltage areas of the system are indicated with high-voltage warning labels. |
![]() | Caution: Power off the system only after the system software has been shut down in an orderly manner. If you power off the system before you halt the operating system, data may be corrupted. |
![]() | Note: A lithium battery is installed on the IO9 card located in the IX-brick. |
The system console enables you to perform the following activities:
Monitor your system by reading the system's status and error message information generated and displayed by the SGI system's L1 controller.
Enter L1 and L2 controller commands to monitor or change particular system functions. You can, for example, monitor the speed of fans for a particular brick. See your SGI L1 and L2 Controller Software User's Guide for descriptions of these commands.
Power on or power off individual bricks or all bricks (except for the storage modules) in your system.
If you connect a console that contains SGIconsole software, you can perform the functions listed above and have boundary scan capabilities. Boundary scan capabilities enable an SGI system support engineer (SSE) to test the functionality of your system.
This section explains how to power on and power off individual bricks or your entire Altix 3000 series system, as follows:
For servers with an L2 controller, you can power on and power off individual bricks or the entire system at the L2 controller touch display on the rear door of the server rack.For servers with a system console, you can power on and power off individual bricks or the entire system at the system console.
If you are using an SGIconsole, you can monitor and manage your server from a remote location. You can also monitor and manage your server with tools such as VACM, Console Manager, and PCP. For details, see the documentation for the particular tool.
The Embedded Support Partner (ESP) program enables you and your SGI system support engineer (SSE) to monitor your server remotely and resolve issues before they become problems. For details on this program, see “Using Embedded Support Partner (ESP) ”.
This section describes how to prepare to power on your system, and how to power on your system by using either of the following:
L2 controller touch display
System console
To prepare to power on your system, follow these steps:
Check to ensure that the cabling between the power distribution unit (PDU) and the wall power-plug receptacle is secure.
For each individual brick that you want to power on, make sure that the PWR (power) switch is set to the 1 (on) position, as shown in Figure 1-1 for the IX–brick. This will start the L1 controller(s) for the brick(s) when the brick(s) are powered on. The 12-VDC LED on an individual brick illuminates green when an L1 controller has powered on successfully.
If you plan to power on a D–brick2, or an entire server that includes a D–brick2, make sure that the power switch on the rear of each PSU/cooling module (two per D-brick2) is in the 1 (on) position. The location of the power switch on the D–brick2 PSU/cooling module is shown in Figure 1-2.
The standard configuration of the TP900 storage module has one power supply, as shown in Figure 1-3. You can add an optional second power supply to the TP900 system to provide n+1 redundant power. The second power supply is identical to the first power supply and would be located in the lower-right section of the TP900 storage module. Both power switches must be in the 1 (on) position to provide redundant power.
Make sure that the PDU and PDS circuit breaker switches shown in Figure 1-4 are turned on to give power to the server system when the system is powered on.
This section describes how to power on individual bricks or the entire system from the L2 controller touch display (rack display) shown in Figure 1-5, which is located on the rear door of rack 001. If you have multiple compute racks whose L2 controllers are interconnected at an Ethernet hub, you can power on any brick in those racks or the entire system.
For instructions on navigating the L2 controller touch display, see “Using the L2 Controller Touch Display” in Chapter 2.
To power on selected bricks or the entire server, follow these steps:
Select Power UP from the home window shown in Figure 1-6.
The Power UP confirmation window appears, as shown in Figure 1-7.
To power on the system, press the OK button.
If the command is not executed successfully, a Command Error/Timeout window appears, which shows the error message. Press Cancel and then return to the home window.
![]() | Note: Verify that the LEDs turn on and light green and that your controllers display that the system is powered on for each segment of the procedure, which indicates that the power-on procedure is proceeding properly. If you have a problem while powering on and an error message appears on the L1 controller, the L2 controller touch display, or the system console, see your online log files and the information in “L1 Controller Error Messages” in Chapter 13 to learn what the error message indicates and how to resolve the problem. |
The power-on procedure at a system console varies with your server setup, as follows:
If you have a system console without a hardware L2 controller, you can toggle between L1 and console mode. This enables you to power on your server with L1 commands and view the activity by changing to the console mode.
If you have a system console connected to a server with a hardware L2 controller, you can toggle between L2, L1, and console mode, and power on your server with L1 or L2 controller commands, and monitor the power-on activity by changing to the console mode.
For detailed instructions on using a system console running L2 software, see “About the L2 Controller Firmware” in Chapter 2. For detailed instructions on using the L2 mode, see “Operating the L2” in Chapter 2. For detailed instructions on using a system console using the L1 mode, see “Operating the L1” in Chapter 2.
The following sections describe how to power on your system in either the L2 mode or the L1 mode.
To power on your system while in the L2 mode, follow these steps:
At your console, switch to the L2 mode by entering the following command:
$> Ctrl+T |
From the L2 prompt (L2>), power on an individual brick by entering the following command. (If you want to power on the entire server, proceed to step 3.)
L2> r <rack#> s <slot#> pwr u |
For example, to power on a C–brick in rack 1, slot 10, enter the following:
L2> r 1 s 10 pwr u |
The slot number is the unit position number located on the rack, slightly above where the bottom of the brick sits. Each rack unit position number is located toward the top of the two lines that mark off the unit position that the number represents. For example, the rack numbering for a brick in slot 10 would appear on the left-front side of the rack, as shown in Figure 1-8:
If you want to power on several selected bricks from a rack at the same time, you must enter the rack number followed by the slot numbers of the bricks you want to power on. For example, to power on bricks in slots 7 and 10 for rack 4, enter the following:
L2> r 4 s 7,10 pwr u |
If you want to power on a brick for several racks, you must enter the number of the racks followed by the slot number of the brick you want to power on for all the racks. For example, to power on the brick in slot 10 for racks 3 and 4, enter the following:
L2> r 3, 4 s 10 pwr u
![]() | Note: To avoid problems with your system, do not try to power on multiple slots for multiple racks at the same time. |
If you want to power on the entire server, enter the following command:
L2> pwr u |
(The default setting for the pwr u command is all racks and all slots.)
From the L2 prompt, display the system configuration by entering the following command:
L2> config |
This command lists the bricks in the system and each brick's system controller address.
The L1 controller display, located on the front of each brick, should display L1 running once the power-on procedure starts.
![]() | Note: If you have a problem while powering on and an error message appears on your console display, see “L1 Controller Error Messages” in Chapter 13 to learn what the error message indicates and how to resolve the problem. |
![]() | Note: You do not want to power on from the L1 mode if a hardware L2 controller is available on your system. |
To power on your system while in the L1 mode, follow these steps:
The prompt on your system shows the rack and slot number of the C–brick to which you have connected your console. If you want to power on the C–brick (001c05 in our example) indicated in the prompt, enter the following command. (If you want to power on the bricks connected to the C–brick, proceed to the next step.)
001c05-L1> power up |
If you want to power on the bricks connected to the C–brick, enter the following command:
oo3c01-L1> * power up (* indicates all) |
From the L1 prompt, display the brick configuration information by entering the following command:
001c05-L1> config |
In L1 mode, you can obtain only limited information about the system configuration. A C–brick only has information about its attached I/O brick and, if another C–brick is attached to the I/O brick, information about that C–brick and its attached I/O brick. An I/O brick only has information about its attached C–brick and an R-brick only has information about itself.
The L1 controller display, located on the front of each brick, should display L1 running once the power-on procedure starts.
You can power off individual bricks or your entire server system from the L2 controller touch display (located on the rear door of rack 001) or from the server system console, as explained in the sections that follow.
![]() | Note: Verify that the LEDs turn off and that your L1 controllers display that the system is powering off for each segment of the procedure, which indicates that the power-off procedure is proceeding properly. If you have a problem while powering off and an error message appears on your L2 controller touch display or your console, see your online log files and the information in “L1 Controller Error Messages” in Chapter 13 to learn what the error message indicates and how to resolve the problem. |
This section describes how to power off individual bricks or the entire system from the L2 controller touch display. If you have multiple racks whose L2 controllers are interconnected at an Ethernet hub, you can power on any brick in those racks or the entire system at the L2 controller touch display on the rear door of rack 001.
For instructions on navigating the L2 controller touch display, see “Using the L2 Controller Touch Display” in Chapter 2.
To power off an individual brick or your entire server, follow these steps:
Select Power DOWN from the home window shown in Figure 1-9.
The Power DOWN confirmation window appears, as shown in Figure 1-10.
To power off the system, press the OK button.
If the command is not executed successfully, a Command Error/Timeout window appears, which shows the error message. Press Cancel and then return to the home window.
The power-off procedure at a system console varies with your server setup, as follows:
If you have a system console connected to a server with a hardware L2 controller, you can toggle between L2, L1, and console mode, power off your server with L1 or L2 controller commands, and monitor the power-off activity by changing to the console mode.
If your system console is running on a server that has no hardware L2 controller, you can toggle between L1 and console mode. This enables you to power off your server with L1 commands and view the activity by changing to the console mode.
For detailed instructions on using a system console running L2 software, see “About the L2 Controller Firmware” in Chapter 2. For detailed instructions on using the L2 mode, see “Operating the L2” in Chapter 2. For detailed instructions on using a system console in the L1 mode, see “Operating the L1” in Chapter 2.
The following sections describe how to power off your system in the L2 mode and the L1 mode.
To power off your system while in the L2 mode, follow these steps:
At your system console, switch to the L2 mode (L2 software) by entering the following command:
$> Ctrl+T |
From the L2 prompt (L2>), power off an individual brick by entering the following command. (If you want to power off the entire server, proceed to step 3.)
L2> r <rack#> s <slot#> pwr d |
For example, to power off a C–brick in rack 1, slot 10, enter the following:
L2> r 1 s 10 pwr d |
The slot number is the unit position number located on the rack, slightly above where the bottom of the brick sits. Each rack unit position number is located toward the top of the two lines that mark off the unit position the number represents. For example, the rack numbering for a brick in slot 10 would appear on the left-front side of the rack as shown in Figure 1-11:
If you want to power off several selected bricks from a rack at the same time, you must enter the rack number followed by the slot numbers of the bricks you want to power off. For example, to power off bricks in slots 7 and 10 for rack 4, enter the following:
L2> r 4 s 7,10 pwr d |
If you want to power off a brick for several racks, you must enter the number of the racks followed by the slot number of the brick you want to power off for each rack. For example, to power off the brick in slot 10 for racks 3 and 4, enter the following:
L2> r 3, 4 s 10 pwr d
![]() | Note: To avoid problems with your system, do not try to power off several slots for several racks at the same time. |
If you want to power off all the bricks in all the racks, enter the following command:
L2> pwr d (The default setting for this command is all racks and all slots.) |
From the L2 prompt, display the brick configuration information by entering the following command:
L2> config |
This command lists all the bricks in the system and each brick's system controller address.
The L1 controller display for each brick should display L1 running once the power-off procedure starts.
![]() | Note: If you have a problem while powering off and an error message appears on your console display, see “L1 Controller Error Messages” in Chapter 13 to learn what the error message indicates and how to resolve the problem. |
![]() | Note: You do not want to power off from the L1 mode if L2 software is available on your system. |
To power off your system while in the L1 mode, follow these steps:
The prompt on your system shows the rack and slot number of the C–brick to which you have connected your console. If you want to power off the C–brick (001c05 in our example) indicated in the prompt, enter the following command. (If you want to power off the bricks connected to the C–brick, proceed to the next step.)
001c05-L1> power down |
If you want to power off the bricks connected to the C–brick, enter the following command:
001c05-L1> * power down (* indicates all) |
From the L1 prompt, display the brick configuration information by entering the following command:
001c05-L1> config |
In L1 mode, you can obtain only limited information about the system configuration. A C–brick only has information about its attached I/O brick and, if another C–brick is attached to the I/O brick, information about that C–brick and its attached I/O brick.
An I/O brick only has information about its attached C–brick and an R-brick only has information about itself.
Embedded Support Partner (ESP) automatically detects system conditions that indicate potential future problems and then notifies the appropriate personnel. This enables you and SGI system support engineers (SSEs) to proactively support systems and resolve issues before they develop into actual failures.
ESP enables users to monitor one or more systems at a site from a local or remote connection. ESP can perform the following functions:
Monitor the system configuration, events, performance, and availability.
Notify SSEs when specific events occur.
Generate reports.
ESP also supports the following:
Remote support and on-site troubleshooting.
System group management, which enables you to manage an entire group of systems from a single system.
You can monitor your Altix 3000 server from the following sources:
On the L1 controller's display at the right-hand corner of the front of each brick (except the TP900 and D-brick2), as shown in Figure 1-12, you can monitor brick items. For example, you can see if the fans on a particular brick are operating properly.
If your server has an L2 controller, you can connect the SGIconsole to the Ethernet port on the L2 controller. If you have multiple L2 controllers, connect the SGIconsole to the Ethernet hub that is connected to the Ethernet port of the multiple L2 controllers. You will need to connect either a local or remote workstation to monitor the servers as shown in Figure 1-13.
These console connections to the L2 controller enable you to view the status and error messages generated by both the L1 controllers and the L2 controller on your server rack. You can also use these consoles to input L1 and L2 commands to manage and monitor your system.
![]() | Note: The SGIconsole enables you to use various software tools, such as VACM, Console Manager, and PCP, to manage and monitor your system. See the SGIconsole Start Here guide for descriptions of these tools and for references to other documents with information about these tools. |
If your system does not have an L2 controller, you can connect a dumb terminal to the C–brick console port. If you have multiple C-bricks, connect the dumb terminal to the bottom-most C–brick in the rack, as shown in Figure 1-14. This connection enables you to view the status and error messages generated by the L1 controller and to enter L1 commands to manage and monitor your system.
Besides adding a system console, you can add or replace the following hardware items on your Altix 3000 series server:
Peripheral component interface (PCI) cards into your IX-brick or PX-brick.
Disk drives to your IX-brick, TP900 disk storage module, or D-brick2 storage module.
The sections that follow discuss these activities in more detail.
![]() | Warning: You can add or replace only the items listed in this section. For your safety and for the protection of your server system, contact your SGI system support engineer (SSE) to install any hardware items not listed in this section. |
![]() | Warning: Before installing, operating, or servicing any part of this product, read the “Safety Information” in Appendix B. |
The PCI-X based I/O system, an industry standard for connecting peripherals to a processor, is the primary I/O system for the Altix 3000 series server.
The IX-brick and PX-brick subsystems provide PCI and PCI-X support for the Altix 3000 server system. The IX-brick provides 12 PCI slots. One of the 12 slots is reserved for a system PCI card. If additional PCI slots are required beyond the 11 remaining slots in the IX-brick, a PX-brick is configured into the system.
See “Adding or Replacing a PCI or PCI–X Card” in Chapter 12 for detailed instructions on installing or removing PCI cards.
The IX-brick supports the system boot functions and has one or two low-profile SCSI disk drives. See “Installing or Replacing a Disk Drive in the IX-brick ” in Chapter 12 for detailed instructions on installing or removing disk drives.
The TP900 is a SCSI-based disk storage module that provides JBOD (just a bunch of disks) mass storage and that has 8 disk drives in each storage module. See “Replacing a TP900 Drive Carrier Module” in Chapter 12 for detailed instructions on installing or removing disk drives.
The D-brick2 is a Fibre Channel-based disk storage enclosure that provides JBOD mass storage and that has 16 disk drives in each storage enclosure. See “Installing or Replacing a D-brick2 Drive Carrier Module” in Chapter 12 for detailed instructions on installing or removing disk drives.