The TPF system is loaded to the online system in 2 steps:
- The online keypoints and the file resident and main storage resident
program segments are loaded to the loader general file by the system loader
offline segment. See TPF System Installation
Support Reference for more information.
- The online keypoints and the file resident and main
storage resident program segments are loaded from the loader general file to
the online files by IPLing the loader general file.
- Note:
- The loader general file and the online files must be initialized by ICKDSF,
an MVS utility, to include the IPL program
and be formatted with the real-time disk formatter (FMTR). See
TPF Database Reference for more information about
FMTR.
The TPF system is started by IPLing the prime (or backup) online system
pack. See TPF Main Supervisor Reference for
more information about the initial program load (IPL).
All of the processors in a loosely coupled complex are IPLed from the same
online files. The first IPLed processor sets the time-of-day (TOD)
clock value for the complex.
Thereafter, when the remaining processors are IPLed, the TOD clocks are
synchronized with the TOD clock of the first processor or with the Sysplex
Timer, if it is available.
Installations that use the multiple database function (MDBF) configure all
subsequent processors from 2 sources: processor-unique keypoints and the
MDBF configuration supplied by the first processor through shared
keypoints. To IPL the first processor in a complex:
- Perform a hardware IPL for the processor.
- When you are prompted, enter the TPF logical CPU ID of the processor that
is being IPLed. TPF logical processor IDs are chosen by the system
programmer during the system generation process.
- When you are prompted, specify the type of IPL that you want to
perform.
- If you specified an IPL with image selection, enter the name of the TPF
image when you are prompted.
- When you are prompted, enter the MDBF configuration information for the
complex (if MDBF was generated in the system).
The system is now cycled to 1052 state.
- Set the system interprocessor status variables using the ZSIPC
command. The Multi-Processor Interconnect Facility (MPIF) system and
complex variables must be set and at least one device, path class, and path
must be defined using the ZMPIF command.
- Note:
- It is recommended that you dedicate a path class to MPIF IPC and that you
define this path class as a protected class.
- Perform either a software or hardware IPL. A software IPL will be
completed without operator interaction.
The system is placed in 1052 state. Other processors can now be
added to the complex.
After the first processor is IPLed, all the other processors can be added
to the complex using the following procedure.
- Perform a hardware IPL for the processor.
- When you are prompted, enter the TPF logical CPU ID of the TPF system that
you want to IPL.
- Note:
- Only one physical CPU can have a particular TPF logical CPU ID at any one
time. When adding a processor to a complex, use a logical ID that is
not already active in the complex.
- The MPIF system name must be set and at least one path must be defined
using the ZMPIF command. Perform either a software or hardware
IPL. A software IPL will be completed without operator
interaction.
- The IPL procedure will recognize that other processors in the complex
exist and use their MDBF configuration if MDBF was generated in the
system.
- While the system is in restart mode, the system will force a
software-initiated IPL of the processor to synchronize the processor-shared
tables with those of the first processor. This process is completed
without operator intervention.
The IPL is complete when the system is placed in 1052 state.
If 2 or more TPF processors are IPLed at the same time, each one will be
unaware of the other until they go through MPIF restart.
If the TPF systems that are being IPLed are currently active in the loosely
coupled complex, the IPLs will continue correctly, generally without operator
intervention.
If the TPF systems that are being IPLed are not currently active, but are
joining a currently active TPF loosely coupled complex, they will go through a
forced software IPL that will bring them into the complex. The software
IPLs will ensure that the TPF processor ID you specified is unique and that
all keypoint data is current. This forced IPL is the normal mechanism
for adding processors to an active TPF loosely coupled complex. (See
the previous procedure for more information.)
If there is no currently active loosely coupled complex (all TPF processors
are deactivated), each system that is being IPLed will perform as if it is the
first TPF system to be IPLed. If more than one TPF system performs as
if it is the first system being IPLed, confusing and occasionally
contradictory messages may be displayed on the consoles of the system being
IPLed. With MDBF, each TPF system will request the MDBF configuration,
possibly getting different responses from you. With the Record Cache
RPQ, each TPF system will attempt to
reinitialize the record cache and reset the record cache operational
parameters, possibly causing I/O errors on the other systems attempting to
IPL.
There are a series of checks during system restart to find keypoint
conditions that are out of sync. These conditions can cause database
damage. The checks will also show if more than one processor determined
that it is the first processor in the complex, and will remove all but one of
the processors that are being IPLed. These processors can then be
brought into the loosely coupled complex by performing an additional
IPL.
When creating a TPF loosely coupled complex (for example, IPLing the first
TPF processor), do not IPL additional processors until the first processor is
in 1052 state. Once the TPF complex is created, new processors can be
added at any time.
- Note:
- Adding TPF processors to a loosely coupled complex places an I/O load against
the module that is being IPLed. Installations that are very close to
the DASD I/O access rate limit on the IPL device should IPL sequentially to
reduce device contention.
The following information lists and explains the program status words
(PSWs) for the IPL program wait states.
- PSW
- Explanation
- 000A0000 000003FF
- Loaded by IPLB at the normal end of dump processing.
- 000A0000 000000FB
- Loaded by IPLB when a CC3 is returned from a set clock (SCK) instruction
indicating that the TOD clock is not operational.
- 000A0000 0000DEAD
- Loaded by IPLB when an irrecoverable I/O error is found during dump
processing.
- 000A0000 00DEAD00
- Loaded by IPL2, IPLA, and IPLB when an unexpected machine check interrupt
is found.
- 000A0000 00FFFFF0
- Loaded by IPL2, IPLA, and IPLB when an unexpected external interrupt is
found.
- 000A0000 00FFFFF2
- Loaded by IPL2, IPLA, and IPLB when an unexpected supervisor call
interrupt is found.
- 000A0000 00FFFFF4
- Loaded by IPL2, IPLA, and IPLB when an unexpected program check interrupt
is found.
- 000A0000 00FFFFF8
- Loaded by IPLA and IPLB when an unexpected I/O interrupt is found.
It is loaded by IPL2 when an irrecoverable error, such as CC3 on a start
subchannel (SSCH) or test subchannel (TSCH), is found before the I/O
interrupt.
- 000A0000 00xxxxF6
- Loaded by IPL2 when an incorrect record ID is found in the IPLA chain,
where xxxx is the ID of the incorrect record.
- 000A0000 00xxxxF8
- Loaded by IPL2 when an irrecoverable I/O error is detected, where
xxxx is the cumulative device and subchannel status from the failing
device.
- 000A0000 FFFFFFFA
- Loaded by IPLB (IB01) when storage for the prefix pages, common I/O (CIO),
or the CIO work areas cannot be allocated without overlaying something that
was loaded by IPLA.
- 000A0000 FFFFFFFC
- Loaded by IPLB (IB01) when fast recovery objects overlap.
- 000A0000 FFFFFFFE
- Loaded by IPLB (IB01) when storage for the IPL restart area, the segment
table, or the page tables cannot be allocated without overlaying common I/O
(CIO).
- 000A0000 xxxxxxxx
- Loaded by IPLB when a program calls the dump routine a second time while a
dump is in progress. The second call can be a result of a program check
or a branch from an IPLB error routine. If a program check occurred,
the program interruption code in the prefix page indicates the type of
exception. The addresses of the page and segment tables and the
register save area are contained in the IPL restart area, where
xxxxxxxx is the address of the IPL restart area (defined by the IB5CT
DSECT in IB0CT). The page and segment tables and the register save area
can be dumped using the stand-alone dump (SADUMP) utility.
The following information lists and explains the PSWs for the system
initialization wait states.
- PSW
- Explanation
- 040AC000 00FFFFFF
- Loaded by CCCTIN to indicate the normal end of the CCCTIN micro dump or to
indicate that an irrecoverable error was detected while attempting to write to
the system console.
- 000A0000 C3E3F8F5
- Loaded by CT85 when an application I-stream (for example, the MPIF
I-stream) fails to restart. IPL the TPF system to recover.
The following information lists and explains the PSWs for the system error
program wait states. The low-order word of each disabled wait PSW
loaded by CPSE does not address an instruction but, instead, contains a unique
3-byte identifier:
- PSW
- Explanation
- 040AC000 00CCD01A
- Catastrophic error on a general file (GF) IPL. A dump was taken,
but no recovery is attempted.
- 000AC000 00CCD01E
- System virtual memory (SVM) is not stable; unable to switch from real
mode to home address space. Inspect the control registers.
- 000AC000 00CCD01F
- CCCPSF recursion failed; unable to continue.
- 040AC000 00CCD09E
- Program error in CPSE on an application I-stream.
- 000A0000 00DEAD09
- An unexpected first-level interrupt handler (FLIH) machine check
occurred.
- 040A0000 00CCD708
- CFLF restart failed on a processor other than the first in the
complex.
The following information lists and describes the disabled wait states
related to the machine check interruption handler. The low-order 3
bytes of each PSW contains an identifier.
- Note:
- Recover from these disabled wait states by performing an IPL with the CLEAR
option.
- PSW
- Explanation
- 000A0000 00DEAD01
- The machine check FLIH was interrupted twice by machine check
interruptions (for example, double machine check recursion).
- 000A0000 00DEAD02
- A machine check interruption reported channel subsystem damage.
- 000A0000 00DEAD03
- A machine check interruption on an application I-stream reported
catastrophic damage while system error recovery was active on the main
I-stream, or a check-stop condition was reported to an application I-stream
while system error recovery was active on the main I-stream.
- 000A0000 00DEAD04
- While handling a machine check interruption that was reporting
catastrophic damage, the machine check FLIH could not perform a SIGP STOP for
one of the I-streams; or while handling a check-stop condition, the
check-stop handler was unable to perform a SIGP STOP for one of the
I-streams.
- 000A0000 00DEAD05
- While handling a machine check interruption that was reporting
catastrophic damage, the machine check FLIH could not perform a SIGP SENSE for
one of the I-streams; or while handling a check-stop condition, the
check-stop handler could not perform a SIGP SENSE for one of the
I-streams.
- 000A0000 00DEAD06
- While handling a machine check interruption that was reporting
catastrophic damage, the machine check FLIH determined that the main I-stream
had performed a check stop; or while handling a check-stop condition, the
check-stop handler determined that the main I-stream had performed a check
stop.
- 000A0000 00DEAD07
- While handling a machine check interruption that was reporting
catastrophic damage, the machine check FLIH could not perform a SIGP RESTART
for the main I-stream; or while handling a check-stop condition, the
check-stop handler could not perform a SIGP RESTART for the main
I-stream.
- 000A0000 00DEAD08
- While handling a machine check interruption that was reporting
catastrophic damage, the machine check FLIH was interrupted by a machine
interruption on the main I-stream.
- 000A0000 00DEAD09
- An unexpected FLIH machine check occurred.