Redp 5649
Redp 5649
Redpaper
IBM Redbooks
November 2024
REDP-5649-01
Note: Before using this information and the product it supports, read the information in “Notices” on
page vii.
© Copyright International Business Machines Corporation 2021, 2024. All rights reserved.
Note to U.S. Government Users Restricted Rights -- Use, duplication or disclosure restricted by GSA ADP Schedule
Contract with IBM Corp.
Contents
Notices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vii
Trademarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . viii
Preface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ix
Authors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ix
Now you can become a published author, too! . . . . . . . . . . . . . . . . . . . . . . . . . x
Comments welcome. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . x
Stay connected to IBM Redbooks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xi
Contents v
Abbreviations and acronyms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 161
This information was developed for products and services offered in the US. This material might be available
from IBM in other languages. However, you may be required to own a copy of the product or product version in
that language in order to access it.
IBM may not offer the products, services, or features discussed in this document in other countries. Consult
your local IBM representative for information on the products and services currently available in your area. Any
reference to an IBM product, program, or service is not intended to state or imply that only that IBM product,
program, or service may be used. Any functionally equivalent product, program, or service that does not
infringe any IBM intellectual property right may be used instead. However, it is the user’s responsibility to
evaluate and verify the operation of any non-IBM product, program, or service.
IBM may have patents or pending patent applications covering subject matter described in this document. The
furnishing of this document does not grant you any license to these patents. You can send license inquiries, in
writing, to:
IBM Director of Licensing, IBM Corporation, North Castle Drive, MD-NC119, Armonk, NY 10504-1785, US
This information could include technical inaccuracies or typographical errors. Changes are periodically made
to the information herein; these changes will be incorporated in new editions of the publication. IBM may make
improvements and/or changes in the product(s) and/or the program(s) described in this publication at any time
without notice.
Any references in this information to non-IBM websites are provided for convenience only and do not in any
manner serve as an endorsement of those websites. The materials at those websites are not part of the
materials for this IBM product and use of those websites is at your own risk.
IBM may use or distribute any of the information you provide in any way it believes appropriate without
incurring any obligation to you.
The performance data and client examples cited are presented for illustrative purposes only. Actual
performance results may vary depending on specific configurations and operating conditions.
Information concerning non-IBM products was obtained from the suppliers of those products, their published
announcements or other publicly available sources. IBM has not tested those products and cannot confirm the
accuracy of performance, compatibility or any other claims related to non-IBM products. Questions on the
capabilities of non-IBM products should be addressed to the suppliers of those products.
Statements regarding IBM’s future direction or intent are subject to change or withdrawal without notice, and
represent goals and objectives only.
This information contains examples of data and reports used in daily business operations. To illustrate them
as completely as possible, the examples include the names of individuals, companies, brands, and products.
All of these names are fictitious and any similarity to actual people or business enterprises is entirely
coincidental.
COPYRIGHT LICENSE:
This information contains sample application programs in source language, which illustrate programming
techniques on various operating platforms. You may copy, modify, and distribute these sample programs in
any form without payment to IBM, for the purposes of developing, using, marketing or distributing application
programs conforming to the application programming interface for the operating platform for which the sample
programs are written. These examples have not been thoroughly tested under all conditions. IBM, therefore,
cannot guarantee or imply reliability, serviceability, or function of these programs. The sample programs are
provided “AS IS”, without warranty of any kind. IBM shall not be liable for any damages arising out of your use
of the sample programs.
The following terms are trademarks or registered trademarks of International Business Machines Corporation,
and might also be trademarks or registered trademarks in other countries.
AIX® IBM FlashSystem® Power8®
Db2® IBM Spectrum® Power9®
Easy Tier® IBM Z® PowerVM®
IBM® Micro-Partitioning® Redbooks®
IBM Cloud® POWER® Redbooks (logo) ®
IBM FlashCore® Power Architecture®
The registered trademark Linux® is used pursuant to a sublicense from the Linux Foundation, the exclusive
licensee of Linus Torvalds, owner of the mark on a worldwide basis.
LTO, Ultrium, the LTO Logo and the Ultrium logo are trademarks of HP, IBM Corp. and Quantum in the U.S.
and other countries.
Red Hat, Ansible, OpenShift, are trademarks or registered trademarks of Red Hat, Inc. or its subsidiaries in
the United States and other countries.
Other company, product, or service names may be trademarks or service marks of others.
This IBM Redpaper publication provides a broad understanding of a new architecture of the
IBM Power E1080 (also known as the Power E1080) server that supports IBM AIX®, IBM i,
and selected distributions of Linux operating systems. The objective of this paper is to
introduce the Power E1080, the most powerful and scalable server of the IBM Power portfolio,
and its offerings and relevant functions:
Designed to support up to four system nodes and up to 240 IBM Power10 processor cores
The Power E1080 can be ordered as a 1-, 2-, 3-, or four-node configuration with each
node providing up to 60 Power10 processor cores to a maximum 240 Power10 processor
cores in a 4-node configuration.
Designed to support up to 64 TB of memory
The Power E1080 can be initially ordered with the total memory capacity of up to 8 TB. It
supports up to 64 TB in a full combined four-system nodes server.
Designed to support up to 32 Peripheral Component Interconnect Express (PCIe) Gen 5
slots in a full combined four-system nodes server, and up to 192 PCIe Gen 3 slots with
expansion I/O drawers
Up to over 4,000 directly attached serial-attached SCSI (SAS) disks or solid-state drives
(SSDs)
Up to 1,000 virtual machines (VMs) with logical partitions (LPARs) per system
System control unit (SCU), providing redundant system master Flexible Service Processor
(FSP)
Supports IBM Power Private Cloud Solution with Dynamic Capacity
This publication is for professionals who want to acquire a better understanding of Power
servers. The intended audience includes the following roles:
Customers
Sales and marketing professionals
Technical support professionals
IBM Business Partners
Independent software vendors (ISVs)
This paper does not replace the current marketing materials and configuration tools. It is
intended as an extra source of information that, together with existing sources, can be used to
enhance your knowledge of IBM® server solutions.
Authors
This paper was produced by a team of specialists from around the world working at
IBM Redbooks, Poughkeepsie Center.
Tim Simon is an IBM Redbooks® Project Leader who is based in Tulsa, Oklahoma, US. He
has over 40 years of experience with IBM, primarily in a technical sales role working with
customers to help them create IBM solutions to solve their business problems. He holds a BS
degree in Math from Towson University in Maryland. He has worked with many IBM products
and has extensive experience creating customer solutions by using IBM Power, IBM Storage,
and IBM Z® throughout his career.
Tsvetomir Spasov is a Power Servers Hardware Product Engineer in Sofia, Bulgaria. He has
8 years of experience with IBM in RTS, SME, and PE roles. His main area of expertise is
Hardware Management Console (HMC), FSP, eBMC, POWERLC, and GTMS. He holds a
masters degree in Electrical Engineering from Technical University of Sofia.
Authors of the first edition, IBM Power E1080 Technical Overview and Introduction,
REDP-5649, published in October 2021, were:
Scott Vetter, Giuliano Anselmi, Manish Arora, Ivaylo Bozhinov, Dinil Das, Turgut Genc,
Bartlomiej Grabowski, Madison Lee, Armin Röllr
Here’s opportunity to spotlight your skills, grow your career, and become a published author—
all at the same time! Join an IBM Redbooks residency project and help write a book in your
area of expertise, while honing your experience using leading-edge technologies. Your efforts
will help to increase product acceptance and customer satisfaction, as you expand your
network of technical contacts and relationships. Residencies run from two to six weeks in
length, and you can participate either in person or as a remote resident working from your
home base.
Find out more about the residency program, browse the residency index, and apply online at:
[Link]/redbooks/[Link]
Comments welcome
Your comments are important to us!
We want our papers to be as helpful as possible. Send us your comments about this paper or
other IBM Redbooks publications in one of the following ways:
Use the online Contact us review Redbooks form found at:
[Link]/redbooks
Send your comments in an email to:
redbooks@[Link]
Mail your comments to:
IBM Corporation, IBM Redbooks
Dept. HYTD Mail Station P099
2455 South Road
Poughkeepsie, NY 12601-5400
Preface xi
xii IBM Power E1080 Technical Overview and Introduction
Summary of changes
This section describes the technical changes that were made in this edition of the paper and
in previous editions. This edition might also include minor corrections and editorial changes
that are not identified.
New information
Added information on Double Data Rate 5 (DDR5)-based differential dual inline memory
modules (DDIMMs), which provide increased memory throughput and reduced latency
compared to the Double Data Rate 4 (DDR4)-based memory that originally was delivered
with IBM Power10.
Added information about the Peripheral Component Interconnect Express (PCIe) Gen 4
Expansion drawer, which replaces the previous PCIe Gen 3 Expansion drawer. This
expansion drawer can hold more PCIe adapters for your system, and provides improved
throughput over the previous version.
Added information about the NED24 Non-Volatile Memory Express (NVMe) Expansion
drawer. This new expansion drawer provides more NVMe slots for U2 NVMe devices.
Each NED24 expansion drawer can provide up to 153 TB of capacity. Up to three NED24
drawers are supported per system node, providing up to 460 TB of capacity per drawer or
over 1840 TB per system.
This section also includes information about the multipath support in the NED24
expansion drawer.
Added information about new adapters that are supported in the Power E1080.
Changed information
Noted that the PCIe Gen 3 Expansion drawer is withdrawn from marketing. The
replacement is the PCIe Gen 4 expansion drawer.
Noted that the EXP24SX serial-attached SCSI (SAS) expansion drawer is withdrawn from
marketing. The recommended replacement is the NED24 expansion drawer because it
generally provides better performance and often lower costs compared to equivalent SAS
storage options.
The Power E1080 simplifies end-to-end encryption and brings AI where your data is stored
for faster insights. This configuration helps enable greater workload deployment flexibility and
agility while accomplishing more work.
The Power E1080 can help you to realize the following benefits:
Protect trust from core to cloud
Protect data that is in-transit and at-rest with greatly simplified end-to-end encryption
across hybrid cloud without affecting performance.
Enjoy enterprise quality of service (QoS)
The Power E1080 can detect, isolate, and recover from soft errors automatically in the
hardware without taking an outage or relying on an operating system to manage the faults.
Drive greater efficiency with sustainable and scalable compute
The processor performance, massive system throughput, and memory capacity qualify the
Power E1080 server to be the perfect workload consolidation platform. This performance
leads to significant savings in floor space, energy consumption, and operational
expenditure costs.
A system node is an enclosure that provides the connections and supporting electronics to
connect the processor with the memory, internal disk, adapters, and the interconnects that
are required for expansion. A combination of one, two, three, or four system nodes per server
is supported.
Each system node provides four sockets for Power10 processor chips and 64 differential dual
inline memory module (DDIMM) slots for Double Data Rate 5 (DDR5) or Double Data Rate 4
(DDR4) technology DIMMs.
Each socket holds one Power10 single-chip module (SCM). An SCM can contain 10, 12, or
15 Power10 processor cores. It also holds the extra infrastructure logic to provide electric
power and data connectivity to the Power10 processor chip.
A 4-node Power E1080 server scales up to 16 processor sockets and 160, 192, or 240 cores,
depending on the number of cores that is provided by the configured SCM type.
The processor configuration of a system node is defined by the selected processor feature.
Each feature defines a set of four Power10 processors chips with the same core density (10,
12, or 15). All system nodes within a Power E1080 server must be configured with the same
processor feature.
Each system node can support up to a maximum of 16 TB of system memory by using the
largest available memory DIMM. A fully configured 4-node Power E1080 can support up to
64 TB of memory.
To provide internal boot capability, each system node enables up to four Non-Volatile Memory
Express (NVMe) drive bays. More drive bays can be configured by using expansion drawers.
The system control unit (SCU) is required in all configurations. The SCU provides system
hardware, firmware, and virtualization control through redundant Flexible Service Processors
(FSPs). Only one SCU is required and supported for every Power E1080 server. For more
information about the SCU, see 1.3, “System control unit” on page 10.
For more information about the environmental and physical aspects of the server, see 1.4,
“Server specifications” on page 12.
An expansion drawer provides more PCIe slots for I/O connectivity if needed. There are two
versions of the PCIe expansion drawer: the original version supporting PCIe Gen 3 I/O slots1,
and a newer version supporting PCIe Gen 4 I/O slots. There is no upgrade path from the
Gen3 version to the Gen4 version, but you may mix the two I/O expansion drawers within a
single system.
The I/O expansion drawer is a 19-inch PCIe 4U enclosure that provides up to 12 PCIe slots
and connects to the system node with a pair of PCIe x16 to CXP converter cards (fanout
adapters) that are housed in the system node. Each system node can support up to three
Gen 4 I/O expansion drawers or four Gen 3 I/O expansion drawers, for a total of
48 PCIe Gen 3 slots or 36 PCIe Gen 4 slots. A fully configured Power E1080 can support a
maximum of 16 Gen 3 I/O expansion drawers, providing a total of 192 PCIe Gen 3 slots or a
maximum of 12 Gen 4 I.O expansion drawers for a total of 144 Gen 4 I/O slots.
An optional NVMe expansion capability is provided by the NED24 NVMe expansion drawer.
The NED24 attaches to the system units by using the same fanout adapters that are used by
the I/O expansion drawers and provides up to 24 NVMe U2 drive slots. Each NVMe drive can
be individually assigned to its own partition, providing great flexibility for boot images and
more data for each partition.
More serial-attached SCSI (SAS) storage can be implemented by using the EXP24SX SAS
storage enclosure2, which provides twenty-four 2.5-inch small form factor (SFF) SAS bays. It
supports up to 24 hot-swap hard disk drives (HDDs) or solid-state drives (SSDs) in only 2U
rack units of space in a 19-inch rack. The EXP24SX is connected to the Power E1080 server
by using SAS adapters that are plugged into PCIe slots either in the system nodes or in an I/O
expansion drawer.
For more information about enclosures and drawers, see 1.6, “I/O drawers” on page 25.
More storage can also be allocated externally by using Storage Network Connections through
the appropriate adapters that are installed in the system PCIe slots. For more information
about IBM storage products, see this web page.
1
The PCIe Gen 3 I/O drawer was withdrawn from marketing as of January 2024.
2 The EXP24SX was withdrawn from marketing, but is still supported.
The massive computational power, exceptional system capacity, and the unprecedented
scalability of the Power E1080 server hardware are provided by unique enterprise class
firmware and system software capabilities and features. The following important
characteristics and features are offered by the IBM Power enterprise platform:
Support for IBM AIX, IBM i, and Linux operating system environments, including support
for Red Hat OpenShift Cloud Platform.
An innovative dense math engine (DME) that is integrated in each Power10 processor
core to accelerate AI-inferencing workloads.
Optimized encryption units that are implemented in each Power10 processor core.
Dedicated data compression engines that are provided by the Power10 processor
technology.
Hardware- and firmware-assisted and enforced security provide trusted boot and
pervasive memory encryption support.
Up to 1,000 virtual machines (VMs) or logical partitions (LPARs) per system.
Dynamic LPAR (DLPAR) support to modify available processor and memory resources
according to workload, without interruption of the business.
Capacity on Demand (CoD) processor and memory options to help respond more rapidly
and seamlessly to changing business requirements and growth.
IBM Power Private Cloud Solution with Dynamic Capacity featuring Power Enterprise
Pools 2.0 that supports unsurpassed enterprise flexibility for real-time workload balancing,
system maintenance and operational expenditure cost management.
Table 1-1 compares important technical characteristics of the Power E1080 server with the
Power E980 server, based on IBM Power9® processor-based technology.
Table 1-1 Comparing the Power E980 and the Power E1080 server
Features Power E980 server Power E1080 server
System configuration options 1-, 2-, 3-, and 4-node systems 1-, 2-, 3-, and 4-node systems
PCIe slots per node Eight PCIe Gen 4 slots Eight PCIe Gen 5 slots
Internal storage bays per node Four NVMe PCIe Gen 3 baysb Four NVMe PCIe Gen 4 bays
Figure 1-1 Power E1080 4-node server mounted in an S42 rack with a #ECRT door
The maximum number of supported PCIe Gen 3 I/O expansion drawers is four per system
node, which can be mixed with PCI Gen 4 expansion drawers (a maximum of three Gen 4
drawers are supported). Each I/O expansion drawer can be populated with two Fanout
Modules. Each Fanout Module is connected to a system node through one PCIe x16 to CXP
Converter Card.
In August 2024, IBM announced new memory features for the Power10 server line that use
DDR5 technology. New memory features are available for the DDR5-based memory. The
Power E1080 supports DDR5 and DDR4 DDIMMs in the same system, but all memory in
each node must be the same type. The DDR5 DDIMMs provide increased memory
bandwidth, which can result in improved performance of the system.
Each system node is 5U rack units high and holds four air-cooled Power10 SCMs that are
optimized for performance, scalability, and AI workloads. An SCM is constructed of one
Power10 processor chip and more logic, pins, and connectors that enable plugging the SCM
into the related socket on the system node system board.
The Power E1080 Power10 SCMs are available in 10-core, 12-core, or 15-core capacity. Each
core can run in an eight-way simultaneous multithreading (SMT) mode, which delivers eight
independent hardware threads of parallel execution power.
The 10-core SCMs are ordered in a set of four per system node through processor feature
#EDP2. In this way, feature #EDP2 provide 40 cores of processing power to one system node
and 160 cores of total system capacity in a 4-node Power E1080 server. The maximum
frequency of the 10-core SCM is specified with 3.9 GHz, which makes this SCM suitable as a
building block for entry class Power E1080 servers.
The 12-core SCMs are ordered in a set of four per system node through processor feature
#EDP3. In this way, feature #EDP3 provides 48 cores capacity per system node and a
maximum of 192 cores per fully configured 4-node Power E1080 server. This SCM type offers
the highest processor frequency at a maximum of 4.15 GHz, which makes it a perfect choice
if highest thread performance is one of the most important sizing goals.
The 15-core SCMs are ordered in a set of four per system node through processor feature
#EDP4. In this way, feature #EDP4 provides 60 cores per system node and an impressive
240 cores total system capacity for a 4-node Power E1080. The 15-core SCMs run with a
maximum of 4.0 GHz and meet the needs of environments with demanding thread
performance and high compute capacity density requirements.
Note: All Power10 SCMs within a system node must be of the same type: 10-core,
12-core, or 15-core. Also, all system nodes within a specific Power E1080 server must be
configured with identical processor features.
Three PowerAXON3 18-bit wide buses per Power10 processor chip are used to span a fully
connected fabric within a CEC drawer. Each SCM within a system node is directly connected to
every other SCM of the same drawer at 32 GBps speed. This on-system board interconnect
provides 128 GBps chip-to-chip data bandwidth, which marks an increase of 33% relative to the
previous Power9 processor-based on-system board interconnect implementation in Power E980
servers. The throughput can be calculated as 16 lanes with * 32 GBps = 64 GBps per direction *
2 directions for an aggregated rate of 128 GBps.
Each of the four Power10 processor chips in a Power E1080 CEC drawer is connected directly to
a Power10 processor chip at the same position in every other CEC drawer in a multi-node
system This connection is made by using an SMP PowerAXON 18-bit wide bus per connection
running at 32 Gbps speed.
The Power10 processor technology introduces a new memory connection technology that uses
the open memory interface (OMI). Using the OMI provides an abstraction layer for the memory
that allows the system memory to be introduced without changing the system board
connections. This capability allowed the introduction of DDR5-based memory.
The 16 available high-speed OMI links are driven by eight on-chip memory controller units
(MCUs) that provide a total aggregated bandwidth of up to 409 GBps per SCM. This design
represents a memory bandwidth increase of 78% compared to the Power9 processor-based
technology capability.
Every Power10 OMI link is directly connected to one memory buffer-based DDIMM slot.
Therefore, the four sockets of one system node offer a total of 64 DDIMM slots with an
aggregated maximum memory bandwidth of 1636 GBps. The DDIMM densities that are
supported in Power E1080 servers are 32 GB, 128 GB, and 256 GB, all of which use either
DDR5 or DDR4 technology.
The Power E1080 memory options are available as 128 GB (#EMC1), 256 GB (#EMC2),
512 GB (#EMC3), and 1024 GB (#EMC4) memory features when using DDR4 memory. For the
DDR5-based memory, the options are 128 GB (#EMFM), 256 GB (#EMFN), 512 GB (#EMFP),
and 1024 GB (#EMFQ). Each memory feature provides four DDIMMs.
Each system node supports a maximum of 16 memory features that cover the 64 DDIMM slots.
For the Power E1080, you may mix different DIMM sizes within a socket with the new Feature
Code #ECMC. This intermix is supported only when using DDR5-based DDIMMs.
Using the 1024 GB DDIMM features yields a maximum of 16 TB per node. Using the maximum
of four nodes provides a maximum of 64 TB capacity. The minimum number of DDIMMs per
node is eight. Activation of a minimum of 50% of the installed memory capacity is required
unless you use Power Enterprise Pools 2, in which case the minimum activation is 256 GB. For
more information about Power Enterprise Pools, see 2.4, “Capacity on Demand” on page 72.
The Power10 processor I/O subsystem is driven by 32 GHz differential PCIe 5.0 (PCIe Gen 5)
buses that provide 32 lanes that are grouped in two sets of 16 lanes. The 32 PCIe lanes deliver
an aggregate bandwidth of 576 GBps per system node and are used to support 8 half-length,
low-profile (LP) (half-height) adapter slots for external connectivity and 4 NVMe mainstream
SSDs of form factor U.2. for internal storage.
Six of the eight external PCIe slots can be used for PCIe Gen 4 x16 or PCIe Gen 5 x8 adapters
and the remaining two offer PCIe Gen 5 x8 capability. All PCIe slots support earlier generations of
the PCIe standard, such as PCIe Gen 1 (PCIe 1.0), PCIe Gen 2 (PCIe 2.0), PCIe Gen 3 (PCIe
3.0), and PCIe Gen 4 (PCIe 4.0).
For extra connectivity, up to four 19-inch PCIe Gen 3 4U high I/O expansion units (#EMX0) or up to
three PCIe Gen 4 expansion drawers can be attached to one system node. Each expansion
drawer contains one or two PCIe Fanout Modules with six PCIe Gen 3 full-length, full-height slots
each.
A fully configured 4-node Power E1080 server offers a total of 32 internal PCIe slots and up to
192 PCIe slots through I/O expansion units.
Figure 1-3 shows the rear view of a system node with the locations of the external ports and
features.
Figure 1-4 Top view of a Power E1080 system node with the top cover assembly removed
One SCU is required for each Power E1080 server (any number of system nodes), and
depending on the number of system nodes, the SCU is powered according to the following
rules:
Two universal power interconnect (UPIC) cables are used to provide redundant power to
the SCU.
In a Power E1080 single system node configuration, both UPIC cables are provided from
the single system node to be connected to the SCU.
For a two, three, or four system nodes configuration, one UPIC cable is provided from the
first system node and the second UPIC cable is provided from second system node to be
connected to the SCU.
The set of two cables facilitate a 1+1 redundant electric power supply. If there is a failure of
one cable, the remaining UPIC cable is sufficient to feed the needed power to the SCU.
Four FSP ports per FSP card provide redundant connection from the SCU to each system
node. System nodes connect to the SCU by using the cable features #EFCH, #EFCE, #EFCF,
and #EFCG. Feature #EFCH connects the first system node to the SCU and it is included by
default in every system node configuration. It provides FSP, UBIC, and USB cables, but no
SMP cables. All the other cable features are added depending on the number of extra system
nodes that are configured and includes FSP and SMP cables.
Figure 1-5 shows the front and rear view of a SCU with the locations of the external ports and
features.
Figure 1-5 Front and rear view of the system control unit
For more information about the comprehensive Model 9080-HEX server specifications and
product documentation, see IBM Documentation.
Each system component must be mounted in a 19-inch industry standard rack. The SCU
requires 2U rack units and each system node requires 5U rack units. Thus, a single-node
system requires 7U, a two-node system requires 12U, a three-node system requires 17U, and
a four-node system requires 22U rack units. More rack space must be allotted; for example, to
PCIe I/O expansion drawers, an HMC, flat panel console kit, network switches, power
distribution units (PDUs), and cable egress space.
Table 1-2 lists the physical dimensions of the Power E1080 server control unit and a
Power E1080 system node. The component height is also given in Electronic Industries
Alliance (EIA) rack units. (One EIA unit corresponds to one rack unit (U) and is defined as
1 3/4 inch or 44.45 mm respectively).
Height 86 mm (3.39 in) / 2 EIA units 217.25 mm (8.55 in) / 5 EIA units
Lift tools
It is a best practice to have a lift tool available at each site where one or more Power E1080
servers are located to avoid any delays when servicing systems. An optional lift tool #EB2Z is
available for order with a Power E1080 server. One #EB2Z lift tool can be shared among
many servers and I/O drawers. The #EB2Z lift tool provides a hand crank to lift and position
up to 159 kg (350 lb). The #EB2Z lift tool is 1.12 meters x 0.62 meters (44 in x 24.5 in).
Attention: A single system node can weigh up to 86.2 kg (190 lb). Also available are a
lighter, lower-cost lift tool (#EB3Z) and wedge shelf toolkit for #EB3Z with Feature Code
#EB4Z.
Depending on the specific Power E1080 configuration, the power for the SCU is provided
through two UPIC cables that are connected to one or two system nodes, as described in 1.2,
“System nodes” on page 7.
Table 1-3 lists the electrical characteristics per Power E1080 system node. For planning
purposes, use the maximum values that are provided. However, the power draw and heat
load depend on the specific processor, memory, adapter, and expansion drawer configuration
and the workload characteristics.
Phase Single
Note: The Power E1080 must be installed in a rack with a rear door and side panels for
electromagnetic compatibility compliance.
Environmental assessment: The IBM Systems Energy Estimator tool can provide more
accurate information about the power consumption and thermal output of systems based
on a specific configuration
Operating Non-operating
Recommended Allowable
A comprehensive list of noise emission values for various different Power E1080 server
configurations is provided by the Power E1080 product documentation. For more information
about noise emissions, search for “Model 9080-HEX server specifications” at IBM
Documentation.
Note: Government regulations, such as those regulations that are issued by the
Occupational Safety and Health Administration (OSHA) or European Community
Directives, can govern noise level exposure in the workplace and might apply to you and
your server installation. The Power E1080 server is available with an optional acoustical
door feature that can help reduce the noise that is emitted from this system.
The sound pressure levels in your installation depend on various factors, including the
number of racks in the installation; the size, materials, and configuration of the room where
you designate the racks to be installed; the noise levels from other equipment; the room
ambient temperature, and employees' location in relation to the equipment.
Also, compliance with such government regulations depends on various other factors,
including the duration of employees’ exposure and whether employees wear hearing
protection. As a best practice, consult with qualified experts in this field to determine
whether you are in compliance with the applicable regulations.
An overview of various Feature Codes and the essential information is also presented that
can help users design their system configuration with suitable features that can fulfill the
application compute requirement. This information also helps with building a highly available,
scalable, reliable, and flexible system around the application.
Table 1-5 lists the Power E1080 server configuration with minimal features.
Processor EDP2 40-core (4x10) Typical 3.65 - 3.90 GHz (max) 1 of any
EDP3 Power10 Processor with 5U system node Feature Code
EDP4 drawer
48-core (4x12) Typical 3.60 - 4.15 GHZ (max)
Power10 Processor with 5U system node
drawer
60-core (4x15) Typical 3.55 - 4.00 GHz (max)
Power10 Processor with 5U system node
drawer
Data protection 0040 Mirrored System Disk Level, Specify Code 1 if IBM i
0 for other
OS
UPIC cables EFCH System Node to System Control Unit Cable Set for 1
Drawer 1
Note: The minimum configuration that is generated by the IBM configurator includes more
administrative and indicator features.
EDP2 5C6C 40-core (4x10) Typical 3.65 GHz - 3.90 AIX, IBM i, and Linux
GHz (max) Power10 Processor with 5U
system node drawer
EDP3 5C6D 48-core (4x12) Typical 3.60 - 4.15 GHZ AIX, IBM i, and Linux
(max) Power10 Processor with 5U system
node drawer
EDP4 5C6E 60-core (4x15) Typical 3.55 GHz - 4.00 AIX, IBM i, and Linux
GHz (max) Power10 Processor with 5U
system node drawer
Each Feature Code provides four processor SCMs. The Feature Code identifies the number
of processors on each SCM, and a system configuration requires either one, two, three, or
four of the same processor feature, where the number of Feature Codes corresponds to the
number of system nodes.
The system nodes connect to other system nodes and to the SCU through cable connect
features. Table 1-7 lists the set of cable features that are required for one-, two-, three-, and
four-node configurations.
1-node 1 0 0 0
2-node 1 0 0 1
3-node 1 0 1 1
4-node 1 1 1 1
Every Feature Code that is listed in Table 1-6 provides the processor cores, not their
activation. The processor core must be activated to be assigned an LPAR. The activations are
offered through multiple permanent and temporary activation features. For more information
about these options, see 2.4, “Capacity on Demand” on page 72.
Table 1-8 lists the processor Feature Codes and the associated permanent activation
features. Any of these activation Feature Codes can permanently activate one core.
A minimum of 16 processor cores must always be activated with the static activation features,
regardless of the Power E1080 configuration. Also, if the server is associated to a PEP 2.0, a
minimum of one base activation is required.
For more information about other temporary activation offerings that are available for the
Power E1080 server, see 2.4, “Capacity on Demand” on page 72.
Regular and PEP 2.0 associated activations for Power E1080 are listed in Table 1-9. The
Order type table column includes the following designations:
Initial Denotes the orderability of a feature for only the new purchase of the
system.
MES Denotes the orderability of a feature for only the Miscellaneous
Equipment Specification (MES) upgrade purchases on the system.
Both Denotes the orderability of a feature for new and MES upgrade
purchases.
Supported Denotes that a feature is not orderable, but is supported. That is, the
feature can be migrated only from existing systems.
EPS0 1 core Base Proc Act (Pools 2.0) for #EDP2 any OS (from Static) MES
EPS1 1 core Base Proc Act (Pools 2.0) for #EDP3 any OS (from Static) MES
EPS2 1 core Base Proc Act (Pools 2.0) for #EDP4 any OS (from Static) MES
EPS5 1 core Base Proc Act (Pools 2.0) for #EDP2 Linux (from Static) MES
EPS6 1 core Base Proc Act (Pools 2.0) for #EDP3 Linux (from Static) MES
EPS7 1 core Base Proc Act (Pools 2.0) for #EDP4 Linux (from Static) MES
EPSK 1 core Base Proc Act (Pools 2.0) for #EDP2 any OS (from Prev) MES
EPSL 1 core Base Proc Act (Pools 2.0) for #EDP3 any OS (from Prev) MES
EPSM 1 core Base Proc Act (Pools 2.0) for #EDP4 any OS (from Prev) MES
EDP2 40-core (4x10) Typical 3.65 GHz - 3.90 GHz (max) Power10 Both
Processor with 5U system node drawer
EDP3 48-core (4x12) Typical 3.60 to 4.15 GHZ (max) Power10 Processor Both
with 5U system node drawer
EDP4 60-core (4x15) Typical 3.55 GHz - 4.00 GHz (max) Power10 Both
Processor with 5U system node drawer
EPDC 1 core Base Processor Activation (Pools 2.0) for EDP2 any OS Both
EPDD 1 core Base Processor Activation (Pools 2.0) for EDP3 any OS Both
EPDS 1 core Base Processor Activation (Pools 2.0) for EDP4 any OS Both
EPDU 1 core Base Processor Activation (Pools 2.0) for EDP2 Linux only Both
EPDW 1 core Base Processor Activation (Pools 2.0) for EDP3 Linux only Both
EPDX 1 core Base Processor Activation (Pools 2.0) for EDP4 Linux only Both
Because the DDIMMs are designed based on the OMI, it was possible to integrate new memory
technology without having to replace the memory connections in the system. In addition, several
enhancements were built into the DDR5-based DDIMMs that provided significant performance
improvements compared to the DDR4-based DDIMMs. For example, all the DDR5-based
memory runs at 4000 MHz, but when using the DDR4-based memory, the 32 GB and 64 GB
DDIMMs run at 3200 MHz frequency, and the 128 GB and 256 GB DDIMMs run at 2933 MHz
frequency. In addition, IBM added more connections to the DDIMMs to increase bandwidth
between the system and the memory. For more information about DDR5 memory, see 2.3,
“Memory subsystem” on page 68.
Each system node provides 64 DDIMM slots that support a maximum of 16 TB memory. A
4-system node E1080 can support a maximum of 64 TB memory. DDIMMs are ordered by
using memory Feature Codes, which include a bundle of four DDIMMs with the same
capacity.
Table 1-10 lists the available memory DDIMM Feature Codes for the Power E1080.
EMC1 128 GB (4x32 GB) DDIMMs, 3200 MHz, 16 Gbit DDR4 Memory AIX, IBM i, and
Linux
EMC2 256 GB (4x64 GB) DDIMMs, 3200 MHz, 16 Gbit DDR4 Memory AIX, IBM i, and
Linux
EMC3 512 GB (4x128 GB) DDIMMs, 2933 MHz, 16 Gbit DDR4 AIX, IBM i, and
Linux
EMC4 1 TB (4x256 GB) DDIMMs, 2933 MHz, 16 Gbit DDR4 AIX, IBM i, and
Linux
EMFM 128 GB (4x32 GB) DDIMMs, 4000 MHz, 16 Gbit DDR5 Memory AIX, IBM i, and
Linux
EMFN 256 GB (4x64 GB) DDIMMs, 4000 MHz, 16 Gbit DDR5 Memory AIX, IBM i, and
Linux
EMFP 512 GB (4x128 GB) DDIMMs, 4000 MHz, 16 Gbit DDR5 AIX, IBM i, and
Linux
EMFQ 1 TB (4x256 GB) DDIMMs, 4000 MHz, 16 Gbit DDR5 AIX, IBM i, and
Linux
A server administrator or user cannot control which physical memory DDIMM features are
activated when memory activations are used.
The amount of memory to activate is determined by the Feature Code that is ordered. For
example, ordering two instances of Feature Code EDAB (100 GB DDR4 Mobile Memory
Activation for HEX) activates a total of 200 GB of installed physical memory. The
IBM PowerVM® hypervisor recognizes the total quantity of each type of memory that is
activated on the server. Then, it manages the activation and allocation of the physical DDIMM
memory to the LPARs.
Similar to processor core activation features, different types of permanent memory activation
features are offered on the Power E1080 server. For more information about the available
types of activations, see 1.5.2, “Processor features” on page 17.
Orders for memory activation features must consider the following rules:
The system must have a minimum of 50% activated physical memory. It can be activated
by using static or static and mobile memory activation features.
The system must have a minimum of 25% of physical memory that is activated by using
static memory activation features.
When a Power E1080 is part of a PEP 2.0 environment, the server must have a minimum
of 256 GB of base memory activations.
Table 1-11 lists the available memory activation Feature Codes for Power E1080. The Order
type column indicates whether the Feature Code is available for an initial order only, or also
with a MES upgrade on an existing server only, or both.
EDAQ 100 GB Base Memory activation (Pools 2.0) from Static Both
EDAR 512 GB Base Memory activation (Pools 2.0) from Static Both
EDAS 500 GB Base Memory activation (Pools 2.0) from Static Both
EDAU 100 GB Base Memory activation (Pools 2.0) MES only Both
EDAV 100 GB Base Memory Activation (Pools 2.0) from Mobile Both
EDAW 500 GB Base Memory Activation (Pools 2.0) from Mobile Both
Table 1-12 PCIe adapters that are supported on a Power E1080 system node
Feature Code CCIN Description OS support
EN1A 578F PCIe Gen 3 32 Gb 2-port Fibre Channel AIX, IBM i, and Linux
Adapter
EN1B 578F PCIe Gen 3 LP 32 Gb 2-port Fibre Channel AIX, IBM i, and Linux
Adapter
EN1C 578E PCIe Gen 3 16 Gb 4-port Fibre Channel AIX, IBM i, and Linux
Adapter
EN1D 578E PCIe Gen 3 LP 16 Gb 4-port Fibre Channel AIX, IBM i, and Linux
Adapter
EN1F 579A PCIe Gen 3 LP 16 Gb 4-port Fibre Channel AIX, IBM i, and Linux
Adapter
EN1H 579B PCIe Gen 3 LP 2-Port 16 Gb Fibre Channel AIX, IBM i, and Linux
Adapter
EN1K 579C PCIe Gen 4 LP 32 Gb 2-port Optical Fibre AIX, IBM i, and Linux
Channel Adapter
EN2A 579D PCIe Gen 3 16 Gb 2-port Fibre Channel AIX, IBM i, and Linux
Adapter
EN2B 579D PCIe Gen 3 LP 16 Gb 2-port Fibre Channel AIX, IBM i, and Linux
Adapter
5260 576F PCIe2 LP 4-port 1 GbE Adapter AIX, IBM i, and Linux
5899 576F PCIe2 4-port 1 GbE Adapter AIX, IBM i, and Linux
EC2T 58FB PCIe Gen 3 LP 2-Port 25/10 Gb NIC&ROCE AIX, IBM i, and Linux
SR/Cu Adaptera
EC2U 58FB PCIe Gen 3 2-Port 25/10 Gb NIC&ROCE AIX, IBM i, and Linux
SR/Cu Adaptera
EC67 2CF3 PCIe Gen 4 LP 2-port 100 Gb ROCE EN LP AIX, IBM i, and Linux
adapter
EN0S 2CC3 PCIe2 4-Port (10 Gb+1 GbE) SR+RJ45 AIX, IBM i, and Linux
Adapter
EN0T 2CC3 PCIe2 LP 4-Port (10 Gb+1 GbE) SR+RJ45 AIX, IBM i, and Linux
Adapter
EN0X 2CC4 PCIe2 LP 2-port 10/1 GbE BaseT RJ45 AIX, IBM i, and Linux
Adapter
a. Requires Service Focal Point (SFP) to provide 10 Gb, 2 Gb, or 1 Gb BaseT connectivity.
If more NVMe capacity is required, NVMe drives can be placed in the NED24 disk enclosure.
For more information, see 2.7, “External I/O subsystems” on page 94.
The Power E1080 system node does not offer an integrated USB port. The USB 3.0 adapter
Feature Code EC6J is required to provide connectivity to an optional external USB DVD drive
and requires one system node or I/O expansion drawer PCIe slot. The adapter connects to
the USB port in the rear of the SCU with the cable that is associated to Feature Code EC6N.
Because this cable is 1.5 m long, if there is a Power E1080 with more than one system node,
the USB 3.0 adapter can be used in the first or the second system node only.
The USB 3.0 adapter Feature Code EC6J supports the assignment to an LPAR and can be
migrated from an operating LPAR to another, including the connected DVD drive. This design
allows it to assign the DVD drive feature to any LPAR according to the need.
Dynamic allocation of system resources such as processor, memory, and I/O is also referred
to as dynamic logical partition (DLPAR).
For more information about the USB subsystem, see 2.5.3, “USB subsystem” on page 87.
Four power cords from the PDUs drive these power supplies, which connect to four C13/C14
type receptacles on the power cord conduit in the rear of the system. The power cord conduit
sources power from the rear and connects to the PSUs in the front of the system.
The system design provides N+2 redundancy for system bulk power, which allows the system
to continue operation with any two of the PSUs functioning. The failed units must remain in
the system until new PSUs are available for replacement.
The PSUs are hot-swappable, which allows replacement of a failed unit without system
interruption. The PSUs are placed in front of the system, which makes any necessary service
simpler.
Figure 1-7 on page 25 shows the PSUs and their physical locations, which are marked as E1,
E2, E3, and E4 in the system.
If more PCIe slots beyond the system node slots are required, the Power E1080 server
supports adding I/O expansion drawers. Two different I/O expansion drawers are supported: the
PCIe Gen 4 and the PCIe Gen3 expansion drawers.4 To connect an I/O expansion drawer, a
PCIe interconnect feature (fanout module) is required. The fanout module requires a PCIe slot
and connects to a 6-slot expansion module in the I/O drawer. Up to two expansion modules are
supported per I/O expansion drawer. For more information about the I/O drawers, see 1.6.2,
“I/O expansion drawers” on page 26.
To increase the amount of internal disk in the system, the Power E1080 supports two different
disk drawers: a 24-bay SAS connected enclosure (the EXP24SX SAS storage enclosure)5 and
a 24-bay NVMe enclosure (the NED24 NVMe expansion drawer). The SAS enclosure uses
SAS adapters in the system to connect to the enclosure. The NVMe enclosure uses the same
fanout modules that are used by the I/O expansion drawers. These fanout adapters are
described in 1.6.1, “System node PCIe interconnect features” on page 25. For more information
about the drive enclosures, see 2.7, “External I/O subsystems” on page 94.
4
The PCIe Gen 3 Drawer was withdrawn from marketing in January 2024.
5 The EXP24SX SAS Storage Enclosure was withdrawn from marketing in October 2023.
Table 1-13 PCIe slots availability for different system nodes configurations
Number of Number of I/O LP slotsa Full-height slots Full-height slots
system expansions with Gen 3 drawer with Gen 4 drawer
nodes
1 4 8 48 36
2 8 16 96 72
3 12 24 144 108
4 16 32 192 144
a. For the Gen 3 drawer case, all LP slots are occupied with #EJ24. For the Gen 4 case, there are two LP
slots that are available in the system node.
Important: When using both the I/O expansion drawers and the NED24 disk expansion
drawers, the combined maximum of NED24 NVMe Expansion Drawer (#ESR0), PCIe
Gen4 I/O Expansion Drawer (#ENZ0), and PCIe Gen3 I/O Expansion Drawer (#EMX0) is
half of the maximum of controller cards #EJ24 that are allowed per server.
Each fanout module connects to the system by using a pair of CXP cable features. Select a
longer-length Feature Code for inter-rack connection between the system node and the
expansion drawer. The same cable features are used to connect either the Gen 4 expansion
drawer or the NED24 NVMe expansion drawer, but different cable features are used to
connect the Gen 3 expansion drawer (they are not interchangeable).
Both the CXP optical cable pair and the optical cable adapter features are concurrently
maintainable. Therefore, careful balancing of I/O, assigning adapters through redundant
EMX0 expansion drawers, and different system nodes can help ensure high availability for I/O
resources that are assigned to partitions.
Restriction: Only two different fanout modules are supported for use in I/O expansion
drawers for Power10 servers. Feature EMXF supports the Gen 3 expansion drawer, and
feature ENZF supports the Gen 4 expansion drawer. Previous versions are not supported.
For more information abut internal buses and the architecture of internal and external I/O
subsystems, see 2.5, “Internal I/O subsystem” on page 79.
A PCIe CXP converter adapter and Active Optical Cables (AOCs) connect the system node to
each PCIe fanout module in the I/O expansion drawer. Each expansion drawer has two power
supplies.
A blind-swap cassette (BSC) houses the full-high adapters that are installed in these slots.
The drawer includes a full set of BSCs, even if the BSCs are empty. Drawers can be added to
a server dynamically. Concurrent repair and adding or removing expansion drawers and PCIe
adapters is done through HMC-guided menus or by operating system support utilities.
Careful balancing of I/O, assigning adapters through redundant EMX0 expansion drawers,
and connectivity to different system nodes can help ensure high availability for I/O resources
that are assigned to LPARs.
Figure 1-9 shows the rear view of the PCIe Gen 3 I/O Expansion Drawer with the location
codes for the PCIe adapter slots in the PCIe Gen 3 6-slot Fanout Module.
Each of the 24 NVMe bays in the NED24 drawer is separately addressable and can be
assigned to a specific LPAR or Virtual I/O Server (VIOS) to provide native boot support for up
to 24 partitions. At the time of writing, each drawer can support up to 153 TB.
Figure 1-10 is a view of the front of the NED24 NVMe Expansion Drawer.
Up to 24 U.2 NVMe devices can be installed in the NED24 drawer by using 15 mm Gen3
carriers. The 15-mm carriers can accommodate either 7 mm or 15 mm NVMe devices.
The NED24 drawer is supported in the Power E1080 by using the same interconnect card
that is used for the PCIe Gen 4 and PCIe Gen 3 expansion drawers. A maximum of three
NED24 NVMe expansion drawers is supported per system node in the E1080. When mixing
the different expansion drawers, the maximum number of drawers that are supported is based
on the number of EJ24 fanout cards that are supported.
For more information about the NED24 Drawer, see 2.8.1, “NED24 NVMe Expansion Drawer”
on page 106.
The EXP24SX drawer is a storage expansion enclosure with 24 2.5-inch SFF SAS bays. It
supports up to 24 hot-plug HDDs or SSDs in only 2 EIA of space in a 19-inch rack. The
EXP24SX SFF bays use SFF Gen2 (SFF-2) carriers or trays.
With AIX/Linux/VIOS, the EXP24SX can be ordered with four sets of 6 bays (mode 4), two
sets of 12 bays (mode 2) or one set of 24 bays (mode 1). With IBM i one set of 24 bays
(mode 1) is supported. It is possible to change the mode setting in the field by using software
commands along with a documented procedure.
The attachment between the EXP24SX drawer and the PCIe Gen 3 SAS adapter is through
SAS YO12 or X12 cables. The PCIe Gen 3 SAS adapters support 6 Gb throughput. The
EXP24SX drawer can support up to 12 Gb throughput if future SAS adapters support that
capability.
The EXP24SX drawer includes redundant AC power supplies and two power cords.
For more information about SAS cabling and cabling configurations, see IBM Documentation.
Note: For the EXP24SX drawer, a maximum of twenty-four 2.5-inch SSDs or 2.5-inch
HDDs are supported in the #ESLS 24 SAS bays. HDDs and SSDs cannot be mixed in the
same mode-1 drawer. HDDs and SSDs can be mixed in a mode-2 or mode-4 drawer, but
they cannot be mixed within a logical split of the drawer. For example, in a mode-2 drawer
with two sets of 12 bays, one set can hold SSDs and one set can hold HDDs, but you
cannot mix SSDs and HDDs in the same set of 12 bays.
IBM DS8900F can provide 100% data encryption at-rest, in-flight and in the cloud. This
flexible storage supports IBM Power, IBM Z, and IBM LinuxONE. For more information, see
this web page.
IBM SAN Volume Controller provides a complete set of data resilience capabilities with high
availability, business continuance, and data security features. Storage supports automated
tiering with AI-based IBM Easy Tier® that can help improve performance at a lower cost. For
more information, see this web page.
Important: As a best practice, order the Power E1080 server with an IBM 42U enterprise
rack 7965-S42 feature. This rack provides a complete and higher-quality environment for
IBM Manufacturing system assembly and testing, and provides a complete package.
If a system is installed in a rack or cabinet that is not from IBM, help ensure that the rack
meets the requirements that are described in 1.7.7, “Original equipment manufacturer racks”
on page 38.
Attention: All PDUs that are installed in a rack that contains a Power E1080 server must
be installed horizontally to allow for cable routing in the sides of the rack.
Compared to the 7965-94Y Slim Rack, the Enterprise Slim Rack provides extra strength and
shipping and installation flexibility.
The 7965-S42 rack includes space for up to four PDUs in side pockets. Extra PDUs beyond
four are mounted horizontally and each uses 1U of rack space.
The Enterprise Slim Rack front door, which can be Basic Black/Flat (#ECRM) or High-End
appearance (#ECRT), has perforated steel, which provides ventilation, physical security, and
visibility of indicator lights in the installed equipment within.
Standard is a lock that is identical to the locks in the rear doors. The door (#ECRG) can be
hinged on the left or right side.
Two possible PDU ratings are supported: 60A/63A (orderable in most countries) and
30A/32A. Consider the following points:
The 60A/63A PDU supports four system node power supplies and one I/O expansion
drawer or eight I/O expansion drawers.
The 30A/32A PDU supports two system node power supplies and one I/O expansion
drawer or four I/O expansion drawers.
High-function PDUs provide more electrical power per PDU and offer better “PDU footprint”
efficiency. In addition, they are intelligent PDUs that provide insight into power usage by
receptacle and remote power on and off capability for support by individual receptacle. The
new PDUs are orderable as #ECJJ, #ECJL, #ECJN, and #ECJQ.
Two or more PDUs can be installed horizontally in the rear of the rack. Mounting PDUs
horizontally uses 1U per PDU and reduces the space that is available for other racked
components. When mounting PDUs horizontally, the preferred approach is to use fillers in the
EIA units that are occupied by these PDUs to facilitate proper air-flow and ventilation in the
rack.
Each PDU requires one PDU-to-wall power cord. Various power cord features are available
for various countries and applications by varying the PDU-to-wall power cord, which must be
ordered separately.
Each power cord provides the unique design characteristics for the specific power
requirements. To match new power requirements and save previous investments, these
power cords can be requested with an initial order of the rack or with a later upgrade of the
rack features.
Table 1-15 Wall power cord options for the PDU and iPDU features
Feat Wall plug Rated voltage Phase Rated Geography
code (V AC) amperage
6654 NEMA L6-30 200 - 208, 240 1 24 amps US, Canada, LA, and
Japan
6655 RS 3750DP 200 - 208, 240 1 24 amps US, Canada, LA, and
(watertight) Japan
6492 IEC 309, 2P+G, 200 - 208, 240 1 48 amps US, Canada, LA, and
60 A Japan
Important: Help ensure that the suitable power cord feature is configured to support the
power that is being supplied. Based on the power cord that is used, the PDU can supply
between 4.8 - 19.2 kVA. The power of all the drawers that are plugged into the PDU must
not exceed the power cord limitation.
To better enable electrical redundancy, each CEC has four power supplies that must be
connected to separate PDUs, which are not included in the base order.
For maximum availability, a preferred approach is to connect power cords from the same
system to two separate PDUs in the rack, and to connect each PDU to independent power
sources.
For more information about power requirements of and the power cord for the 7965-94Y rack,
see IBM Documentation.
The IBM System Storage 7226 Multi-Media Enclosure supports LTO Ultrium and DAT160
Tape technology, DVD-RAM, and RDX removable storage requirements on the following IBM
systems:
IBM POWER6 processor-based systems
IBM POWER7 processor-based systems
IBM Power8® processor-based systems
IBM Power9 processor-based systems
IBM Power10 processor-based systems
The IBM System Storage 7226 Multi-Media Enclosure offers an expansive list of drive feature
options, as listed in Table 1-17.
5763 DVD Front USB Port Sled with DVD-RAM USB Drive Available
Removable RDX drives are in a rugged cartridge that inserts in to an RDX removable (USB)
disk docking station (#1103 or #EU03). RDX drives are compatible with docking stations,
which are installed internally in Power8, Power9, and Power10 processor-based servers,
where applicable.
Figure 1-13 shows the IBM System Storage 7226 Multi-Media Enclosure.
The IBM System Storage 7226 Multi-Media Enclosure offers a customer-replaceable unit
(CRU) maintenance service to help make the installation or replacement of new drives
efficient. Other 7226 components are also designed for CRU maintenance.
The IBM System Storage 7226 Multi-Media Enclosure is compatible with most Power8,
Power9, and Power10 processor-based systems that offer current level AIX, IBM i, and Linux
operating systems.
For a complete list of host software versions and release levels that support the IBM System
Storage 7226 Multi-Media Enclosure, see IBM System Storage Interoperation Center (SSIC).
Note: Any of the existing 7216-1U2, 7216-1U3, and 7214-1U2 multimedia drawers are
also supported.
The Model TF5 is a follow-on product to the Model TF4 and offers the following features:
A slim, sleek, and lightweight monitor design that occupies only 1U (1.75 in.) in a 19-inch
standard rack
A 18.5 inch (409.8 mm x 230.4 mm) flat panel TFT monitor with truly accurate images and
virtually no distortion
The IBM Documentation provides the general rack specifications, including the following
information:
The rack or cabinet must meet the EIA Standard EIA-310-D for 19-inch racks that was
published August 24, 1992. The EIA-310-D standard specifies internal dimensions, for
example, the width of the rack opening (width of the chassis), the width of the module
mounting flanges, and the mounting hole spacing.
The front rack opening must be a minimum of 450 mm (17.72 in.) wide, and the
rail-mounting holes must be 465 mm plus or minus 1.6 mm (18.3 in. plus or minus 0.06 in.)
apart on center (horizontal width between vertical columns of holes on the two
front-mounting flanges and on the two rear-mounting flanges).
Figure 1-14 is a top view showing the rack specification dimensions.
The vertical distance between mounting holes must consist of sets of three holes that are
spaced (from bottom to top) 15.9 mm (0.625 in.), 15.9 mm (0.625 in.), and 12.7 mm (0.5
in.) on center, which makes each three-hole set of vertical hole spacing 44.45 mm
(1.75 in.) apart on center.
The following rack hole sizes are supported for racks where IBM hardware is mounted:
– 7.1 mm (0.28 in.) plus or minus 0.1 mm (round)
– 9.5 mm (0.37 in.) plus or minus 0.1 mm (square)
The rack or cabinet must be capable of supporting an average load of 20 kg (44 lb.) of product
weight per EIA unit. For example, a four EIA drawer has a maximum drawer weight of 80 kg
(176 lb.).
Note: The recovery media for V10R1 is the same for 7063-CR2 and 7063-CR1.
Any customer with valid contract can download it from Entitled Systems Support (ESS) site or
it can be included within an initial Power E1080 order.
The following minimum requirements must be met to install the virtual HMC:
16 GB of memory
Four virtual processors
Two network interfaces (maximum 4 allowed)
One disk drive (500 GB available disk drive)
For an initial Power E1080 order with the IBM configurator (e-config), HMC virtual appliance
can be found by selecting add software → Other System Offerings (as product selections)
and then:
5765-VHP for IBM HMC Virtual Appliance for Power V10
5765-VHX for IBM HMC Virtual Appliance x86 V10
For more information about an overview of the Virtual HMC, see this web page.
For more information about how to install the virtual HMC appliance and all requirements, see
IBM Documentation.
The 7063-CR2 provides two network interfaces (eth0 and eth1) for configuring network
connectivity for BMC on the appliance.
Each interface maps to a different physical port on the system. Different management tools
name the interfaces differently. The HMC task Console Management → Console
Settings → Change BMC/IPMI Network Settings modifies only the Dedicated interface.
The main difference is that the shared and dedicated interface to the BMC can coexist. Each
has its own LAN number and physical port. Ideally, the customer configures one port, but both
can be configured. Connecting Power servers to the HMC rules remain the same as previous
versions.
New features and capabilities are implemented with firmware updates, and many firmware
updates require a minimum version of HMC code. Help ensure that your installed firmware
and HMC levels support the features that you plan to implement.
The IBM Power processor-based architecture has always ranked highly in terms of
end-to-end security, which is why it remains a platform of choice for mission-critical enterprise
workloads.
A key aspect of maintaining a secure Power environment is helping ensure that the HMC (or
virtual HMC) is current and fully supported (including hardware, software, and Power
firmware updates).
Outdated or unsupported HMCs represent a technology risk that can quickly and easily be
mitigated by upgrading to a current release.
Always complete performance sizing at the application workload environment level and
evaluate performance by using real-world performance measurements and production
workloads.
Midplane
To CEC #2
To CEC #2
Power-In
Power-In
To CEC #1
To CEC #1
Service Service
processor #0 processor #1
USB port USB port
128 GBps
128 GBps
128 GBps
128 GBps
128 GBps
128 GBps
128 GBps
128 GBps
128 GBps
128 GBps
128 GBps
128 GBps
128 GBps
128 GBps
128 GBps
x16 PCIe Gen4 x16 PCIe Gen4 x16 PCIe Gen4 x8 PCIe Gen5 x16 PCIe Gen4 x8 PCIe Gen5 x16 PCIe Gen4 x16 PCIe Gen4
or or or or or or
x8 PCIe Gen5 x8 PCIe Gen5 x8 PCIe Gen5 x8 PCIe Gen5 x8 PCIe Gen5 x8 PCIe Gen5
SMP connectors to
SMP connectors to SMP connectors to additional SMP connectors to
additional additional CEC drawer or additional
CEC drawer or CEC drawer or OpenCAPI to CEC drawer or
OpenCAPI to OpenCAPI to external I/O drawer 128 GBps OpenCAPI to
external I/O drawer external I/O drawer external I/O drawer
PHB
PHB
PHB
PHB
PHB
PHB
PHB
PHB
PHB
PHB
PHB
PHB
PHB
PHB
OP5
OP6 OP6 OP5 OP6 OP6
OP4 OP4 OP4 OP4
OP2 OP2 OP2 OP2
OP1 Processor chip OP1 Processor chip OP1 Processor chip OP1 Processor chip
OP0 socket #0 OP5 128 GBps OP7 socket #2 OP0 128 GBps OP3 socket #3 OP5 128 GBps OP7 socket #1 OP3 128 GBps
OP7 OP7 128 GBps
PHB PHB
Memory
Four memory controllers Memory
Four memory controllers
Memory Four memory controllers Four memory controllers
two
controller
channels#0 for each controller two
controller
channels#0 for each
controller
controller
#1 PHB PHB two channels for each controller two channels for each controller
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
DDIMMs
Control signals from
system control unit
Extender card for flash drives
The IBM Power10 processor was introduced to the public on August 17, 2020 at the 32nd
HOT CHIPS1 semiconductor conference. At that meeting, the new capabilities and features of
the latest POWER processor microarchitecture and the Power Instruction Set Architecture
(ISA) v3.1B were revealed and categorized according to the following Power10 processor
design priority focus areas:
Data plane bandwidth focus area
Terabyte per-second signaling bandwidth on processor functional interfaces, petabyte
system memory capacities, 16-socket symmetric multiprocessing (SMP) scalability, and
memory clustering and memory inception capability.
Powerful enterprise core focus area
New core micro-architecture, flexibility, larger caches, and reduced latencies.
End-to-end security focus area
Hardware enabled security features that are co-optimized with PowerVM hypervisor
support.
1 Source: [Link]
The remainder of this section provides more specific information about the Power10
processor technology as it is used in the Power E1080 scale-up enterprise class server.
The IBM’s Power10 Processor session material as presented at the 32nd HOT CHIPS
conference is available through the HC32 conference proceedings archive at this web page.
Each core has private access to 2 MB L2 cache and local access to 8 MB of L3 cache
capacity. The local L3 cache region of a specific core is also accessible from all other cores
on the processor chip. The cores of one Power10 processor share up to 120 MB of latency
optimized non-uniform cache access (NUCA) L3 cache.
The processor supports the following three distinct functional interfaces, which all are capable
running with a signaling rate of up to 32 giga-transfers per second (GTps):
Open memory interface (OMI)
The Power10 processor has eight memory controller unit (MCU) channels that support
one OMI port with two OMI links each2. One OMI link aggregates 8 lanes running at
32 GTps speed and connects to one memory buffer based differential dual inline memory
module (DDIMM) slot to access main memory. Physically, the OMI interface is
implemented in two separate die areas of 8 OMI links each. The maximum theoretical
full-duplex bandwidth aggregated over all 128 OMI lanes is 1 TBps.
SMP fabric interconnect (PowerAXON)
A total of 144 lanes are available in the Power10 processor to facilitate the connectivity to
other processors in an SMP architecture configuration. Each SMP connection requires 18
lanes, eight data lanes plus one spare lane per direction (2 x(8+1)). In this way the
processor can support a maximum of eight SMP connections with at total of 128 data
lanes per processor. This configuration yields a maximum theoretical full-duplex
bandwidth aggregated over all SMP connections of 1 TBps.
The generic nature of the interface implementation also allows the use of 128 data lanes
to potentially connect accelerator or memory devices through the OpenCAPI protocols.
Also, it can support memory cluster and memory interception architectures.
Because of the versatile characteristic of the technology, it is also referred to as
PowerAXON interface (Power A-bus/X-bus/OpenCAPI/Networking3). The OpenCAPI and
the memory clustering and memory interception use cases can be pursued in the future
and are not used by available technology products at the time of writing.
2
The OMI links are also referred to as OMI subchannels.
3
A-busses and X-busses provide SMP fabric ports that are used between central electronic complex (CEC) drawers
or within CEC drawers respectively.
Figure 2-2 shows the Power10 processor die with several functional units labeled. Note, 16
SMT8 processor cores are shown, but only 10-, 12-, or 15-core processor options are
available for Power E1080 server configurations.
Figure 2-2 The Power10 processor chip (Die photo courtesy of Samsung Foundry)
Table 2-1 Summary of the Power10 processor chip and processor core technology
Technology Power10 processor
Processor compatibility modes Support for Power ISAa of Power8 and Power9
a. Power ISA
The Power10 processor is packaged as a single-chip module (SCM) for exclusive use in the
Power E1080 servers. The SCM contains the Power10 processor plus more logic that is
needed to facilitate power supply and external connectivity to the chip. It also holds the
connectors to plug SMP cables directly onto the socket to build 2-, 3-, and 4-node
Power E1080 servers.
72 A-bus/X-bus/OpenCAPI lanes
to SMP connectors on top of module
OP1
OP2
OP4
OP6
E0 E1
32 PCIe5 lanes to
bottom of module
The second set of ports (OP0, OP3, OP5, and OP7) are used to implement the fully
connected SMP fabric between the four sockets within a system node. Eight OMI ports (OMI0
to OMI7) with two OMI links each provide access to the buffered main memory DDIMMs. The
32 PCIe Gen 5 lanes are grouped in two PCIe host bridges (PHBs) (E0, E1).
Figure 2-4 shows a physical diagram of the Power10 SCM. The eight SMP connectors
(OP1A, OP1B, OP2A, OP2B, OP4A, OP4B, OP6A, and OP6B) externalize 4 SMP busses,
which are used to connect system node drawers in 2-, 3-, and 4-node Power E1080
configurations. The OpenCAPI connectivity options are also indicated, although they are not
used by any commercially available product.
SMP connector
on module
OP4A
(8x per SCM)
OP6A
OP4B
AXON
AXON
OP6
OP4
OMI
AX
OP1
OP2
AX
OP1A
The SMT8 core includes two execution resource domains. Each domain provides the
functional units to service up to four hardware threads.
Figure 2-5 shows the functional units of an SMT8 core where all 8 threads are active. The two
execution resource domains are highlighted with colored backgrounds in two different shades
of blue.
Each of the two execution resource domains supports between one and four threads and
includes 4 vector scalar units (VSUs) of 128-bit width, two Matrix Math Accelerator (MMA)
accelerators, and one quad-precision floating-point (QPFP) and decimal floating-point (DFP)
unit.
One VSU and the directly associated logic are called an execution slice. Two neighboring
slices can also be used as a combined execution resource which is then named super-slice.
When operating in SMT8 mode, eight simultaneous multithreading (SMT) threads are
subdivided in pairs that collectively run on two adjacent slices as indicated through colored
backgrounds in different shades of green.
In SMT4 or lower thread modes, one to two threads each share a 4-slice resource domain.
Figure 2-5 also indicates other essential resources that are shared among the SMT threads,
such as instruction cache, instruction buffer and L1 data cache.
The SMT8 core supports automatic workload balancing to change the operational SMT
thread level. Depending on the workload characteristics, the number of threads that are
running on one chiplet can be reduced from four to two and even further to only one active
thread. An individual thread can benefit in terms of performance if fewer threads run against
the core’s executions resources.
The Power10 processor core includes the following key features and improvements that affect
performance:
Enhanced load and store bandwidth
Deeper and wider instruction windows
Enhanced data prefetch
Branch execution and prediction enhancements
Instruction fusion
Enhancements in the area of computation resources, working set size, and data access
latency are described next. The change in relation to the Power9 processor core
implementation is provided in parentheses.
Micro-architectural innovations that complement physical and logic design techniques and
specifically address energy efficiency include the following examples:
Improved clock-gating
Reduced flush rates with improved branch prediction accuracy
Fusion and gather operating merging
Reduced number of ports and reduced access to selected structures
Effective address (EA)-tagged L1 data and instruction cache yield ERAT access only on a
cache miss
If more than one hardware thread is active, the processor runs in SMT mode. In addition to
the ST mode, the Power10 processor supports the following different SMT modes:
SMT2: Two hardware threads are active.
SMT4: Four hardware threads are active.
SMT8: Eight hardware threads are active.
SMT enables a single physical processor core to simultaneously dispatch instructions from
more than one hardware thread context. Computational workloads can use the processor
core’s execution units with a higher degree of parallelism. This ability significantly enhances
the throughput and scalability of multi-threaded applications and optimizes the compute
density for ST workloads.
Table 2-2 lists a historic account of the SMT capabilities that are supported by each
implementation of the IBM Power Architecture® since POWER4.
IBM POWER4 32 ST 32
To efficiently accelerate MMA operations, the Power10 processor core implements a dense math
engine (DME) microarchitecture that effectively provides an accelerator for cognitive computing,
machine learning, and AI inferencing workloads.
The DME encapsulates compute efficient pipelines, a physical register file, and associated
data-flow that keeps resulting accumulator data local to the compute units. Each MMA pipeline
performs outer-product matrix operations, reading from and writing back a 512-bit accumulator
register.
Power10 implements the MMA accumulator architecture without adding an architected state. Each
architected 512-bit accumulator register is backed by four 128-bit Vector Scalar eXtension (VSX)
registers.
Code that uses the MMA instructions is included in OpenBLAS and Eigen libraries. This library can
be built by using the most recent versions of the GNU Compiler Collection (GCC) compiler. The
latest version of OpenBLAS is available at this web page.
OpenBLAS is used by Python-NumPy library, PyTorch, and other frameworks, which make it
simple to use the performance benefit of the Power10 MMA accelerator for AI workloads.
For more information about the implementation of the Power10 processor’s high throughput math
engine, see the white paper A matrix math facility for Power ISA processors.
For more information about fundamental MMA architecture principles with detailed instruction set
usage, register file management concepts, and various supporting facilities, see Matrix-Multiply
Assist Best Practices Guide, REDP-5612.
Depending on the specific settings of the PCR, the Power10 core runs in a compatibility mode
that pertains to Power9 (Power ISA version 3.0) or Power8 (Power ISA version 2.07)
processors. The support for processor compatibility modes also enables older operating
systems versions of AIX, IBM i, Linux, or Virtual I/O Server (VIOS) environments to run on
Power10 processor-based systems.
The Power10 processors of the Power E1080 server support the following compatibility
modes:
Power8: Enabled by firmware level FW810
Power9 Base: Enabled by firmware level FW910
Power9: Enabled by firmware level FW940
Power10: Enabled by firmware level FW1010
Table 2-3 lists the processor Feature Codes that are available for Power E1080 servers.
Power10 processors automatically optimize their core frequencies based on workload
requirements, thermal conditions, and power consumption characteristics. Each processor
Feature Code that is listed is associated with a processor core frequency range within which
the SCM cores typically operate. Also, the maximum frequency that is supported by the
relevant SCM type is indicated.
Intelligent L3 cache management enables the Power10 processor to optimize the access to L3
cache lines and minimize cache latencies. The L3 includes a replacement algorithm with data type
and reuse awareness. It also supports an array of prefetch requests from the core, including
instruction and data, and works cooperatively with the core, memory controller, and SMP
interconnection fabric to manage prefetch traffic, which optimizes system throughput and data
latency.
The aggregated maximum theoretical full-duplex bandwidth of the OMI interface culminates
at 2 x 512 GBps = 1 TBps per Power10 SCM.
The OMI physical interface enables low-latency, high-bandwidth, and technology-neutral host
memory semantics to the processor. It enables attaching established and emerging memory
elements. With the Power E1080 server, OMI initially supported one main tier, low-latency,
and enterprise-grade Double Data Rate 4 (DDR4) DDIMM per OMI link. In August 2024, IBM
announced support for a new DDIMM that uses Double Data Rate 5 memory. The DDR5
technology DDIMMs are available in the same capacities as the DDR4 features, but have a
higher memory throughput. The higher memory throughput is due to two enhancements in
the new DDR5-based DDIMM:
The DDR5 memory runs at 4000 MHz versus a maximum of 3200 MHz for some of the
DDR4-based DDIMMS (the larger DDIMMS ran at 2933 MHz). This new memory provides
25% - 36% bandwidth improvement).
The DDR5 DDIMM includes a second DRAM port from the OMI buffer in the DDIMM to the
DRAM. It provides up to 50% more sustained total bandwidth with mixed read/write traffic.
This DDIMM also reduces the latency when memory is loaded by up to 15%.
This configuration yields a total memory capacity of 16 DDIMMs per SCM and 64 DDIMMs
per Power E1080 system node. The memory bandwidth depends on the DDIMM density that
is configured for a specific Power E1080 server.
The maximum theoretical duplex memory bandwidth when using DDR4-based DDIMMs is
409 GBps per SCM if 32 GB or 64 GB DDIMMs that are running at 3200 MHz are used. The
maximum memory bandwidth is reduced to 375 GBps per SCM if 128 GB or 256 GB
DDIMMs that are running at 2933 MHz are used. When using the DDR5-based memory, the
maximum theoretical duplex memory bandwidth is 1024 GBps regardless of the DDIMM
capacity because all DDR5 DDIMMs run at 4000 MHz.
In summary, the Power10 SCM supports 128 OMI lanes with the following characteristics:
32 Gbps signaling rate
Eight lanes per OMI link
Two OMI links per OMI port (2 x 8 lanes)
Eight OMI ports per SCM (16 x 8 lanes)
The Power E1080 servers support the AES CTR mode for pervasive memory encryption.
Each Power10 processor holds a 128-bit encryption key that is used by the processor’s MCU
to encrypt the data of the DDIMMs that are attached to the OMI links.
The MCU crypto engine is transparently integrated into the data path, which helps ensure that
the data fetch and store bandwidth are not compromised by the AES CTR encryption mode.
Because the encryption has no noticeable performance effect and because of the obvious
security benefit, the pervasive memory encryption is enabled by default and cannot be turned
off through any administrative interface.
To facilitate LPM data compression and encryption, the hypervisor on the source system
presents the LPM buffer to the on-chip nest accelerator (NX) unit as part of the process in
Step 2. The reverse decryption and decompress operation is applied on the target server
as part of the process in Step 4.
The pervasive memory encryption logic of the MCU decrypts the memory data before it is
compressed and encrypted by the NX unit on the source server. It also encrypts the data
before it is written to memory but after it is decrypted and decompressed by the NX unit of
the target server.
Note: The pervasive memory encryption of the Power10 processor does not affect the
encryption status of a system dump content. All data that is coming from the DDIMMs is
decrypted by the MCU before it is passed onto the dump devices under the control of the
dump program code. This statement applies to the traditional system dump under the
operating system control and the firmware assist dump utility.
Each one of the AES/SHA engines, data compression, and Gzip units consist of a
coprocessor type and the NX unit features three coprocessor types. The NX unit also
includes more support hardware to support coprocessor invocation by user code, use of EAs,
high-bandwidth storage accesses, and interrupt notification of job completion.
The direct memory access (DMA) controller of the NX unit helps to start the coprocessors
and move data on behalf of coprocessors. SMP interconnect unit (SIU) provides the interface
between the Power10 SMP interconnect and the DMA controller.
The NX coprocessors can be started transparently through library or operating system kernel
calls to speed up operations that are related to data compression, LPM migration, IPsec,
JFS2 encrypted file systems, PKCS11 encryption, random number generation, and the most
recently announced logical volume encryption.
In effect, this on-chip NX unit on Power10 systems implements a high throughput engine that
can perform the equivalent work of multiple cores. The system performance can benefit by
offloading these expensive operations to on-chip accelerators, which in turn can greatly
reduce the CPU usage and improve the performance of applications.
The accelerators are shared among the logical partitions (LPARs) under the control of the
PowerVM hypervisor and accessed by way of hypervisor call. The operating system, along
with the PowerVM hypervisor, provides a send address space that is unique per process
requesting the coprocessor access. This configuration allows the user process to directly post
entries to the first in - first out (FIFO) queues that are associated with the NX accelerators.
Each NX coprocessor type has a unique receive address space corresponding to a unique
FIFO for each of the accelerators.
For more information about the use of the xgzip tool that uses the gzip accelerator engine,
see the following resources:
Using the Power9 NX (gzip) accelerator in AIX
Power9 GZIP Data Acceleration with IBM AIX
Note: The OpenCAPI interface and the memory clustering interconnect are Power10
technology options for future use.
Because of the versatile nature of signaling technology, the 32 Gbps interface is also referred
to as Power/A-bus/X-bus/OpenCAPI/Networking (PowerAXON) interface. The IBM
proprietary X-bus links connect two processors on a system board with a common reference
clock. The IBM proprietary A-bus links connect two processors in different drawers on
different reference clocks by using a cable.
OpenCAPI is an open interface architecture that allows any microprocessor to attach to the
following items:
Coherent user-level accelerators and I/O devices
Advanced memories are accessible through read/write or user-level DMA semantics
The PowerAXON interface is implemented on dedicated areas that are at each corner of the
Power10 processor die. The Power E1080 server uses this interface to implement
single-drawer chip-to-chip and drawer-to-drawer chip interconnects.
The Power E1080 single-drawer chip-to-chip SMP interconnect features the following
properties:
Three (2 x 9)-bit on system board buses per Power10 SCM
Eight data lanes, plus one spare lane in each direction per chip-to-chip connection
32 Gbps signaling rate providing 128 GBps per chip-to-chip SMP connection bandwidth,
an increase of 33% compared to the Power E980 single-drawer implementation
4-way SMP architecture implementations build out of four Power10 SCMs per drawer in a
1-hop topology
Figure 2-7 shows the SMP connections for a fully configured 4-node 16-socket Power E1080
system. The blue lines represent the chip-to-chip connections within one system node. The
green lines represent the drawer-to-drawer SMP connections.
From the drawing that is shown in Figure 2-7, you can deduce that each socket is directly
connected to any other socket within one system node and only one intermediary socket is
required to get from a chip to any other chip in another CEC drawer.
Based on the extensive experience that was gained over the past few years, the Power10
EnergyScale technology evolved to use the following effective and simplified set of
operational modes:
Power-saving mode
Static mode (nominal frequency)
Maximum performance mode (MPM)
The Power10 processor-based Power E1080 system features MPM enabled by default. This
mode dynamically adjusts processor frequency to maximize performance and enable a higher
processor frequency range. Each of the power saver modes delivers consistent system
performance without any variation if the nominal operating environment limits are met.
For Power10 processor-based systems that are under control of the PowerVM hypervisor, the
MPM is a system-wide configuration setting, but each processor module frequency is
optimized separately.
The following factors determine the maximum frequency that a processor module can run at:
Processor utilization: Lighter workloads run at higher frequencies.
Number of active cores: Fewer active cores run at higher frequencies.
Environmental conditions: At lower ambient temperatures, cores are enabled to run at
higher frequencies.
Figure 2-8 shows the comparative frequency ranges for the Power10 power-saving mode,
static or nominal mode, and the MPM. The frequency adjustments for different workload
characteristics, ambient conditions, and idle states are also indicated.
Figure 2-8 Power10 power management modes and related frequency ranges
Table 2-4 shows the power-saving mode and the static mode frequencies, and the frequency
ranges of the MPM for all three processor module types that are available for the
Power E1080 server.
Table 2-4 Characteristic frequencies and frequency ranges for Power E1080 server
Feature Cores per Power-saving Static mode MPM frequency range
code SCM mode frequency frequency [GHz]
[GHz] [GHz]
Figure 2-9 shows the ASM interface menu for Power and Performance Mode Setup on a
Power E1080 server.
Figure 2-9 Power E1080 ASMI menu for Power and Performance Mode setup
Figure 2-10 Power E1080 HMC menu for Power and Performance Mode setup
Table 2-5 compares key features and characteristics of the Power10, Power9, and Power8
processor implementations as used in enterprise class scale-up servers.
Table 2-5 Comparing the Power10 processor technology to prior processor generations
Characteristics Power10 Power9 Power8
Technology 7 nm 14 nm 22 nm
Maximum cores 15 12 12
Maximum static frequency / 3.75 - 4.15 GHZ 3.9 - 4.0 GHz 4.15 GHz
high-performance frequency
range
Power System E980 SMP cables cannot be used on Power E1080 because the Power10
processor-based server uses a different set of cables.
Figure 2-11 shows how each Power10 processor socket in a system node drawer has the bus
routed to the rear tail stock of the chassis. The Power10 processor chip in the first socket (see
processor chip socket #0 in Figure 2-11) uses ports T0 and T1 for the OP6 bus connection
(see upper left in Figure 2-11).
T0 T2 T4 T6 T8 T10 T12 T14 T16 T18 T20 T22 T24 T26 T28 T30
T1 T3 T5 T7 T9 T11 T13 T15 T17 T19 T21 T23 T25 T27 T29 T31
Simplified view of the rear of system node and connection between processor chip socket to SMP ports
Figure 2-11 Logical connection from processor chip sockets to SMP external ports
T0 T1 T2 T3 T4 T5 T6 T7 T8 T9 T10 T11 T12 T13 T14 T15 T16 T17 T18 T19 T20 T21 T22 T23 T24 T25 T26 T27 T28 T29 T30 T31
Node 0
T0 T1 T2 T3 T4 T5 T6 T7 T8 T9 T10 T11 T12 T13 T14 T15 T16 T17 T18 T19 T20 T21 T22 T23 T24 T25 T26 T27 T28 T29 T30 T31
The length of every cable is 775 mm (2.5 ft.) and its tape color is violet. As Figure 2-12 shows,
16 cables are needed for this configuration.
The Power E1080 supports concurrent repair of the external SMP cables; therefore, the entire
server does not need to be shut down to replace an external SMP cable.
The concurrent repair process enables the system to restore the OP-bus to the full function
without a system outage.
Every Power10 chip MCU provides the system memory interface between the on-chip SMP
interconnect fabric and the OpenCAPI memory interface (OMI) links. Logically eight,
essentially independent MCU channels are on every chip that interface to a total of 16
high-speed OMI links.
Each memory channel supports two OMI links (referred to as subchannels A and B) as shown
in Figure 2-13, which also shows the memory slot locations that are associated to the related
subchannels, channels, and MCUs.
All 16 of the OMI subchannels come off the Power10 SCM and are wired to 16 memory buffer
based DDIMM slots, with one DDIMM connected to each OMI subchannel.
IBM designed a new DDIMM to support DDR5 memory DIMMS. The new DDIMM connects to
the same interfaces in the Power E1080 nodes. DDR5 is designed to run at higher speeds,
and the new DDIMMs all run at 4000 MHz, regardless of the DIMM capacity. This DDIMM is
differentiated from the existing DDR4-based DDIMMs where the larger DDIMMs ran at a
slower speed than the smaller capacity DIMMs.
With this new memory speed, IBM added an extra port on the memory buffer to connect to
the DIMMs in the DDIMM. This extra port enables the processor to leverage the new memory
speed, improves memory throughput, and reduces latency.
The DDIMM densities that are supported in the Power E1080 are 32 GB, 64 GB, 128 GB, and
256 GB, and any memory Feature Code provides four DDIMMs. The DDIMMs have their own
voltage regulators and are n+1. redundant. Two Power Management Integrated Circuits
(PMICs) plus two spares are used.
It takes two PMICs to supply all the voltage levels that are required by the DDIMM. On the
DDIMM, four PMICs are used, which consist of two redundant pairs. If one PMIC in each of
the redundant pairs is still functional, the DDIMM is not called for replacement.
As described in 1.5.3, “Memory features” on page 19, the Power E1080 requires the
activation of 50% or more of the installed physical memory.
The Power E1080 offers four different DDIMM sizes in both DDR4 and DDR5 technologies:
32 GB, 64 GB, 128 GB, and 256 GB. When using DDR4 DDIMMS, the 32 GB and 64 GB
DDIMMs run at a data rate of 3200 Mbps and the 128 GB and 256 GB DDIMMs at a data rate
of 2933 Mbps. When using DDR5 DDIMMs, all the DDIMMs run at 4000 MHz regardless of
the capacity. DDR4 and DDR5 technology DDIMMs can be mixed within a system, but not
within a single node.
Important: You may not mix DDR4 and DDR5 memory in a node. DDR5 nodes operate at
3200 Mhz if they are in a system with DDR4 nodes.
Table 2-6 lists the available DDIMM capacities and their related maximum theoretical
bandwidth figures per OMI link and Power10 SCM.
Important: For the best possible performance, it is a best practice that memory is installed
in all memory slots and evenly across all system node drawers and all SCMs in the system.
Balancing memory across the installed system board cards enables memory access in a
consistent manner and typically results in better performance for your configuration.
The Active Memory Expansion (AME) feature is an option that can increase the effective
memory capacity of the system. AME uses the on-chip NX unit to compress or decompress
memory content. This separately priced hardware feature is enabled at the partition (LPAR)
level. The related LPAR profile parameter allows it to configure the compression factor in
accordance with memory capacity needs and performance tradeoff consideration.
DDIMMs cannot be installed in such a way that an odd number of DDIMM slots are populated
behind an SCM. DDIMMs also cannot be installed in a configuration that results in 10 or 14
DDIMM slots being populated behind an SCM. The only valid number of DDIMMs plugged
into slots that are connected to a specific SCM is 8, 12, or 16.
The Power E1080 server must adhere to the following DDIM plug rules:
The minimum memory that is allowed is 1 TB (see 1.5.1, “Minimum configuration” on
page 15).
All DDR4 DDIMMs under each SCM must have the same capacity. DDR5 DDIMMs can be
intermixed on the same SCM with Feature Code EMCM.
DDIMMs must be plugged in as groups of four (quads).
Each DDIMM quad must be the same size and type (identical to each other) and must be
the same as the other DDIMM quads under each SCM when using DDR4 memory, but
they can be different from other DDIMM quads from other SCMs. DDR5 memory
capacities can be mixed within a single SCM with Feature Code EMCM.
For example, in a one system node configuration:
– The first SCM can be connected to 8x32 GB DDIMMS.
– The second SCM can be connected to 8x64 GB DDIMMS.
– The third SCM can be connected to 8x128 GB DDIMMS.
– The Fourth SCM can be connected to 8x256 GB DDIMMS.
Overall system performance improves when all the quads match each other. Physical
memory features from a previous generation of servers cannot be migrated to Power E1080.
For the best possible performance, it is recommended that memory is installed in all memory
slots and evenly across all system node drawers and all SCMs in the system.
Balancing memory across the installed system board cards enables memory access in a
consistent manner and typically results in better performance for your configuration.
Account for any plans for future memory upgrades when you decide which memory feature
size to use at the time of the initial system order.
Spreading the DDIMMs evenly between the two memory controllers provides maximum
performance. The DDIMMs within each quad must be the same size and type and all the
DDIMMs under each SCM must be the same capacity.
A DDIMM quad can be different from the DDIMM quads plugged elsewhere.
For the next quad of DDIMMs to be installed in slots that are connected to one of the four
pSCMs, consider the following rules:
The DDIMMs within each quad must be the same size and type.
All DDIMMs under each SCM must have the same capacity.
The DDIMM quad can be different from the DDIMM quads plugged elsewhere, including
the other quad connected to the same SCM.
Plug the remaining quad of DDIMMs in any of the remaining open quad of DIMM slots and
all DDIMMs under each SCM must have the same capacity
Adhere to the following order and use Figure 2-13 on page 68 as reference:
C52, C53, C54, C55 (slots connected to SCM P1) or
C60, C61, C62, C63 (slots connected to SCM P3) or
C68, C69, C70, C71 (slots connected to SCM P2) or
C76, C77, C78, C79 (slots connected to SCM P0) or
C56, C57, C58, C59 (slots connected to SCM P1) or
C64, C65, C66, C67 (slots connected to SCM P3) or
C72, C73, C74, C75 (slots connected to SCM P2) or
C80, C81, C82, C83 (slots connected to SCM P0)
Note: In a multi-system node configuration, a drawer can be fully populated. The other
drawers can be partially populated or have all the drawers symmetrically populated.
All installed processors and memory on systems in a pool are activated and made available
for immediate use when a pool is started. Processor and memory usage on each server are
tracked by the minute and aggregated across the pool.
The capacity in this model consists of Base Activations and Capacity Credits, which are
shared across the pool without having to move them from server to server. The unpurchased
capacity in the pool can be used on a pay-as-you-go basis.
Figure 2-15 IBM Power Private Cloud with Shared Utility Capacity
IBM Power Private Cloud with Shared Utility Capacity solution is supported only on specific
Power9 and Power10 processor-based systems. Power E1080 servers can co-exist with
Power E980 systems in the same pool.
A single Power Enterprise Pool 2.0 supports up to 2000 virtual machines (VMs); up to 1000
VMs are supported per HMC. At the time of this writing, up to 200 VMs are supported per
Power E1080 server, which is planned to be increased to 750 VMs per Power E1080.
In addition to processor and memory metering, Power Enterprise Pools 2.0 enables metering
AIX and IBM i software entitlements and SUSE Linux Enterprise Server and Red Hat
Enterprise Linux (RHEL) subscriptions in the pool.
For more information and requirements, see IBM Power Systems Private Cloud with Shared
Utility Capacity: Featuring Power Enterprise Pools 2.0, SG24-8478.
Table 2-7 lists the Static, Mobile, and Base processor activation features that are available for
initial order on the Power E1080 server.
Table 2-8 lists the Static, Mobile, and Base memory activations that are available for initial
order on the Power E1080 server.
Active processor cores or memory units are processor cores or memory units that are
available for use on your server when it comes from the manufacturer. Inactive cores or units
are processor cores or memory units that are included with your server, but not available for
use until you activate them.
With the CUoD offering, you can purchase more static processors or memory capacity and
dynamically activate them without restarting your server or interrupting your business. All the
static processor or memory activations are restricted to a single server.
CUoD features several benefits that enable a more flexible environment. One benefit is
reducing the initial investment in a system. Traditional projects that use other technologies
mean that a system must be acquired with all the resources available to support the entire
lifecycle of the project. As a result, you pay up front for capacity that you do not need until the
later stages of the project or possibly at all, which affects software licensing costs and
Software Maintenance (SWMA).
By using CUoD, a company starts with a system with enough installed resources to support
the entire project lifecycle, but uses only active resources that are necessary for the initial
project phases. Resources can be added as the project proceeds by activating resources as
needed. Therefore, a company can reduce the initial investment in hardware and acquire
software licenses only when they are needed for each project phase, which reduces the total
cost of ownership (TCO) and total cost of acquisition (TCA) of the solution.
Figure 2-16 shows a comparison between two scenarios: a fully activated system versus a
system with CUoD resources being activated along the project timeline.
80
Projected
70
60
50
Without CuOD
Cores 40
With CuOD
30
20
10
0
Time
Core Activations
Processors and memory can be activated and turned off an unlimited number of times when
more resources are needed.
This offering provides a system administrator an interface at the HMC to manage the
activation and deactivation of resources. Before temporary capacity is used on your server,
you must enable your server. To enable your server, an enablement feature (MES only) must
be ordered and the required contracts must be in place. The 90-day enablement feature is not
available for the Power E1080 processors.
For more information about enablement and usage of Elastic CoD, see Power Capacity on
Demand.
IBM Power Enterprise Pools 1.0 is a technology for dynamically sharing processor and
memory activations among a group (or pool) of IBM Power servers. By using Mobile CoD
activation codes, the systems administrator can perform tasks without contacting IBM.
With this capability, you can move resources between Power E1080 and Power E980 systems
and have unsurpassed flexibility for workload balancing and system maintenance. The
supported Power Enterprise Pool 1.0 members are listed by pool type in Table 2-9.
Note: Only two consecutive generations of Power servers are supported in the same
Power Enterprise Pools 1.0; therefore, Power E1080 can be mixed with Power E980
servers only.
Table 2-9 Supported Power Enterprise Pool 1.0 members by pool type
Power Enterprise Pool type Pool members
Midrange Power Enterprise Pool 1.0 770+, E870, E870C, and E880C
High-end Power Enterprise Pool 1.0 780+, 795, E880, E870C, and E880C
Power Enterprise Pool 1.0 E870, E880, E870C, E880C, and E980
Power Enterprise Pool 1.0 (with Power E1080) E980 and E1080
A pool can support systems with different clock speeds or processor generations.
An HMC can manage multiple IBM Power Enterprise Pools and systems that are not part of
an IBM Power Enterprise Pool. Systems can belong to only one IBM Power Enterprise Pool at
a time. Powering down an HMC does not limit the assigned resources of participating
systems in a pool, but does limit the ability to perform pool change operations.
After an IBM Power Enterprise Pool is created, the HMC can be used to perform the following
tasks:
Mobile CoD processor and memory resources can be assigned to systems with inactive
resources. Mobile CoD resources remain on the system to which they are assigned until
they are removed from the system.
New systems can be added to the pool and existing systems can be removed from the
pool.
New Mobile CoD processor and memory resources can be added to the pool.
Pool information can be viewed, including pool resource assignments, compliance, and
history logs.
For the mobile activation features to be configured, an IBM Power Enterprise Pool and the
systems that are going to be included as members of the pool must be registered with IBM.
Also, the systems must have #EB35 for mobile enablement configured, and the required
contracts must be in place.
Note: The following convention is used in the Order type column in the tables in this
section:
Initial: Only available when ordered as part of a new system
MES: Only available as a MES upgrade
Both: Available with a new system or as part of an upgrade
Supported: Unavailable as a new purchase, but supported when migrated from another
system or as part of a model conversion
Table 2-10 lists the mobile processor and memory activation features that are available for the
Power E1080 server.
An exception request for Trial CoD requires you to complete a form that includes contact
information and VPD from your Power E1080 server with inactive CoD resources. An
exception request activates all inactive processors or all inactive memory (or all inactive
processor and memory) for 30 days. An exception request can be made only once over the
life of the machine. An HMC is required to manage Trial CoD activations.
To request either a Standard or an Exception Trial, see Power Capacity on Demand: Trial
Capacity on Demand.
The PHBs provide different type of end-point connections. Both the PHBs from Power10
sockets P0 and P1 and one PHB from sockets P2 and P3 connect directly to a PCIe Gen 4
x16 or PCIe Gen 5 x8 slot to provide six PCIe Gen 4 x16 or PCIe Gen 5 x8 slots per system
node.
The following theoretical maximum I/O bandwidths in each system node are available:
Six PCIe Gen 4 x16 / PCIe Gen 5 x8 slots at 64 GBps = 384 GBps
Two PCIe Gen 5 x8 Slots at 64 GBps = 128 GBps
Four NVMe slots at 16 GBps = 64 GBps
The rear view of a Power E1080 system node is as shown in Figure 2-17 and Figure 2-18 on
page 81.
EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE
EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE
C0 C0 C0 C0 C0 C0 C0 C0
! ! ! !
! ! ! !
C0 C1 C2 C3
C0 (x16) C1 (x16) C2 (x16) C3 (x16 x8) ! C4 (x16 x8) C5 (x16) C6 (x16) C7 (x16)
P1
T1 T1
! !
C8 C9
E4 E3 !
ND ! E2 E1
1
Figure 2-17 Rear view of a Power E1080 system node highlighting PCIe slots
Adapters that are used in the Power E1080 server are enclosed in a specially designed
adapter cassette. The cassette includes a special mechanism for holding the adapter within a
socket and helps for insertion and removal of the adapter into a PCIe slot at the rear side of
the server.
The cassette locations are P0-C0 - P0-C7. The adapter locations are P0-C0-C0 - P0-C7-C0.
All slots are of the low-profile (LP) type; that is, half-height, half-length (HHHL). They support
Enhanced Error Handling (EEH) and can be serviced with the system power turned on. The
slots also support Gen 1 to Gen 3 adapters.
All the PCIe slots support Single Root I/O Virtualization (SR-IOV) adapters.
Currently, no adapter can dissipate more than 25 watts. Special considerations are needed to
exceed this limit.
The Power E1080 server supports concurrent maintenance (hot plugging) of PCIe adapters.
The adapter cassette is provided with three LEDs that can indicate power and activity (green),
identify function (amber) for the adapter (labeled C0), and fault function LED (amber) for the
cassette. The server can be located by using the blue identify LED on the enclosure.
PHB0 P0-C7-C0
P10 (1)
PHB1 P0-C6-C0
P0-C13
PHB0 P0-C5-C0
P10 (3) PHB1 P0-C4-C0
P0-C14
NVMe Slot P1-C3
Figure 2-19 Power E1080 system node to PCIe slot internal connection schematics
Slot locations and descriptions for the Power E1080 servers are listed in Table 2-12.
PCIe x16 cards are supported only in PCIe x16 slots and the slot priority is (1, 3, 7, 2, 6, 8).
PCIe adapters with x8 and lower lanes are supported in all the PCIe slots and the slot priority
is (1, 7, 3, 6, 2, 8, 4, 5).
Place 8x and 16x adapters in same size slots first before mixing connector size with slot size.
Adapters with smaller connectors are allowed in larger PCIe slot sizes, but larger connectors
are not compatible in smaller PCIe slots sizes.
The system nodes allow for eight PCIe slots of which six are PCIe Gen4 x16 or PCIe Gen5 x8
slots and two are PCIe Gen5 x8 slots. Slots can be added by attaching PCIe expansion
drawers; serial-attached SCSI (SAS) disks can be attached to EXP24S small form factor
(SFF) Gen2 expansion drawers. and more NVMe drives can be added by using the NED24
NVMe expansion drawer. For more information on expansion drawers, see 1.6, “I/O drawers”
on page 25.
The PCIe expansion drawer and the NED24 NVMe expansion drawer are connected by using
an #EJ24 adapter. The EXP24S storage enclosure can be attached to SAS adapters on the
system nodes or on the PCIe expansion drawer.
For a list of adapters and their supported slots, see 2.6, “Supported PCIe adapters” on
page 89.
Disk support: SAS disks that are directly installed on the system nodes and PCIe
Expansion Drawers are not supported. If directly attached SAS disks are required, they
must be installed in a SAS disk drawer and connected to a supported SAS controller in one
of the PCIe slots.
As a best practice, the adapters that do not require high bandwidth should be placed in an
external expansion drawer. Populating a low-latency, high-bandwidth slot with a
low-bandwidth adapter is not the best use of system resources.
As a best practice, use the high-profile version of all adapters whenever the system includes a
PCIe expansion drawer. This configuration allows the user to place most of the adapters
within the PCIe expansion drawer. Typically, it is advised to use node slots for the cable
adapters (#EJ24) or high bandwidth PCIe Gen 4 or PCIe Gen 5 adapters.
EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE
EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE EE
C0 C0 C0 C0 C0 C0 C0 C0
! ! ! !
! ! ! !
C0 C1 C2 C3
C0 (x16) C1 (x16) C2 (x16) C3 (x16 x8) ! C4 (x16 x8) C5 (x16) C6 (x16) C7 (x16)
P1
T1 T1
! !
C8 C9
E4 E3 !
ND ! E2 E1
1
Figure 2-20 Rear view of a E1080 system node highlighting NVMe slots
The internal 7 mm NVMe devices support only one namespace. The primary use of the
devices is to host VIOSs. None of the 7 mm devices support IBM i, IBM i support requires the
#EJBD or EJBE backplane and 15 mm NVMe devices.
The location of the NVMe U.2 drives in the system node is shown in Figure 2-21 on page 85.
Each slot is driven from a x4 PCIe connection, with two SSDs that are connected to the PHB
on the P2 CPU socket and two SSDs that are connected to the PHB of the P3 CPU socket, as
shown in Figure 2-19 on page 82. The buses are routed through the system board from the
CPU to the NVMe slot interface card, and then up through the SSD riser card and into the
SSD.
The NVMe slots support U.2 NVMe flash SSD drives. Four 2.5-inch 7 mm form-factor SSDs
or two 2.5-inch 15 mm NVMe SSDs can be used. To determine which drives are available for
the Power E1080, see the E1080 Sales Manual.
Slot locations P1 - C2 and P1 - C3 remain empty if #EJBD backplane is selected to allow for
spacing of the 15 mm option.
Important: Choose the type of NVMe SSD drives carefully based on your planned usage.
Match the drive characteristics such as DWPD and number of namespaces to your
intended use case.
The location of the NVMe U.2 drive LEDs in the system is shown in Figure 2-22.
Figure 2-22 Rear view of a Power E1080 system node with the NVMe slot LED location
You can find the remaining life on a NVMe device from the LPAR that owns the device.
To determine the remaining life of an NVMe device, complete the following steps:
For an IBM AIX operating system:
a. From the AIX CLI, enter diag and press Enter.
b. From the Function Selection menu, select Task Selection → NVMe general health
information.
c. Select the NVMe device that you want to check the remaining life for and press Enter.
d. View the Percentage of NVM subsystem life used field.
For a Linux operating system:
a. From the Linux CLI, enter the following command and press Enter:
nvme smart-log /dev/nvmeX -H
Where nvmeX is the resource name of the NVMe device.
b. View the Percentage used field.
When the operating system writes data to a read-only device, the write operations are
rejected, and the operating system considers the device as if a failure occurred. To support
normal write operations, the NVMe device must be replaced.
A USB Cable (#EC6N) is used for connecting the #EC6J adapter port to the rear side USB
port in the system control unit (SCU), which is internally routed to a front-accessible USB port
on the SCU (see Figure 2-23).
Figure 2-23 Rear view of the SCU showing the USB port for connection
The following USB devices are supported and available for the Power E1080:
#EUA5: Stand-alone USB DVD Drive with cable
#EUA4: RDX USB External Docking Station
The front view of the SCU in which the System USB port and Flexible Service Processor
(FSP) USB ports are highlighted is shown in Figure 2-24.
Figure 2-24 Front view of SCU showing the System USB port and FSP USB ports
You can connect up to four I/O expansion drawer features EMX0 per node with the slot
capacity that is listed in Table 2-13.
Table 2-13 PCIe slots availability for different system node configurations
System node number I/O expansion number LP slots Full-height slots
1 4 8 48
2 8 16 96
3 12 24 144
4 16 32 192
Each I/O expansion drawer consists of two Fanout Module features EMXH, each providing six
PCIe slots. Each Fanout Module connects to the system by using a pair of CXP cable
features. The CXP cable features are listed in Table 2-14. The RPO Only cables in this list are
not available for ordering new or MES upgrade but for migration from a source system. Select
a longer length feature for inter-rack connection between the system node and the expansion
drawer.
The one pair of CXP optical cables connects to a system node by using one 2-ports PCIe
optical cable adapter feature EJ24 placed in the CEC.
The CXP optical cable pair and the optical cable adapter features are concurrently
maintainable. Therefore, careful balancing of I/O, assigning adapters through redundant
EMX0 expansion drawers, and different system nodes can help ensure high availability for the
I/O resources that are assigned to partitions.
For more information about internal buses and the architecture of internal and external I/O
subsystems, see 2.5, “Internal I/O subsystem” on page 79.
The Order type table column in the following subsections is defined as:
Initial Denotes the orderability of a feature for only the new purchase of the
system.
MES Denotes the orderability of a feature for only the MES upgrade
purchases on the system.
Both Denotes the orderability of a feature for both new and MES upgrade
purchases.
Supported Denotes that a feature is not orderable, but is supported; that is, the
feature can be migrated only from existing systems.
5899 576F PCIe2 4-port 1 GbE Adapter 0 192 AIX, IBM I, Supported
and Linux
EC2U 58FB PCIe Gen 3 2-Port 25/10 Gb 0 192 AIX, IBM i, Both
NIC&ROCE SR/Cu Adaptera and Linux
EN0S 2CC3 PCIe2 4-Port (10 Gb+1 GbE) 0 192 AIX, IBM i, Supported
SR+RJ45 Adapter and Linuxb
EN0T 2CC3 PCIe2 LP 4-Port (10 Gb+1 GbE) 0 32 AIX, IBM i, Supported
SR+RJ45 Adapter and Linuxa
EN0W 2CC4 PCIe2 2-port 10/1 GbE BaseT 0 192 AIX, IBM i, Both
RJ45 Adapter and Linuxb
EN0X 2CC4 PCIe2 LP 2-port 10/1 GbE BaseT 0 32 AIX, IBM i, Both
RJ45 Adapter and Linuxa
EN24 EC2A PCIe4 LPX 4-Port 25/10/1 GbE 0 32 AIX, IBM i, Both
RoCE SFP28 Adapter and Linux
EN25 EC2A PCIe4 4-Port 25/10/1 GbE RoCE 0 32 AIX, IBM i, Both
SFP28 Adapter and Linux
EC2W 2F04 PCIe3 4-port 10GbE BaseT RJ45 0 64 AIX, IBM i, Both
Adapter and Linuxb
EJ0J 57B4 PCIe Gen 3 RAID SAS 0 128 AIX, IBM i, Both
Adapter Quad-port 6 Gb x8 and Linux
The PCIe Gen4 I/O Expansion Drawer has two redundant, hot-plug power supplies. Each
power supply has its own separately ordered power cord. The two power cords plug in to a
power supply conduit that connects to the power supply. The single-phase AC power supply is
rated at 1025 W and can use 100 - 127 V or 200 - 240 V. It is a best practice that the power
supply connects to a power distribution unit (PDU) in the rack. IBM Power PDUs are designed
for a 200 - 240 V electrical source.
A blind-swap cassette (BSC) is used to house the full-height adapters that are installed in
these slots. The BSC is the same BSC that is used with previous generation PCIe Gen 3 12X
attached I/O drawers (#5802, #5803, #5877, #5873, and #EMX0). The drawer includes a full
set of BSCs, even if the BSCs are empty.
Concurrent repair, and adding or removing PCIe adapters, is done by HMC-guided menus or
by operating system support utilities.
IBM PCIe Gen4 I/O expansion drawer BSCs are mechanically the same as the previous PCIe
Gen3 I/O expansion drawers, but the color is different for the touch points. Instead of the
terracotta color, the BSCs in the Gen4 I/O expansion drawer are blue. For that reason,
cassettes from Gen3 I/O expansion drawers should not be moved to the Gen4 I/O expansion
drawer.
Figure 2-26 on page 95 shows a PCIe Gen4 expansion drawer 3D from the rear.
PCI slots that are available in the PCIe Gen4 expansion drawer
Table 2-23 lists the PCI slots in the PCIe Gen4 I/O expansion drawer that is equipped with two
PCIe4 6-slot fan-out modules.
Table 2-23 PCIe slot configuration in the PCIe Gen4 I/O expansion drawer
Slot Location Code Description
Table 2-24 lists the maximum number of fanout modules that are supported and the total
number of cables per system for each Power E1080 server.
IBM Power E1080 PCIe4 cable adapter (Feature Code EJ24; CCIN 6B92)
The supported cables and part numbers are listed in Table 2-26.
A PCIe Gen4 I/O expansion drawer with two I/O fanout modules is connected to one host
system node through two PCIe4 cable adapters with four expansion drawer cables (two
expansion drawer cable pairs). One pair is used for each of the PCIe4 6-slot fanout modules.
Figure 2-27 on page 97 illustrates the connection of two expansion drawer cable pairs for two
PCIe4 6-slot Fanout Modules.
The PCIe Gen3 I/O Expansion Drawer (#EMX0) is a 4U high, PCIe Gen3-based and
rack-mountable I/O drawer. It offers two PCIe Fanout Modules (#EMXH), each providing six
PCIe Gen3 full-high, full-length slots (two x16 and four x8).
The Power E1080 supports four PCIe Gen3 I/O drawer per system node, which yields a
maximum of eight I/O drawers per 2-node Power E1080 server configuration. One I/O drawer
supports two Fanout Modules that offer six PCIe Gen 3 adapter slots each. This configuration
delivers an extra 48 PCIe Gen 3 slot capacity per system node and a maximum of 96 PCIe
Gen 3 slots per 2-node server.
All eight slots in a system node can be used to cable the four I/O drawers. It is always a best
practice to balance the I/O Drawer connectivity across the available system nodes.
With the availability of 3-node and 4-node Power System E1080 configurations in the later GA
period, the number of PCIe Gen 3 I/O drawers scales with a maximum of 16 per 4-node
Power E1080. A maximum of 48 PCIe Gen 3 slots per system node and a maximum of 192
PCIe Gen 3 slots per 4-node Power E1080 server to be available at that date.
Each Fanout Module in the PCIe Gen 3 I/O Expansion Drawer (#EMX0) is attached by one
optical cable adapter, which occupies one x16 PCIe Gen 4 slot of a system node.
4 The PCIe Gen 3 I/O Expansion drawer was withdrawn from marketing.
Table 2-27 Maximum number of supported I/O drawers and the total number of PCIe slots
System nodes Maximum #EMX0 Total number of slots
drawers
PCIe Gen 3, PCIe Gen 3, Total PCIe
x16 x8 Gen 3
The older Fanout Modules (#EMXF and #EMXG) that are used by Power E870, Power
E870C, Power E880, and Power E880C systems cannot be ordered and is not supported by
the Power E1080.
For more information about the dimensions of the drawer, see 1.6, “I/O drawers” on page 25.
PCIe Gen 4 cable adapter (#EJ24) CXP 16X AOCs connects the system node to a PCIe
Fanout Module in the I/O expansion drawer.
Concurrent repair and addition or removal of PCIe adapters is done by using HMC-guided
menus or operating system support utilities.
A BSC is used to house the full-high adapters that are installed into these slots. The BSC is
the same BSC that was used with the previous generation server’s #5802, #5803, #5877, and
#5873 12X attached I/O drawers.
Figure 2-28 on page 99 shows the rear view of the PCIe Gen3 I/O Expansion Drawer.
Although these cables are not redundant, the loss of one cable reduces the I/O bandwidth
(that is, the number of lanes that are available to the I/O module) by 50%.
A total of eight PCIe Gen 4 cable adapters (#EJ24) are allowed per system node to connect to
a maximum of eight PCIe Gen 3 6-slot Fanout Modules (#EMXH). Each Fanout Module must
go with one CXP 16x AOC cable pair with both ports wired to the same PCIe Gen 4 cable
adapter from the system node.
The cable adapter can be placed in any slot in the system node. The following adapter
placement sequence in the system node is recommended:
The maximum number of adapters that are supported is eight per system node and 32 per
system, having four system nodes.
Three different types of CXP 16x AOC cable pairs are available for the connection between
PCIe Gen 4 cable adapter (#EJ24) and PCIe Gen 3 fanout module (#EMXH) of the I/O
Expansion Drawer:
2.0 M (#ECCR)
This cable is used when I/O drawer is in the same rack as the E1080 server.
10.0 M (#ECCY)
This cable is used when the I/O drawer is in a different rack from the E1080 server.
20.0 M (#ECCZ)
Like the 10.0 M (#ECCY) cable, these cables are used when I/O drawer is in a different
rack from the E1080 server.
Note: One #ECCR, one #ECCY, or one #ECCZ includes two AOC cables.
Cable lengths: Use the 2.0 m cables for intra-rack installations. Use the 10.0 m or 20.0 m
cables for inter-rack installations.
Older cables #ECC6, #ECC8, and #ECC9 that are used in Power E980 are not supported or
orderable on Power E1080. They are supported only as model conversions and announced
as RPO.
Drawer connections: Each Fanout Module in a PCIe Gen 3 Expansion Drawer (#EMX0)
can be connected to only a single PCIe Gen 4 Cable Adapter (#EJ24). However, the two
Fanout Modules in a single I/O expansion drawer can be connected to different system
nodes in the same server.
Figure 2-30 shows the connector locations of the PCIe Gen3 I/O Expansion Drawer.
Figure 2-30 Connector locations for the PCIe Gen3 I/O Expansion Drawer
Table 2-28 lists the PCIe slots in the PCIe Gen3 I/O Expansion Drawer.
Table 2-28 PCIe slot locations and descriptions for the PCIe Gen3 I/O Expansion Drawer
Slot Location code Description
All slots support full-length, regular-height adapters or short (LP) adapters with a
regular-height tail stock in single-wide, Gen3 BSCs.
Slots C1 and C4 in each PCIe Gen 3 6-slot Fanout Module are x16 PCIe Gen 3 buses,
and slots C2, C3, C5, and C6 are x8 PCIe buses.
All slots support EEH.
All PCIe slots are hot-swappable and support concurrent maintenance.
Table 2-29 PCIe adapter slot priorities for the PCIe Gen3 I/O Expansion Drawer
Feature Code Description Slot priorities
For more information about slot priorities for individual adapters, see IBM Documentation.
If the EMX0 PCIe Gen 3 Expansion Drawer is configured with two PCIe Gen 3 6-slot Fanout
Modules, distribute the PCIe adapters across both I/O modules whenever possible.
Slots 1, 7, 4, and 10 support SR-IOV capabilities. However, if the total amount of physical
memory is less than 128 GB, Slot 4 (P1-C4) and Slot 10 (P2-C4) are not SR-IOV capable.
Figure 2-31 on page 103 shows the connection schematics between system node and PCIe
Gen3 I/O Expansion Drawer.
P1-C1
Fanout module 1
PCIe4 cable adapter P1-C2
T1 P1-T2 P1-C3
EJ24
T0 P1-T1 P1-C4
P1-C5
P1-C6
P2-C1
Fanout module 2
PCIe4 cable adapter P2-C2
T1 P2-T2 P2-C3
EJ24
T0 P2-T1 P2-C4
P2-C5
P2-C6
Figure 2-31 Connection schematics between the system node and the Expansion Drawer
General rules for the PCIe Gen3 I/O Expansion Drawer configuration
The PCIe Gen 4 Cable Adapter (#EJ24) can be in any of the PCIe adapter slots in a
Power E1080 system node. However, it is a best practice to first populate the PCIe adapter
slots according to the slot priorities that are discussed in this section. If an I/O expansion
drawer is present, the PCIe Gen 4 cable adapter (#EJ24) must be given the highest priority.
Each processor module drives two PCIe Gen5 slots, and all slots are equal regarding their
bandwidth characteristics. If you first use the slots following the slot priorities, you help ensure
that one PCIe Gen5 slot per processor module is populated before you use the second PCIe
Gen5 slot of the processor modules.
Table 2-30 lists the PCIe adapter slot priorities in the Power E1080 server. If the sequence is
chosen as listed in the slot priority column, the adapters are assigned to the SCM in
alignment with the internal enumeration order: SCM0, SCM1, SCM2, and SCM3.
Figure 2-32 shows an example of a Power 1080 that features a single system node and four
PCIe Gen3 I/O Expansion Drawers. Each lane between the #EJ24 adapter and I/O drawer
fanout module represents a CXP 16x AOC pair.
Left
Slot 5 (EJ24) Fanout
Right
Slot 4 (EJ24) Fanout
Slot 3 (EJ24)
I/O Expansion Drawer 3
Slot 2 (EJ24)
Left
Fanout
Slot 1 (EJ24)
Right
Fanout
Right
Fanout
Figure 2-32 Single system node to four PCIe Gen3 I/O Expansion Drawer connections
Figure 2-33 on page 105 shows an example of a Power E1080 having two system nodes and
four PCIe Gen3 I/O Expansion Drawers. Each lane between the #EJ24 adapter and I/O
drawer fanout module represents a CXP 16x AOC pair.
Right
Slot 6 (EJ24)
System node 1 Fanout
Slot 5
Right
Slot 6 (EJ24)
System node 2
Fanout
Slot 5
Slot 2 Right
Fanout
Slot 1 (EJ24)
Figure 2-33 Dual-system node to four PCIe Gen3 I/O Expansion Drawer connections
Each of the 24 NVMe bays in the NED24 drawer is separately addressable, and each can be
assigned to a specific LPAR or VIOS to provide native boot support for up to 24 partitions. At
the time of writing, each drawer can support up to 153 TB.
Figure 2-34 shows a view of the front of the NED24 NVMe Expansion Drawer.
Up to 24 U.2 NVMe devices can be installed in the NED24 drawer by using 15 mm Gen3
carriers. The 15 mm carriers can accommodate either 7 mm or 15 mm NVMe devices. The
devices that are shown in Table 2-31 are supported in the NED24 drawer at the time of
writing.
Table 2-31 Devices that are supported in the NED24 Expansion Drawer
Feature Description
ES3H Enterprise 800 GB SSD PCIe4 NVMe U.2 module for AIX/Linux
ES3A Enterprise 800 GB SSD PCIe4 NVMe U.2 module for IBM i
ES3B Enterprise 1.6 TB SSD PCIe4 NVMe U.2 module for AIX/Linux
ES3C Enterprise 1.6 TB SSD PCIe4 NVMe U.2 module for IBM i
ES3D Enterprise 3.2 TB SSD PCIe4 NVMe U.2 module for AIX/Linux
ES3E Enterprise 3.2 TB SSD PCIe4 NVMe U.2 module for IBM i
ES3F Enterprise 6.4 TB SSD PCIe4 NVMe U.2 module for AIX/Linux
ES3G Enterprise 6.4 TB SSD PCIe4 NVMe U.2 module for IBM i
Each NED24 NVMe Expansion Drawer contains two redundant AC power supplies. The AC
power supplies are part of the enclosure base.
Power10 servers
The NED24 drawer is supported in the Power E1080 by using the same interconnect card
that is used for the PCIe Gen 4 and PCIe Gen 3 expansion drawers. A maximum of three
NED24 NVMe expansion drawers is supported per system node in the E1080. When mixing
the different expansion drawers, the maximum number of drawers that are supported is based
on the number of EJ24 fanout cards that are supported.
Two PCIe4 cable adapters are required to connect each NED24 drive enclosure. This adapter
is available as Feature Code EJ24. This adapter is the same one that is used to connect the
PCIe Gen3 I/O expansion drawer. For more information about the EJ24 adapters, see 1.6.1,
“System node PCIe interconnect features” on page 25.
Installation considerations
This section describes installation considerations for installing and connecting the NED24
drawer to your Power10 scale-out server.
Both CXP Converter adapters require one of the following cable features:
#ECLR - 2.0 M AOC x16 Pair for PCIe4 Expansion Drawer
#ECLS - 3.0 M CXP x16 Copper Cable Pair for PCIe4 Expansion Drawer
#ECLX - 3.0 M AOC x16 Pair for PCIe4 Expansion Drawer
#ECLY - 10 M AOC x16 Pair for PCIe4 Expansion Drawer
#ECLZ - 20 M AOC x16 Pair for PCIe4 Expansion Drawer
Note: Each Feature Code provides two cables that connect from the server adapter to one
of the ESMs. The same Feature Code should be used to connect the second server
adapter to the other ESM. Each drawer requires two identical cable Feature Codes to
connect.
VIOS [Link]
Firmware requirements
The minimum system firmware level that is required to support the NED24 drawer is FW1040,
which requires HMC 10.2.1040 or later. When running with FW1040 or FW1050, the NED24
drawer runs in single-path mode. In single-path mode, each drive has a single connection to
one of the ESMs. As a best practice, use OS mirroring to provide high availability. Multipath
connectivity is available starting with FW1060. For more information, see “Multipath support”.
.
Important: The NED24 requires FW1040 or later to be installed on the system that is
connected. If you are running FW 1040, then the following adapters are not supported by
FW1040, and are not concurrently installable with the NED24 drawer with FW 1040.
PCIe3 12 Gb x8 SAS Tape HBA adapter(#EJ2B/#EJ2C)
PCIe4 32 Gb 4-port optical Fibre Channel adapter (#EN2L/#EN2M)
PCIe4 64 Gb 2-port optical Fibre Channel adapter (#EN2N/#EN2P)
Mixed DDIMM support for the Power E1050 server (#EMCM)
100 V power supplies support for the Power S1022s server (#EB3R)
Multipath support
From initial availability with firmware levels FW 1040 and FW 1050, only mode 1 (single
connect) is supported for the NED24 NVMe Expansion drawer. In mode 1, only one of the
ports on the dual-port NVMe drives is enabled and connected to one of the ESMs. As a best
practice, use OS mirroring for critical devices because there is a single point of failure. The
switch in each of the ESMs is configured to logically drive only 12 of the 24 NVMe drives. No
device failover capability is available.
Starting with FW 1060, the NED24 NVMe drawer supports multipath. The multipath function
supports two connections for each drive because each of the ports on the multiport drives is
connected through both ESMs. This function provides more reliability, availability, and
serviceability (RAS) and better performance.
Multipath is automatically enabled with FW 1060 and enabled when the appropriate OS level
is installed.
Figure 2-36 on page 109 shows the multipath support for the NED24 drawer.
Multipath support is provided in the operating systems that are shown in Table 2-33.
AIX AIX Version 7.3 with the 7300-02 Technology Level and Service Pack
7300-02-02-2420 or later
AIX Version 7.2 with the 7200-05 Technology Level and Service Pack
7200-05-08-2420 or later
AIX Version 7.3 with the 7300-01 Technology Level and Service Pack
7300-01-04-2420 or later
Important: Both ESMs must be connected to the same server. Single connections and
multiple server connections are not supported.
Each of these device paths is seen by AIX as a separate device, for example, nvme0 and
nvme1, as shown in “Drive installation order” on page 111.
For the data on the NVMe drive to be seen through both paths, the NVMe namespace that
you are using must be defined as shared. For load balancing of I/O operations between the
two drive paths, the namespaces should be created as shared and attached from both drive
paths.
In NVMe technology, a namespace (NS) is a logical grouping of data blocks that are
accessible to host software. NVMe supports two namespace types:
Private namespaces are exclusive to the controller where they are created and cannot be
accessed by other controllers within a multi-controller SSD.
Shared namespaces can be accessed from multiple controllers (I/O paths) within an
NVMe subsystem, which provides increased flexibility and redundancy.
Figure 2-39 illustrates a shared namespace (B) that is accessible from both controllers.
Namespace management within AIX can be done through the SMIT interface.
Figure 2-40 shows the suggested placement for the first four drives
1 1 13
2 7 19
3 2 14
4 8 20
5 3 15
6 9 21
7 4 16
8 10 22
9 5 17
10 11 23
11 6 18
12 12 24
Summary
The NED24 drawer provides an excellent method of increasing the internal NVMe storage in
the Power10 processor family and should be considered instead of external SAS. NVMe
provides q lower price per GB than SAS-based enclosures and also provides better
performance.
Electronics service module Dual redundant ESMs with 24 PCIe Gen4 lanes each
Power supply Dual ‘EU Regulation 2019 42' Compliant Power Supply
180 - 264 VAC 50/60 MHz
No DC option
N-1 power and cooling
Hot swappable
Major Field-Replaceable Unit NVMe devices, cable card, ESM, power supply unit (PSU) and
(FRU) parts PDB, cables, and mid-plane
The EXP24SX is a storage expansion enclosure with 24 2.5-inch SFF SAS bays. It supports
10K and 15 K SAS HDD having 512 and 4 K format drives and enterprise and mainstream
class SAS SSDs.
The enclosures can be split logically into one, two, or four independent groups:
One set of 24 bays (mode 1) (D1-D24)
Two sets of 12 bays (mode 2) (D1-D12) (D13-D24)
Four sets of 6 bays (mode 4) (D1-D6) (D7-D12) (D13-D18) (D19-D24)
Figure 2-41 Front view of the ESLS Storage Enclosure with mode groups and drive locations
The following PCIe Gen 3 SAS adapters support the EXP24SX drawers:
PCIe Gen 3 RAID SAS Adapter Quad-port 6 Gb x8 (#EJ0J)
PCIe Gen 3 LP RAID SAS Adapter Quad-Port 6 Gb x8 (#EJ0M)
PCIe3 SAS Tape/DVD Adapter Quad-port 6 Gb x8 (#EJ10)
PCIe3 LP SAS Tape/DVD Adapter Quad-port 6 Gb x8 (#EJ11)
PCIe Gen 3 12 GB Cache RAID Plus SAS Adapter Quad-port 6 Gb x8 (#EJ14)
IBM i configurations require the drives to be protected (RAID or mirroring). Protecting the
drives is highly advised, but not required for other operating systems. All Power server
operating system environments that use SAS adapters with write cache require the cache to
be protected by using pairs of adapters.
Notes: Consider the following points about mixing SSDs and HDDs:
SSDs and HDDs cannot be mixed when configured in mode 1.
SSDs and HDDs can be mixed when configured in mode 2: One disk partition can be
SSDs and the other disk partition can be HDDs, but you cannot mix within a disk
partition.
SSDs and HDDs can be mixed when configured in mode 4. Each disk partition can be
SSDs or HDDs, but cannot mix within a disk partition.
For example, in a mode 2 drawer with two sets of 12 bays, one set can hold SSDs and
one set can hold HDDs, but you cannot mix SSDs and HDDs in the same set of
12-bays.
SAS cabling
The cables that are used to connect an #ESLS storage enclosure to a server are different
from the cables that are used with the 5887 disk drive enclosure. Attachment between the
SAS controller and the storage enclosure SAS ports is through the suitable SAS YO12 or X12
cables. The PCIe Gen3 SAS adapters support 6 Gb throughput. The EXP12SX drawer
supports up to 12 Gb throughput if future SAS adapters support that capability.
Six SAS connectors are at the rear of the EXP24SX drawers to which SAS adapters or
controllers are attached. They are labeled T1, T2, and T3; two T1s, two T2s, and two T3s
connectors. Consider the following points:
In mode 1, two or four of the six ports are used. Two T2 ports are used for a single SAS
adapter, and two T2 and two T3 ports are used with a paired set of two adapters or a dual
adapters configuration.
In mode 2 or mode 4, four ports are used, two T2s and two T3 connectors, to access all
the SAS bays.
The T1 connectors are not used.
Shared Ports
Required
X Cables
T3 2 T3 2 2
T3 T3 T3 T3
3 3 3
4 4 4
T2 5 T2 T2 5 T2 5 T2
T2
6 6 6
7 7 7
8 8 8
9 9 9
T1 10 10 10
T1 T1 T1 T1 T1
11 11 11
12 12 12
13 13 13
14 14 14
15 15 15
16 16 16
17 17 17
18 18 18
19 19 19
20 20 20
21 21 21
22 22 22
23 23 23
24 24 24
ESM1 ESM2 ESM1 ESM2 ESM1 ESM2
IBM EXP24S three modes of operation and the associated disk mapping
Figure 2-42 Rear view of the EXP24SX with location codes and different split modes
Mode setting is done by IBM Manufacturing. If you need to change the mode after installation,
ask your IBM System Services Representative (IBM SSR) for support.
For more information about SAS cabling and cabling configurations, see “Connecting an
#ESLS storage enclosure to your system” in IBM Documentation.
Similar to the previous generation of Power System E980 server, two service processors are
available for redundancy. They are hosted in the SCU and communicate with the system
nodes by using the FRU Service Interface (FSI)/Processor Support Interface (PSI) bus
connectors that are at the rear of the SCU and the system nodes.
All the service processor communication between the control unit and the system nodes flows
though the service processor cables. In comparison to previous generations, the
Power E1080 associated SCU is no longer hosting the system clock. Each system node hosts
its own redundant clocks.
The cables that are necessary for each system node are grouped under a single Feature
Code, which allows for an simpler configuration. Each cable set includes a pair of FSP cables,
and when applicable SMP cables and UPIC cables.
Initial orders of Power E1080 server include one #EFCH, which is required to connect the
system node with SCU. This configuration does not require SMP cables, which are necessary
only for configurations with two or more system nodes.
Cable sets Feature Codes are incremental and depend on the number of installed system
nodes:
One system node: #EFCH
Two system nodes: #EFCH and #EFCE
Three system nodes: #EFCH, #EFCE, and #EFCF
Four system nodes: #EFCH, #EFCE, #EFCF, and #EFCG
The redundant FSP provides proprietary interface communication, such as FSI and PSI to the
system nodes.
PSI is used for FSP-host processor unit communication. PSI is a clock synchronous
bidirectional interface for control communication. Each FSP in the SCU has four PSI
interfaces and are connected such that whichever FSP becomes Primary can control the
entire system.
FSI in the Power E1080 server is a serial point-to-point connection that is used for device
communication in the overall System Control Structure design.
The FSI connection network is across FSP to FSP connections inside the SCU, FSP to
system node through clocking and control logic. They connect FSPs in the SCU to system
node elements and the Power10 to Power10 processor chips on the system node system
board inside the system node. Similar to PSI network, whichever FSP becomes Primary can
control the entire server.
Figure 2-43 on page 117 shows the UPIC and FSP cabling between a single system node
and a SCU.
T0 T2 T4 T6 T8 T10 T12 T14 T16 T18 T20 T22 T24 T26 T28 T30
T1 T3 T5 T7 T9 T11 T13 T15 T17 T19 T21 T23 T25 T27 T29 T31
C8 C9
C1 C2
FSP cable
FSP cable
T2 T3 T4 T5
T6 T7 T8 T9
Figure 2-43 UPIC and FSP connection between a SCU and a single system node
Figure 2-44 shows UPIC and FSP cabling between two system nodes and a SCU.
Figure 2-44 UPIC and FSP connection between a SCU and two system nodes
In a single system node configuration, two UPIC cables connect to the SCU and provide a
redundant power source from the system node. The SCU power source for two or more
system nodes is supplied by the first and second system node. When Power E1080 supports
more than two system nodes, power output ports on the third and forth system node are not
used.
The system reference clock source is responsible for providing a synchronized clock signal to
all functional units. Each system node of a Power E1080 server uses its own private set of two
redundant system clock or control cards. If a failure occurs in any of the clock or control cards,
the second card helps ensure continued operation of the system until a replacement is
scheduled.
Figure 2-45 shows UPIC and FSP cabling between three system nodes and a SCU.
Figure 2-45 UPIC and FSP connection between a SCU and three system nodes
T0 T2 T4 T6 T8 T10 T12 T14 T16 T18 T20 T22 T24 T26 T28 T30
T1 T3 T5 T7 T9 T11 T13 T15 T17 T19 T21 T23 T25 T27 T29 T31
C8 C9
T0 T2 T4 T6 T8 T10 T12 T14 T16 T18 T20 T22 T24 T26 T28 T30
T1 T3 T5 T7 T9 T11 T13 T15 T17 T19 T21 T23 T25 T27 T29 T31
C8 C9
T0 T2 T4 T6 T8 T10 T12 T14 T16 T18 T20 T22 T24 T26 T28 T30
T1 T3 T5 T7 T9 T11 T13 T15 T17 T19 T21 T23 T25 T27 T29 T31
FSP cable
C8 C9
T0 T2 T4 T6 T8 T10 T12 T14 T16 T18 T20 T22 T24 T26 T28 T30
T1 T3 T5 T7 T9 T11 T13 T15 T17 T19 T21 T23 T25 T27 T29 T31
C8 C9
C1 C2
FSP cable FSP cable
T2 T3 T4 T5
FSP cable
T6 T7 T8
T6 T9
FSP cable
System control unit
Figure 2-46 UPIC and FSP connection between a SCU and four system nodes
In addition, the VIOS can be installed in special partitions that provide support to other
partitions running AIX, IBM i, or Linux OSes for the use of various features, such as
virtualized I/O devices, or PowerVM LPM.
For more information about the software that is available on Power servers, see IBM Power
Software.
This tool also helps to plan a successful system upgrade by providing the prerequisite
information for features currently in use or planned to be added to a system. It's possible to
choose a machine type and model (9080-HEX for Power E1080) and find out all the
prerequisites, operating system levels supported and other information.
AIX 7.1 must run on an LPAR in Power8 compatibility mode with VIOS-based virtual storage
and networking. The I/O can be supplied only by the VIOS and the system is LPM enabled,
excluding Native VF (SR-IOV) directly in the customer LPAR.
AIX 7.2 can run with physical and virtual I/O and requires Power9 compatibility mode.
NVMe over Fibre Channel fabrics (NVMe-OF) is now available on Power E1080 running AIX
when the 2-port 32 Gb adapter (#EN1A or #EN1B) and the selected IBM FlashSystem
high-end (IBM FlashSystem 9110) and mid-range (IBM FlashSystem 7200) model are used.
IBM periodically releases maintenance packages (service packs or technology levels) for the
AIX operating system. For more information about these packages, downloading, and
obtaining the CD-ROM, see Fix Central.
The Service Update Management Assistant (SUMA), which can help you automate the task
of checking and downloading operating system downloads, is part of the base operating
system. For more information about the suma command, see IBM Documentation.
A major update for Enterprise Cloud Edition Bundles version includes PowerSC 2.0.
Customers that are purchasing a new Power E1080 can get this stack with their system
purchase. Customers with existing systems and active SWMA by using either of the Bundles
can get this Product by using ESS and start using it immediately.
A subscription license is a licensing model that provides access to an IBM program and IBM
SWMA for a specified subscription term (one or three years). The subscription term begins on
the start date and ends on the expiration date, which is reflected in ESS.
Customers are licensed to run the product through the expiration date of the 1- or 3-year
subscription term and then can renew at the end of the subscription to continue using the
product. This model provides flexible and predictable pricing over a specific term, with lower
up-front entry cost.
Another benefit of this model is that the licenses are customer number entitled, which means
they are not tied to a specific hardware serial number as with perpetual licenses. Therefore,
the licenses can be moved between on-premises and cloud if needed, something that is
becoming more of a requirement with hybrid workloads.
The new Product IDs for the subscription licenses are listed in Table 2-37.
The licenses are orderable through IBM configuration tools. The AIX perpetual and monthly
term licenses for standard edition are still available.
2.10.3 IBM i
IBM i is supported on the Power E1080 server by the following minimum required levels:
IBM i 7.3 TR11 or later
IBM i 7.4 TR5 or later
For compatibility information for hardware features and the corresponding IBM i Technology
Levels, see IBM Prerequisites.
IBM i terms and conditions require that IBM i operating system license entitlements remain
with the machine for which they were originally purchased. Under qualifying conditions, IBM
allows the transfer of IBM i processor and user entitlements from one machine to another.
This capability helps facilitate machine replacement, server consolidation, and load
rebalancing while protecting a customer’s investment in IBM i software. When the
requirements are met, the IBM i license transfer can be configured by using IBM configuration
tools.
Each IBM i processor entitlement that is transferred to a target machine includes one year of
new SWMA at no charge. Extra years of coverage or 24x7 support are available options for
an extra charge.
2.10.4 Linux
The following types of Linux distributions are available to run on Power E1080:
With native Power10 processor technology support
Running in Power9 compatibility mode
Distributions with native Power10 processor technology support can also run in Power9
compatibility mode. This feature is important when doing LPM from a Power9
processor-based server to Power E1080.
Red Hat
RHEL version 8.4 and later can run in native Power10 processor mode. At the time of this
writing, RHEL version 8.4 was available.
SUSE
SUSE 15 SP3 is the first version with native Power10 processor technology support. Its
regular support cycle is 18 months, plus long-term SP support.
Canonical and the Ubuntu community work together with IBM to help ensure that Ubuntu
Server and Ubuntu OpenStack work seamlessly with IBM Power and IBM software
applications.
Starting with Ubuntu 22.04 LTS, Power9 and Power10 processors are supported. For more
information, see Ubuntu on IBM Power.
Older distributions
The following selected older distributions also are supported on Power E1080:
RHEL 8.2 is supported in Power9 compatibility mode only
SUSE Linux Enterprise Server 12 SP5 is supported in Power9 compatibility mode only
When an LPAR runs in Power9 compatibility mode, it benefits from most of the features of the
Power10 processor technology, including the full eight threads per core. However, program
and kernel features that use new Power10 instructions or capabilities are not available.
The Power9 compatibility mode is required when moving partitions back and forth between
Power9 processor based-systems and Power10 processor-based systems. After a partition is
moved to a Power10 processor-based systems, it can be upgraded to a distribution with
native Power10 technology support and restarted in native Power10 mode.
Note: No official support for Power9 compatibility mode is available for older service pack
of SUSE Linux Enterprise Server 15, such as SUSE Linux Enterprise Server 15 SP2.
Enhancements
One of the main features of the Power10 processor chip is the possibility to support up to 15
cores per SCM. Therefore, on a Power E1080 that is configured with up to 240 processor
cores (each capable of running eight threads for up to 1920 possible threads), the Linux
distribution with native Power10 technology support can use this capability.
LPARs running in Power9 compatibility mode are restricted to 1536 threads per LPAR, which
is the maximum for a Power9 processor-based server.
The Power10 specific toolchain is available in Advance Toolchain 15.0, which allows
customers and developers to use all new Power10 processor-based technology instructions
when programming. Cross-module function call overhead was reduced because of a new
PC-relative addressing mode.
Network virtualization is an area with significant evolution and improvements, which benefit
virtual and containerized environments. The following recent improvements were made for
Linux networking features on Power E1080:
SR-IOV allows virtualization of network cards at the controller level without the need to
create virtual Shared Ethernet Adapters (SEAs) in the VIOS partition. It is enhanced with
(vNIC) virtual Network Interface Controller, which allows data to be transferred directly
from the partitions to or from the SR-IOV physical adapter without transiting through a
VIOS partition.
Hybrid Network Virtualization (HNV) allows a partition to use the efficiency and
performance benefits of SR-IOV logical ports and participate in mobility operations, such
as active and inactive LPM and Simplified Remote Restart (SRR). HNV is enabled by
selecting a new Migratable option when an SR-IOV logical port is configured.
NVMe over Fibre Channel fabrics (NVMe-OF) is now available on Power E1080 running
Linux when the 2-port 32 Gb adapter (#EN1A or #EN1B) is and the selected IBM
FlashSystem high-end (IBM FlashSystem 9110) and mid-range (IBM FlashSystem 7200)
model are used.
Security
Security is a top priority for IBM and our distribution partners. Linux security on IBM Power is
a vast topic that can be the subject of detailed separate material; however, improvements in
the areas of hardening, integrity protection, performance, platform security, and certifications
are introduced with this section.
Performance is also a security topic because specific hardening mitigation strategies (for
example, against side-channel attacks), can have a significant performance effect. In
addition, cryptography can use significant compute cycles.
The Power E1080 features transparent memory encryption at the level of the controller, which
prevents an attacker from retrieving data from physical memory or storage-class devices that
are attached to the processor bus.
IBM regularly updates the VIOS code. For more information, see IBM Fix Central.
When system firmware updates are applied to the system, UAK and its expiration date are
checked (see Figure 2-48 on page 128). System firmware updates include a release date.
When attempting to apply system firmware updates, if the release date for the firmware
updates passed the expiration date for the UAK, the updates are not processed. As UAKs
expire, they need to be replaced by using the HMC or the ASMI on the service processor.
Determining when the current UAK runs through the HMC, GUI, or CLI as provided in the
following examples. However, it is also possible to display the expiration date by using the
suitable AIX or IBM i command.
The output is similar to the output that is shown in Example 2-1 (the Microcode Entitlement
Date represents the UAK expiration date).
Example 2-1 Output of the command to check the UAK expiration date through AIX 7.1
$ lscfg -vpl sysplanar0 | grep -p "System Firmware"
System Firmware:
...
Microcode Image.............SV860_138 SV860_103 SV860_138
Microcode Level.............FW860.42 FW860.30 FW860.42
Microcode Build Date........20180101 20170628 20180101
Microcode Entitlement Date..20190825
Hardware Location Code......[Link]-Y1
Physical Location: [Link]-Y1
The output is similar to the output that is shown in Example 2-2 (the UAK Exp Date
represents the UAK expiration date).
Example 2-2 Output of the command to check the UAK expiration date through AIX 7.2
$ lscfg -vpl sysplanar0 |grep -p "System Firmware"
System Firmware:
...
Microcode Image.............SV860_138 SV860_103 SV860_138
Microcode Level.............FW860.42 FW860.30 FW860.42
Microcode Build Date........20180101 20170628 20180101
Update Access Key Exp Date..20190825
Hardware Location Code......[Link]-Y1
Physical Location: [Link]-Y1
If the UAK expired, proceed to the ESS website to replace your UAK. Figure 2-47 shows the
output in the IBM i 7.1 and 7.2 releases. In the 7.3 release, the text changes to Update Access
Key Expiration Date. The line that is highlighted in Figure 2-47 is displayed whether the
system is operating system-managed or HMC managed.
2.11 Manageability
Manageability of the system is a key component for customers to manage, service, test, and
monitor the system performance, security, and reliability.
ASMI is a graphical interface that is part of the service processor firmware. The ASMI
manages and communicates with the service processor. The ASMI is required to set up the
service processor and to perform service tasks, such as reading service processor error logs,
reading VPD, and controlling the system power.
The control panel functions allow you to interface with the server. Control panel functions
range in complexity from functions that display a status (such as IPL speed) to service
functions that only service representatives can access.
For more information about a full list of all functions, see this web page.
The HMC GUI/CLI is the main manageable interface for customers. With HMC V10R1, the
layout and the steps changed from previous versions.
3. In the next window, the Readiness Status column shows whether the system is ready. If
the message is too long, hover around the box to see the entire message.
If two or more systems are selected and one is in a ready state and one is not in a ready
state, the system that is not in a ready state is skipped and only the system with the ready
state progresses (see Figure 2-49 on page 129).
Click Next.
4. The System Firmware Type window opens. Two options are available: update or upgrade
(see Figure 2-50). Click Update and then click Next.
6. The Available Levels for Update window opens. In the target level column, the nature of
the update (disruptive or nondisruptive) is shown. If you want to see what is included in the
update or upgrade and which LPARs are affected, select the system.
From the drop-down menu, select View Cover Data or View Impacted LPARs (see
Figure 2-52) and then click Next.
In the system Firmware Update Progress window, window (see Figure 2-54), the estimated
time for and the progress of the installation are displayed.
For example, the first and second steps of I/O firmware update include the License
Agreement and Repository windows, which are similar to the system firmware update
windows.
The next step of the wizard shows the system name and partition name that owns the I/O
device, the current level, the available level of the related device microcode, and suggested
actions and device description (see Figure 2-55).
2.12 Serviceability
Instances exist in which maintenance procedures are performed on systems and
subsystems, including replacing failed components, engineering change implementations,
MES hardware upgrades, and migrations. Any of these procedures need the system
operations to completely halt or run with a degradation in service.
The productive hours that are lost in performing the service negatively affects the business;
therefore, the service strategy includes plans for not only how efficiently the system can be
brought back to resume its operations normally, but also to reduce the hours that are needed
to service the system.
The success of achieving these goals determines the availability of the system; that is,
efficient serviceability means high availability. Power E1080 server carries forward the rich
legacy of enterprise class Power servers to deliver the best enterprise RAS features
compared to any OEM vendor.
Service interface
Service engineers are assisted by multiple system service interfaces that communicate with
the service support applications in a server that is using the operator console, the GUI that is
on the management console or service processor menu, or an operating system terminal.
The service interface helps the support team to efficiently manage system resources and
service information.
Applications that are available through the service interface are configured and placed to give
service engineers access to important service functions. Depending on the system state,
hypervisor, and the operating environment, one or more service interfaces can be useful in
accessing logs and service information and communicating with the system. The following
primary service interfaces are available:
Light path diagnostics (LPD)
Operator panel
Service processor menu
ASMI
Operating system service menu
SFP on the HMC or vHMC
The system can identify components for replacement by using FRU-specific LEDs. The
service engineer can use the identify function to set the FRU level LED to flash, which lights
the blue enclosure locate and system locate LEDs. The enclosure LEDs turn on solid and
guide the service engineer to follow the light path from the system to the enclosure and down
to the specific FRU in error.
Similar to LPD notifications, other interfaces in the previous bulleted list provide tools to
capture logs/dumps and other essential information to identify and detect errors.
For more information about service interfaces and the available service functions, see this
web page.
Within the processor or memory subsystem error checker, error-checker signals are captured
and stored in hardware FIRs. The associated logic circuitry is used to limit the domain of an
error to the first checker that encounters the error. In this way, runtime error diagnostic tests
can be deterministic so that for every check station, the unique error domain for that checker
is defined and mapped to FRUs that can be repaired when necessary.
First-failure data capture (FFDC) is a technique that helps ensure that the root cause of the
fault is captured without the need to re-create the problem or run any extending tracing or
diagnostics program when a fault is detected in a system. For most faults, a good FFDC
design means that the root cause can also be detected automatically without service
engineers’ intervention.
FFDC information, error data analysis, and fault isolation are necessary to implement the
advanced serviceability techniques that enable efficient service of the systems and to help
determine the failing items.
In the rare absence of FFDC and Error Data Analysis, diagnostics are required to re-create
the failure and determine the failing items.
2.12.2 Diagnostics
Diagnostics refers to identification of errors, symptoms, and determination of the potential
causes of the errors identified.
The Power E1080 server is supplemented with several advanced troubleshooting and
diagnostic routines that are available through multiple service interfaces, as discussed in
2.12.1, “Error detection” on page 133. For more information about the available aids, see the
following IBM Documentation web pages:
Analyzing problems
Isolation procedures
Reference codes
These codes represent the system IPL status progress codes, OS IPL progress codes, dump
progress codes, service request numbers (SRNs), and others, which serve as diagnostic aid
to help determine the source of various hardware errors. Diagnostic applications report
problems with SRNs.
The support team and service engineers use this information with reference code-specific
information to analyze and determine the source of errors or find more information about
other isolation procedures.
Automatic diagnostics
The processor and memory FFDC is designed to perform without the need to re-create the
problems or user intervention. Solid and intermittent errors are detected early and isolated at
the time of failure. Runtime and boot-time diagnostics fall into this category.
2.12.3 Reporting
If a system hardware or environmentally induced failure occurs, the system runtime error
diagnostics analyze the hardware error signature to determine the cause of failure.
The analysis is stored in the system NVRAM. The identified errors are reported to the
operating system and recorded in the system logs of the operating system.
For an HMC-managed system in the PowerVM environment, an ELA routine analyzes the
error, forwards the event to the SFP application that is running on the HMC, and notifies the
system administrator about isolation of the likely cause of the system problem. The service
processor event log also records unrecoverable checkstop conditions and forwards them to
the SFP application.
The system can call home from IBM i and AIX operating systems to report platform
recoverable errors and errors that are associated with PCIe adapters and devices. In an
HMC-managed system environment, a Call Home service request is started from the HMC
and the failure report that carries parts information and part location is sent to the IBM service
organization.
Along with such information, customer contact information and system-specific information,
machine type, model, and serial number, and error logs also are sent to IBM service
organization electronically.
2.12.4 Notification
In this section, we describe the types of notifications that are available.
Call Home
HMC can notify service events that need IBM service organization attention, which uses the
Call Home feature. Call Home helps to transmit error logs, server status, or other
service-related information to the IBM service organization electronically.
This feature is optional, but customers can gain significant advantages implementing the
feature, such as faster problem determination and resolution and, usually, without customer
notice or direct involvement. This feature also allows the IBM Service organization to register
auto service tickets and send customer-replaceable units (CRU) directly to the customer or
dispatching a service engineer to the customer location.
ESA monitors and collects system inventory and service information, which is accessible from
a secured web portal.
For more information about ESA planning, implementation, and configuration, see the
following web pages:
For AIX
For IBM i
Customers who are skeptical about the information that is shared with IBM can be assured
that security protocols are in place. Also, by registering with the IBM Electronics Services
portal, the service information is accessible to customers.
Some of the aids that are available with Power 1080 system are discussed next.
Service labels
Placed at various locations on the system hardware, these labels provide the following
ready-to-use, graphics-based information to perform physical maintenance on the system:
Location diagrams
Typically, these diagrams consist of placement view of various hardware components
(FRUs and CRUs) that are inside the server chassis. This information can help the service
engineer to find the location of the FRUs that are being serviced or removed.
Remove/ or replace procedures
A pictorial aid that shows the way to remove hardware components from the server
chassis. For example, a picture can show thumbs that are pressing the latches, which can
help release the component from its place.
Arrows
Numbered arrows are used to indicate the order of actions and the serviceability direction
of components. Some serviceable parts, such as latches, levers, and touch points, must
be pulled or pushed in a specific direction and in a specific order for the mechanical
mechanisms to engage or disengage.
Physical address diagrams
These diagrams can help map the logical address of failed components to the physical
address so that only intended components can be identified and serviced.
Concurrent maintenance
The Power E1080 includes many physical components that allow concurrent maintenance,
which frees the service engineer to bring the system down for maintenance. The following
concurrent maintainable components of E1080 are available:
EXP24S SAS storage enclosure drawer
Drives in the EXP24S storage enclosure drawer
NVMe U.2 drives
PCIe extender cards, optical PCIe link I/O expansion card
PCIe I/O adapters
PCIe I/O drawers
PCIe to USB conversion card
SMP cables
System node AC power supplies: Two functional power supplies must remain installed
always while the system is operating
System node fans
SCU fans
SCU operations panel
Time of Day clock battery
UPIC interface card in SCU
UPIC power cables from system node to SCU
Note: PowerVM Enterprise Edition License Entitlement is included with each Power E1080
server. PowerVM Enterprise Edition is available as a hardware feature (#5228), supports
up to 20 partitions per core, VIOS, multiple shared processor pools (MSPPs) and also
offers LPM.
Combined with features in the Power E1080, the IBM POWER Hypervisor delivers functions
that enable other system technologies, including logical partition (LPAR) technology,
virtualized processors, IEEE virtual local area network (VLAN)-compatible virtual switch,
virtual SCSI adapters, virtual Fibre Channel adapters, and virtual consoles.
The POWER Hypervisor is a basic component of the system’s firmware and offers the
following functions:
Provides an abstraction between the physical hardware resources and the LPARs that use
them.
Enforces partition integrity by providing a security layer between LPARs.
Controls the dispatch of virtual processors to physical processors.
Saves and restores all processor state information during a logical processor context
switch.
Controls hardware I/O interrupt management facilities for LPARs.
Provides VLAN channels between LPARs that help reduce the need for physical Ethernet
adapters for inter-partition communication.
Monitors the Flexible Service Processor (FSP) and performs a reset or reload if it detects
the loss of one of the FSP, notifying the operating system if the problem is not corrected.
The amount of memory for the HPT is based on the maximum memory size of the partition
and the HPT ratio. The default HPT ratio is 1/128th (for AIX, VIOS, and Linux partitions) of the
maximum memory size of the partition. AIX, VIOS, and Linux use larger page sizes (16 KB
and 64 KB) instead of using 4 KB pages. The use of larger page sizes reduces the overall
number of pages that must be tracked; therefore, the overall size of the HPT can be reduced.
For example, the HPT is 2 GB for an AIX partition with a maximum memory size of 256 GB.
When defining a partition, the maximum memory size that is specified is based on the amount
of memory that can be dynamically added to the dynamic logical partition (DLPAR) without
changing the configuration and restarting the partition.
In addition to setting the maximum memory size, the HPT ratio can be configured. The
hpt_ratio parameter for the chsyscfg Hardware Management Console (HMC) command can
be issued to define the HPT ratio that is used for a partition profile. The valid values are 1:32,
1:64, 1:128, 1:256, or 1:512.
Specifying a smaller absolute ratio (1/512 is the smallest value) decreases the overall
memory that is assigned to the HPT. Testing is required when changing the HPT ratio
because a smaller HPT might incur more CPU consumption because the operating system
might need to reload the entries in the HPT more frequently. Most customers choose to use
the IBM provided default values for the HPT ratios.
For physical I/O devices, the base amount of space for the TCEs is defined by the hypervisor
that is based on the number of I/O devices that are supported. A system that supports
high-speed adapters can also be configured to allocate more memory to improve I/O
performance. Linux is the only operating system that uses these extra TCEs so that the
memory can be freed for use by partitions if the system uses only AIX.
The POWER Hypervisor must set aside save areas for the register contents for the maximum
number of virtual processors that are configured. The greater the number of physical
hardware devices, the greater the number of virtual devices, the greater the amount of
virtualization, and the more hypervisor memory is required. For efficient memory
consumption, wanted and maximum values for various attributes (processors, memory, and
virtual adapters) must be based on business needs, and not set to values that are higher than
actual requirements.
The POWER Hypervisor provides the following types of virtual I/O adapters:
Virtual SCSI
The POWER Hypervisor provides a virtual SCSI mechanism for the virtualization of
storage devices. The storage virtualization is accomplished by using two paired adapters:
a virtual SCSI server adapter and a virtual SCSI customer adapter.
Virtual Ethernet
The POWER Hypervisor provides a virtual Ethernet switch function that allows partitions
fast and secure communication on the same server without any need for physical
interconnection or connectivity outside of the server if a Layer 2 bridge to a physical
Ethernet adapter is set in one VIOS partition, also known as Shared Ethernet Adapter
(SEA).
Virtual Fibre Channel
A virtual Fibre Channel adapter is a virtual adapter that provides customer LPARs with a
Fibre Channel connection to a storage area network through the VIOS partition. The VIOS
partition provides the connection between the virtual Fibre Channel adapters on the VIOS
partition and the physical Fibre Channel adapters on the managed system.
Virtual (TTY) console
Each partition must have access to a system console. Tasks, such as operating system
installation, network setup, and various problem analysis activities, require a dedicated
system console. The POWER Hypervisor provides the virtual console by using a virtual
TTY or serial adapter and a set of hypervisor calls to operate on them. Virtual TTY does
not require the purchase of any other features or software, such as the PowerVM Edition
features.
Logical partitioning is the ability to make a server run as though it were two or more
independent servers. When you logically partition a server, you divide the resources on the
server into subsets, called LPARs. You can install software on an LPAR, and the LPAR runs as
an independent logical server with the resources that you allocated to the LPAR.
LPAR is also referred to in some documentation as a virtual machine (VM), which makes it
look similar to what other hypervisors offer. However, LPARs provide a higher level of security
and isolation and other features that are described in this chapter.
Processors, memory, and I/O devices can be assigned to LPARs. AIX, IBM i, Linux, and VIOS
can run on LPARs. VIOS provides virtual I/O resources to other LPARs with general-purpose
operating systems.
LPARs share a few system attributes, such as the system serial number, system model, and
processor Feature Codes. All other system attributes can vary from one LPAR to another.
Micro-Partitioning
When you use the Micro-Partitioning technology, you can allocate fractions of processors to
an LPAR. An LPAR that uses fractions of processors is also known as a shared processor
partition or micropartition. Micropartitions run over a set of processors that is called a shared
processor pool (SPP), and virtual processors are used to enable the operating system
manage the fractions of processing power that are assigned to the LPAR.
On the Power10 processor-based server, a partition can be defined with a processor capacity
as small as 0.05processing units. This number represents 0.05 of a physical core. Each
physical core can be shared by up to 20 shared processor partitions, and the partition’s
entitlement can be incremented fractionally by as little as 0.05 of the processor. The shared
processor partitions are dispatched and time-sliced on the physical processors under the
control of the Power Hypervisor. The shared processor partitions are created and managed
by the HMC.
The Power E1080 supports up to 240 cores in a single system and 1000 micropartitions
(1000 is the maximum that PowerVM supports).
Note: Although the Power E1080 supports up to 1000 micropartitions, the real limit
depends on application workload demands in use on the server.
Processing mode
When you create an LPAR, you can assign entire processors for dedicated use, or you can
assign partial processing units from an SPP. This setting defines the processing mode of the
LPAR.
Dedicated mode
In dedicated mode, physical processors are assigned as a whole to partitions. The SMT
feature in the Power10 processor core allows the core to run instructions from two, four, or
eight independent software threads simultaneously.
Shared mode
In shared mode, LPARs use virtual processors to access fractions of physical processors.
Shared partitions can define any number of virtual processors (the maximum number is 20
times the number of processing units that are assigned to the partition). The Power
Hypervisor dispatches virtual processors to physical processors according to the partition’s
processing units entitlement. One processing unit represents one physical processor’s
processing capacity. All partitions receive a total CPU time equal to their processing unit’s
entitlement. The logical processors are defined on top of virtual processors. Therefore, even
with a virtual processor, the concept of a logical processor exists, and the number of logical
processors depends on whether SMT is turned on or off.
Micropartitions are created and then identified as members of the default processor pool or a
user-defined SPP. The virtual processors that exist within the set of micropartitions are
monitored by the Power Hypervisor. Processor capacity is managed according to
user-defined attributes.
If the Power server is under heavy load, each micropartition within an SPP is assured of its
processor entitlement, plus any capacity that might be allocated from the reserved pool
capacity if the micropartition is uncapped.
If specific micropartitions in an SPP do not use their processing capacity entitlement, the
unused capacity is ceded and other uncapped micropartitions within the same SPP can use
the extra capacity according to their uncapped weighting. In this way, the entitled pool
capacity of an SPP is distributed to the set of micropartitions within that SPP.
All Power servers that support the MSPP capability have a minimum of one (the default) SPP
and up to a maximum of 64 SPPs.
This capability helps customers reduce TCO significantly when the cost of software or
database licenses depends on the number of assigned processor-cores.
The VIOS eliminates the requirement that every partition owns a dedicated network adapter,
disk adapter, and disk drive. The VIOS supports OpenSSH for secure remote logins. It also
provides a firewall for limiting access by ports, network services, and IP addresses.
Virtual SCSI
Physical Ethernet Virtual Ethernet Adapter
Adapter Adapter
Physical
Disk
Virtual I/O Client 2
Physical Disk Virtual SCSI
Adapter Adapter Virtual Ethernet
Adapter
Physical
Disk Virtual SCSI
Adapter
By using the SEA, several customer partitions can share one physical adapter. You can also
connect internal and external VLANs by using a physical adapter. The SEA service can be
hosted only in the VIOS (not in a general-purpose AIX or Linux partition) and acts as a Layer
2 network bridge to securely transport network traffic between virtual Ethernet networks
(internal) and one or more (Etherchannel) physical network adapters (external). These virtual
Ethernet network adapters are defined by the Power Hypervisor on the VIOS.
Virtual SCSI
Virtual SCSI is used to view a virtualized implementation of the SCSI protocol. Virtual SCSI is
based on a customer/server relationship. The VIOS LPAR owns the physical I/O resources
and acts as a server or, in SCSI terms, a target device. The client LPARs access the virtual
SCSI backing storage devices that are provided by the VIOS as clients.
N_Port ID Virtualization
N_Port ID Virtualization (NPIV) is a technology that allows multiple LPARs to access one or
more external physical storage devices through the same physical Fibre Channel adapter.
This adapter is attached to a VIOS partition that acts only as a pass-through that manages
the data transfer through the Power Hypervisor.
Each partition features one or more virtual Fibre Channel adapters, each with their own pair
of unique worldwide port names. This configuration enables you to connect each partition to
independent physical storage on a SAN. Unlike virtual SCSI, only the client partitions see the
disk.
For more information and requirements for NPIV, see IBM PowerVM Virtualization Managing
and Monitoring, SG24-7590.
LPM provides systems management flexibility and improves system availability by avoiding
the following situations:
Planned outages for hardware upgrade or firmware maintenance.
Unplanned downtime. With preventive failure management, if a server indicates a potential
failure, you can move its LPARs to another server before the failure occurs.
For more information and requirements for LPM, see IBM PowerVM Live Partition Mobility,
SG24-7460.
HMCV10R1 and VIOS 3.1.3 or later provide the following enhancements to the LPM Feature:
Automatically choose the fastest network for LPM memory transfer.
Allow LPM when a virtual optical device is assigned to a partition.
AME is an innovative technology that supports the AIX operating system. It helps enable the
effective maximum memory capacity to be larger than the true physical memory maximum.
Compression and decompression of memory content can enable memory expansion up to
100% or more. This expansion can enable a partition to complete more work or support more
users with the same physical amount of memory. Similarly, it can enable a server to run more
partitions and do more work for the same physical amount of memory.
The Power E1080 includes a hardware accelerator that is designed to boost AME efficiency and
uses less processor core resources. Each AIX partition can turn on or turn off AME. Control
parameters set the amount of expansion that is wanted in each partition to help control the
amount of CPU used by the AME function.
An IPL is required for the specific partition that is turning memory expansion. When enabled,
monitoring capabilities are available in standard AIX performance tools, such as lparstat,
vmstat, topas, and svmon.
A planning tool is included with AIX, which enables you to sample workloads and estimate how
expandable the partition's memory is and much CPU resource is needed. The feature can be
ordered with the initial order of the Power E1080 or as a Miscellaneous Equipment Specification
(MES) order. A software key is provided when the enablement feature is ordered, which is applied
to the system node. An IPL is not required to enable the system node. The key is specific to an
individual system and is permanent. It cannot be moved to a different server.
HMC V10R1 provides an enhancement to the Remote Restart Feature that enables remote
restart when a virtual optical device is assigned to a partition.
On Power servers, partitions can be configured to run in several modes, including the
following modes:
Power8
This native mode for Power8 processors implements Version 2.07 of the IBM Power
Instruction Set Architecture (ISA). For more information, see IBM Documentation.
Power9
This native mode for Power9 processors implements Version 3.0 of the IBM Power ISA.
For more information, see IBM Documentation.
Power10
This native mode for Power10 processors implements Version 3.1 of the IBM Power ISA.
For more information, see IBM Documentation.
Processor compatibility mode is important when LPM migration is planned between different
generations of servers. An LPAR that might be migrated to a machine that is managed by a
processor from another generation must be activated in a specific compatibility mode.
SR-IOV is a PCI standard architecture that enables PCIe adapters to become self-virtualizing.
It enables adapter consolidation through sharing much like logical partitioning enables server
consolidation. With an adapter capable of SR-IOV, you can assign virtual slices of a single
physical adapter to multiple partitions through logical ports without using a VIOS.
IBM PowerVC can manage AIX, IBM i, and Linux-based VMs that are running under
PowerVM virtualization that is connected to an HMC or by using NovaLink. This release
supports the scale-out and the enterprise Power servers that are built on IBM Power8,
IBM Power9, and subsequent technologies.
IBM PowerVC is an addition to the PowerVM set of enterprise virtualization technologies that
provide virtualization management. It is based on open standards and integrates server
management with storage and network management.
Because IBM PowerVC is based on the OpenStack initiative, Power servers can be managed
by tools that are compatible with OpenStack standards. When a system is controlled by
IBM PowerVC, it can be managed in one of three ways:
By a system administrator by using the IBM PowerVC GUI (GUI)
By a system administrator that uses scripts that contain the IBM PowerVC
Representational State Transfer (REST) application programming interfaces (APIs)
By higher-level tools that call IBM PowerVC by using standard OpenStack API
The following PowerVC offerings are positioned within the available solutions for a Power
cloud:
IBM PowerVC: Advanced Virtualization Management
IBM PowerVC for Private Cloud: Basic Cloud
IBM Cloud Automation Manager: Advanced Cloud
VMware vRealize: Advanced Cloud
The ability to automate by using Ansible returns valuable time to the system administrators.
Red Hat Ansible Automation Platform for Power is fully enabled, so enterprises can automate
several tasks within AIX, IBM i, and Linux that include deploying applications. Ansible can
also be combined with HMC, PowerVC, and Power Virtual Server to provision infrastructure
anywhere you need, including cloud solutions from other IBM Business Partners or third-party
providers based on Power processor-based servers.
A first task after the initial installation or set up of a new LPAR is to help ensure that the
correct patches are installed. Also, extra software (whether it is open source software,
independent software vendor (ISV) software, or perhaps their own enterprise software) must
be installed. Ansible features a set of capabilities to roll out new software, which makes it
popular in Continuous Delivery/Continuous Integration (CD/CI) environments. Orchestration
and integration of automation with security products represent other ways in which Ansible
can be applied within the data center.
AIX and IBM i skilled customers can also benefit from the extreme automation solutions that
are provided by Ansible.
The Power processor-based architecture features unique advantages over commodity server
platforms, such as x86, because the engineering teams that are working on the processor,
system boards, virtualization. and management appliances collaborate closely to help ensure
an integrated stack that works seamlessly. This approach is in stark contrast to the
multi-vendor x86 processor-based technology approach, in which the processor, server,
management, and virtualization must be purchased from different (and sometimes
competing) vendors.
The Power stack engineering teams work closely to deliver the enterprise server platform,
which results in an IT architecture with industry-leading performance, scalability, and security
(see Figure 3-3).
Every layer in the Power stack is optimized to make the Power10 processor-based technology
the platform of choice for mission-critical enterprise workloads. This stack includes the
Ansible Automation Platform, which is described next.
The various Ansible collections for IBM Power processor-based technology, which (at the time
of this writing) were downloaded more than 25,000 times by customers, are now included in
the Red Hat Ansible Automation Platform. As a result, these modules are covered by
Red Hat’s 24x7 enterprise support team, which collaborates with the respective Power
processor-based technology development teams.
From an IBM i perspective, an example is the ability to run SQL queries against the integrated
IBM Db2® database that comes built into the IBM i platform, manage object authorities, and
other tasks. All of these things mean something to an AIX administrator or IBM i administrator.
Our operating system teams develop modules that are sent to the open source community
(named Ansible Galaxy). Every developer can post any object that can be a candidate for a
collection in the open Ansible Galaxy community and possibly certified to be supported by
IBM with a subscription to Red Hat Ansible Automation Platform (see Figure 3-4).
For more information about Ansible in the IBM Power environment, see Using Ansible for
Automation in IBM Power Environments, SG24-8551.
The collection includes modules and sample playbooks that help to automate tasks and is
available starting at this web page.
For more information about the collection, see this web page.
For more information about this collection, see this web page.
The rapidly evolving cyberthreat landscape requires a focus on cyber resilience. Persistent
and end-to-end security is the only way to reduce exposure to threats.
1
Source: Thales Data Threat Report - Global Edition:
[Link]
The Power E1080 is enhanced to simplify and integrate security management across the
stack, which reduces the likelihood of administrator errors.
In the Power E1080, all data is protected by a greatly simplified end-to-end encryption that
extends across the hybrid cloud without detectable performance impact and prepares for
future cyberthreats.
Also, workloads on the Power E1080 benefit from cryptographic algorithm acceleration, which
allows algorithms, such as Advanced Encryption Standard (AES), SHA2, and SHA3 to run
faster than Power9 processor-based servers on a per core basis. This performance
acceleration allows features, such as AIX Logical Volume Encryption, to be enabled with low
performance overhead.
Quantum-safe cryptography refers to the efforts to identify algorithms that are resistant to
attacks by classical and quantum computers in preparation for the time when large-scale
quantum computers are built.
PowerSC is introducing more features to help customers manage security end-to-end across
the stack to stay ahead of various threats. Specifically, PowerSC 2.0 adds support for
Endpoint Detection and Response (EDR), host-based intrusion detection, block listing, and
Linux.
As the world is set to deploy AI everywhere, attention is turning from how fast to build AI to
how fast to inference with the AI.
To support this shift, the Power E1080 delivers faster business insights by running AI in place
with four Matrix Math Accelerator (MMA) units to accelerate AI in each Power10
technology-based processor-core. The robust execution capability of the processor cores with
MMA AI acceleration, enhanced single instructions multiple data (SIMD), and enhanced data
bandwidth provide an alternative to external accelerators, such as GPUs, and related device
management for execution of statistical machine learning and inferencing workloads. These
features, combined with the possibility to consolidate multiple environments for AI model
execution on a Power E1080 platform together with other different types of environments,
reduces costs and leads to a greatly simplified solution stack for AI.
Operationalizing AI inferencing directly on a Power E1080 brings AI closer to data. This ability
allows AI to inherit and benefit from the Enterprise Qualities of Service (QoS): reliability,
availability, and security of the Power10 processor-based platform and support a performance
boost. Enterprise business workflows can now readily and consistently use insights that are
built with the support of AI.
The use of data gravity on Power10 processor-cores enables AI to run during a database
operation or concurrently with an application, for example. This feature is key for
time-sensitive use cases. It delivers fresh input data to AI faster and enhances the quality and
speed of insight.
Open Neural Network Exchange (ONNX) models can be brought over from x86
processor-based servers or other platforms and small-sized VMs or Power Virtual Server
(PowerVS), for deployment on Power E1080, which gives customers the ability to build on
commodity hardware but deploy on enterprise servers.
IBM optimized math libraries so that AI tools benefit from acceleration that is provided by the
MMA units of the Power10 chip. The benefits of MMA action can be realized for statistical
machine learning and inferencing, which provides a cost-effective alternative to external
accelerators or GPUs.
Power10 cores are equipped with four Math Engines for matrix and tensor math. Applications
can run models with colocated data without the need for external accelerators, GPUs, or extra
AI platforms. Power10 technology uses the “train anywhere, deploy here” principle to
operationalize AI.
A model can be trained on a public or private cloud (see Figure 3-6) by using the following
procedure:
1. The model is registered with its version in the so-called model vault, which is a VM or
LPAR with tools, such as Watson OpenScale, BentoML, or Tensorflow Serving.
2. The model is pushed out to the destination (in this case, it is a VM or an LPAR running a
database with an application) and the model might be used by the database or the
application.
3. Transactions that are received by the database and application trigger model execution
and generate predictions or classifications. These predictions can also be stored locally.
For example, these predictions can be the risk or fraud that is associated with the
transaction or product classifications to be used by downstream applications.
4. A copy of the model output (prediction or classification) is sent to the model operations
(ModelOps) engine for calculation of drift by comparison with Ground Truth.
5. If the drift exceeds a threshold, the model retrain triggers are generated.
6. Retrained models are then taken through steps 1 - 5.
By using a combination of software and hardware innovation, Power E1080 can meet the
model performance, response time, and throughput KPIs of databases and applications that
are infused with AI.
The publications that are listed in this section are considered suitable for a more detailed
description of the topics that are covered in this paper.
IBM Redbooks
The following IBM Redbooks publications provide more information about the topics in this
document. Some publications that are referenced in this list might be available in softcopy
only.
IBM PowerAI: Deep Learning Unleashed on IBM Power Systems Servers, SG24-8409
IBM Power System AC922 Technical Overview and Introduction, REDP-5494
IBM Power System E950: Technical Overview and Introduction, REDP-5509
IBM Power System E980: Technical Overview and Introduction, REDP-5510
IBM Power System L922 Technical Overview and Introduction, REDP-5496
IBM Power Systems H922 and H924 Technical Overview and Introduction, REDP-5498
IBM Power Systems LC921 and LC922: Technical Overview and Introduction,
REDP-5495
IBM Power Systems Private Cloud with Shared Utility Capacity: Featuring Power
Enterprise Pools 2.0, SG24-8476
IBM Power System S822LC for High Performance Computing Introduction and Technical
Overview, REDP-5405
IBM Power Systems S922, S914, and S924 Technical Overview and Introduction
Featuring PCIe Gen 4 Technology, REDP-5595
IBM PowerVM Best Practices, SG24-8062
IBM PowerVC Version 2.0 Introduction and Configuration, SG24-8477
IBM PowerVM Virtualization Introduction and Configuration, SG24-7940
IBM PowerVM Virtualization Managing and Monitoring, SG24-7590
SAP HANA Data Management and Performance on IBM Power Systems, REDP-5570
Using Ansible for Automation in IBM Power Environments, SG24-8551
You can search for, view, download, or order these documents and other Redbooks
publications, Redpapers, web docs, drafts, and additional materials, at the following website:
[Link]/redbooks
REDP-5649-01
ISBN 073846189X
Printed in U.S.A.
®
[Link]/redbooks