7/22/2019 010 VIO Overview
1/40
Copyright IBM Corporation 2006 Course materials may not be reproduced in whole or in part without written permission.3.3
Virtual I/O Server Overview
QPS01 IBM System p APV - 010
7/22/2019 010 VIO Overview
2/40
7/22/2019 010 VIO Overview
3/40
7/22/2019 010 VIO Overview
4/40
Copyright IBM Corporation 2006
IBM System p
QPS01 010 VIO Overview 4
The IBM benchmarks results shown herein were derived using particular, well configured, development-level and generally-available computer systems. Buyersshould consult other sources of information to evaluate the performance of systems they are considering buying and should consider conducting applicationoriented testing. For additional information about the benchmarks, values and systems tested, contact your local IBM office or IBM authorized reseller or accessthe website of the benchmark consortium or benchmark vendor.
IBM benchmark results can be found in the IBM eServer p5, pSeries and IBM RS/6000 Performance Report athttp://www-1.ibm.com/servers/eserver/pseries/hardware/system_perf.html
Unless otherwise indicated for a system, the performance benchmarks were conducted using AIX V4.3 or AIX 5L. IBM C Set++ for AIX and IBM XLFORTRAN for AIX with optimization were the compilers used in the benchmark tests. The preprocessors used in some benchmark tests include KAP 3.2 forFORTRAN and KAP/C 1.4.2 from Kuck & Associates and VAST-2 v4.01X8 from Pacific-Sierra Research. The preprocessors were purchased separately fromthese vendors. Other software packages like IBM ESSL for AIX and MASS for AIX were also used in some benchmarks.
For a definition and explanation of each benchmark and the full list of detailed results, visit the web site of the benchmark consortium or benchmark vendor.
TPC http://www.tpc.org
SPEC http://www.spec.org
Linpack http://www.netlib.org/benchmark/performance.pdf
Pro/E http://www.proe.comGPC http://www.spec.org/gpc
NotesBench http://www.notesbench.org
VolanoMark http://www.volano.com
STREAM http://www.cs.virginia.edu/stream/
SAP http://www.sap.com/benchmark/
Oracle Applications http://www.oracle.com/apps_benchmark/
PeopleSoft - To get information on PeopleSoft benchmarks, contact PeopleSoft directly
Siebel http://www.siebel.com/crm/performance_benchmark/index.shtm
Baan http://www.ssaglobal.com
Microsoft Exchange http://www.microsoft.com/exchange/evaluation/performance/default.asp
Veritest http://www.veritest.com/clients/reports
Fluent http://www.fluent.com/software/fluent/fl5bench/fullres.htm
TOP500 Supercomputers http://www.top500.org/
Ideas International http://www.microsoft.com/exchange/evaluation/performance/default.asp
Storage Performance Council http://www.storageperformance.org/results
Notes on Benchmarks and Values
Revised August 26, 2003
7/22/2019 010 VIO Overview
5/40
Copyright IBM Corporation 2006
IBM System p
QPS01 010 VIO Overview 5
rPerf
rPerf (Relative Performance) is an estimate of commercial processing performance relative to other pSeries systems. It is derived from an IBM
analytical model which uses characteristics from IBM internal workloads, TPC and SPEC benchmarks. The rPerf model is not intended to represent any
specific public benchmark results and should not be reasonably used in that way. The model simulates some of the system operations such as CPU,
cache and memory. However, the model does not simulate disk or network I/O operations.
rPerf estimates are calculated based on systems with the latest levels of AIX 5L and other pertinent software at the time of system announcement.Actual performance will vary based on application and configuration specifics. The IBM pSeries 640 is the baseline reference system and has
a value of 1.0. Although rPerf may be used to approximate relative IBM UNIX commercial processing performance, actual system performance may
vary and is dependent upon many factors including system hardware configuration and software design and configuration.
All performance estimates are provided "AS IS" and no warranties or guarantees are expressed or implied by IBM. Buyers should consult other sources
of information, including system benchmarks, and application sizing guides to evaluate the performance of a system they are considering buying. For
additional information about rPerf, contact your local IBM office or IBM authorized reseller.
Notes on Performance Estimates
Revised June 28, 2004
7/22/2019 010 VIO Overview
6/40
Copyright IBM Corporation 2006
IBM System p
QPS01 010 VIO Overview 6
What is Advanced POWER Virtualization (APV)
APV the hardware feature code for POWER5 servers that enables:
Micro-partitioning fractional CPU entitlements from a shared pool of
processors, beginning at one-tenth of a CPU
Partition Load Manager (PLM) a policy-based, dynamic CPU and memory
reallocation tool
VIO Server (virtual SCSI and Shared Ethernet Adapter)
Physical disks can be shared as virtual disks to client partitions
Shared Ethernet Adapter (SEA) A physical adapter or EtherChannel in
a VIO Server can be shared by client partitions. Clients use virtual
Ethernet adapters
Virtual Ethernet a LPAR-to-LPAR Virtual LAN within a POWER5 Server
Does not require the APV feature code
7/22/2019 010 VIO Overview
7/40
Copyright IBM Corporation 2006
IBM System p
QPS01 010 VIO Overview 7
Why Virtual I/O Server?
POWER5 systems will support more partitions than physical I/O slots available
Each partition still requires a boot disk and network connection, but now
they can be virtual instead of physical
VIO Server allows partitions to share disk and network adapter resources
The Fibre Channel or SCSI controllers in the VIO Server can be accessed
using Virtual SCSI controllers in the clients
A Shared Ethernet Adapter in the VIO Server can be a layer 2 bridge for
virtual Ethernet adapters in the clients
The VIO Server further enables on demand computing and server consolidation
7/22/2019 010 VIO Overview
8/40
Copyright IBM Corporation 2006
IBM System p
QPS01 010 VIO Overview 8
Why Virtual I/O Server?Virtualizing IO saves:
Gbit Ethernet Adapters
2 Gbit Fibre Channel Adapters
PCI slots
Eventually, IO drawers
Server frames?
Floor space?
Electric, HVAC?
Ethernet switch ports
Fibre channel switch ports
Logistics, scheduling, delays of physical Ethernet, SAN attach
Some servers run 90% utilization all the time everyone knows which ones.
However,
Average utilization in the UNIX server farm is closer to 25%. They dont all
maximize their use of dedicated IO devices
VIO is departure from new project, new chassis mindset
7/22/2019 010 VIO Overview
9/40
Copyright IBM Corporation 2006
IBM System p
QPS01 010 VIO Overview 9
Virtual I/O Server Characteristics
Requires AIX 5.3 and POWER5 hardware with APV feature
Installed as a special purpose, AIX-based logical partition
Uses a subset of the AIX Logical Volume Manager and attaches to
traditional storage subsystems
Inter-partition communication (client-server model) provided via the POWER
Hypervisor
Clients see virtual disks as traditional AIX SCSI hdisks, although they may be
a physical disk or logical volume on the VIO Server
One physical disk on a VIO server can provide logical volumes for several
client partitions
7/22/2019 010 VIO Overview
10/40
Copyright IBM Corporation 2006
IBM System p
QPS01 010 VIO Overview 10
Creating the Virtual IO Server
7/22/2019 010 VIO Overview
11/40
Copyright IBM Corporation 2006
IBM System p
QPS01 010 VIO Overview 11
Virtual I/O Server installation
VIO Server code is packaged and shipped as an AIX mksysb image on a VIO CD
Installation methods
CD install
HMC install - Open rshterm and type installios; follow the prompts
The Network Installation Manager (NIM)
VIO Server can support multiple client types
AIX 5.3
SUSE Linux Enterprise Server 9 for POWER
Red Hat Enterprise Linux AS for POWER Version 3 and 4
7/22/2019 010 VIO Overview
12/40
Copyright IBM Corporation 2006
IBM System p
QPS01 010 VIO Overview 12
Virtual I/O Server Administration
The VIO server uses a command line interface running in a restricted shell no
smitty or GUI
There is no root login on the VIO Server
A special user padmin executes VIO server commands
First login after install, user padmin is prompted to change password
After that, padmin runs the command license accept
Slightly modified commands are used for managing devices, networks, code
installation and maintenance, etc.
The padmin user can start a root AIX shell for setting up third-party devices
using the command oem_setup_env
7/22/2019 010 VIO Overview
13/40
Copyright IBM Corporation 2006
IBM System p
QPS01 010 VIO Overview 13
VIO Server/Client Overview
7/22/2019 010 VIO Overview
14/40
Copyright IBM Corporation 2006
IBM System p
QPS01 010 VIO Overview 14
VIO Server Configuration with MPIO
MPIO in client LPAR
automatically configures
Client sees one hdisk with two
MPIO paths lspath l hdisk0
Paths are fail_over only. No
load balancing in client MPIO
hdisk1 in each VIO server
attached to vscsi server
adapter as a raw disk
Set reserve_policy attribute
on hdisk1 to no_reserve ineach VIO server. Also, with two fcs in
a VIO server, change algorithm to
round_robin for hdisk1. SDDPCM
driver installed in each VIO server
LUN appears in each VIOserver as hdisk1
Single RAID5 LUN carved in ESS,
zoned to 4 FC adapters in the two VIO
servers
SDDPCM SDDPCM
7/22/2019 010 VIO Overview
15/40
Copyright IBM Corporation 2006
IBM System p
QPS01 010 VIO Overview 15
Checking for MPIO Capability
Run oem_setup_env to get to a root prompt
Run # odmget qattribute=unique_id CuAt
In the output, look for a field unique_id followed by a value field
lsattr rl hdisk0 a algorithm (both round_robin or fail_over)
With VIOS 1.3, MPIO is available with:
- IBM Enterprise Storage Server and DS6000 with SDDPCM (but notSDD)
- FAStT and DS4000 with RDAC- IBM System Storage N3700 and N5200
- EMC PowerPath
- HDS with HDLM
7/22/2019 010 VIO Overview
16/40
Copyright IBM Corporation 2006
IBM System p
QPS01 010 VIO Overview 16
VIO Server Configuration with LVM Mirroring
7/22/2019 010 VIO Overview
17/40
Copyright IBM Corporation 2006
IBM System p
QPS01 010 VIO Overview 17
Internals AIX
Virtual I/O Server AIX client
LVM
OEM
multipathing
disk DD
HBA DD
VSCSI
client
disk DD
LVM
MPIOPCM
MPIO
PCM
VSCSI
serverinternal orexternal storage
logical volumes
OEM device paths
physical volumes
7/22/2019 010 VIO Overview
18/40
Copyright IBM Corporation 2006
IBM System p
QPS01 010 VIO Overview 18
Virtual I/O Server Resource Configuration
Add physical devices to the VIO Server
If serving LVs as hdisks, create a volume group on one or more disks with
mkvg
- mkvg [-f] [-vg VolumeGroup] PhysicalVolume
- mkvg f vg rootvg_clients hdisk2
rootvg_clients
Create logical volumes on the volume group
- mklv [-lv NewLogicalVolume | -prefix Prefix ]
VolumeGroup Size [PhysicalVolume ]- mklv lv aix_sq07 rootvg_clients 7G hdisk2
7/22/2019 010 VIO Overview
19/40
19 IBM Systems GroupPOWER5 VIO Server 2006 IBM Corporation
DLPAR add of VIO Server Virtual Adapter
List virtual adapter after added by DLPAR
lsdev virtualname status description
ent2 Available Virtual I/O Ethernet Adapter (l-lan)vhost0 Available Virtual SCSI Server Adaptervhost1 Available Virtual SCSI Server Adaptervsa0 Available LPAR Virtual Serial Adapter
Doesnt show up in VIO Server until after running cfgdev
cfgdev dev vio0
List virtual adapter after running cfgdev
lsdev virtualname status description
ent2 Available Virtual I/O Ethernet Adapter (l-lan)vhost0 Available Virtual SCSI Server Adaptervhost1 Available Virtual SCSI Server Adaptervhost2 Available Virtual SCSI Server Adapter vsa0 Available LPAR Virtual Serial Adapter
7/22/2019 010 VIO Overview
20/40
Copyright IBM Corporation 2006
IBM System p
QPS01 010 VIO Overview 20
Virtual IO Server Resource Mapping
Configuring virtual target device
mkvdev vdev aix_sq07 vadapter vhost0 dev vt_aix_sq07
mkvdev vdev hdisk7 vadapter vhost1 dev vt_hdisk7
$ lsdev -virtual
name status description
ent2 Available Virtual I/O Ethernet Adapter (l-lan)
vhost0 Available Virtual SCSI Server Adapter
vhost1 Available Virtual SCSI Server Adapter
vsa0 Available LPAR Virtual Serial Adapter
vt_aix_sq07 Available Virtual Target Device - Logical Volumevt_hdisk7 Available Virtual Target Device - Disk
7/22/2019 010 VIO Overview
21/40
Copyright IBM Corporation 2006
IBM System p
QPS01 010 VIO Overview 21
Confirm what backs up each vhostList all vscsi server adapters with all flag
$ lsmap -all
SVSA Physloc Client Partition ID
--------------- ---------------------------------- ------------------
vhost0 U9111.520.10C1C1C-V3-C2 0x00000001
VTD vtscsi0
LUN 0x8100000000000000
Backing device aix_sq07
Physloc
SVSA Physloc Client Partition ID
--------------- ---------------------------------- ------------------
vhost1 U9111.520.10C1C1C-V3-C4 0x00000001
VTD vtscsi1
LUN 0x8100000000000000
Backing device hdisk7
Physloc U787A.001.DNZ00ZE-P1-C1-T1-L4-L0
List a single vscsi server adapter with vadapter flag
$ lsmap vadapter vhost0
slot #
Logical Volume
hdisk
7/22/2019 010 VIO Overview
22/40
Copyright IBM Corporation 2006
IBM System p
QPS01 010 VIO Overview 22
More Use of lsmap
$for v in `ioscli lsdev -virtual | grep vhost | awk '{print$1}'`
do
ioscli lsmap -vadapter $v -fmt : | awk -F: '{ print $1" "$2""$4" "$6 $8 $10}'
done
vhost0 U9117.570.10C0EDC-V16-C7 virt_aix_sq07 aix_sq07
vhost1 U9117.570.10C0EDC-V16-C8 virt_suse_sq07 suse_sq07
vhost2 U9117.570.10C0EDC-V16-C9 virt_aix_sq08 aix_sq08
vhost3 U9117.570.10C0EDC-V16-C10 virt_suse_sq08 suse_sq08
vhost4 U9117.570.10C0EDC-V16-C11 virt_aix_sq09 aix_sq09
vhost5 U9117.570.10C0EDC-V16-C12 virt_suse_sq09 suse_sq09
vhost6 U9117.570.10C0EDC-V16-C13 virt_aix_sq10 aix_sq10
vhost7 U9117.570.10C0EDC-V16-C14 virt_suse_sq10 suse_sq10
vhost8 U9117.570.10C0EDC-V16-C15 virt_rhel_sq07 rhel_sq07
vhost9 U9117.570.10C0EDC-V16-C16 virt_rhel_sq08 rhel_sq08
vhost10 U9117.570.10C0EDC-V16-C17 virt_rhel_sq09 rhel_sq09
vhost11 U9117.570.10C0EDC-V16-C18 virt_rhel_sq10 rhel_sq10
7/22/2019 010 VIO Overview
23/40
Copyright IBM Corporation 2006
IBM System p
QPS01 010 VIO Overview 23
Client Virtual Disk Attributes
# lsdev -Cc disk
hdisk0 Available Virtual SCSI Disk Drive
# lscfg -vl hdisk0
hdisk0 U9117.570.10C0EDC-V7-C5-T1-L810000000000Virtual SCSI Disk Drive
# lsattr -El hdisk0
PCM PCM/friend/vscsiPath Control Module False
algorithm fail_over
Algorithm Falsemax_transfer 0x20000Maximum TRANSFER Size True
pvid 00cc0edc916c5bd80000000000000000Physical volume identifier False
queue_depth 3Queue DEPTH False
reserve_policy no_reserveReserve Policy False
# lscfg -vl vscsi1
vscsi1 U9117.570.10C0EDC-V7-C6-T1 Virtual SCSI ClientAdapter
Device Specific.(YL)........U9117.570.10C0EDC-V7-C6-T1
7/22/2019 010 VIO Overview
24/40
Copyright IBM Corporation 2006
IBM System p
QPS01 010 VIO Overview 24
Virtual Ethernet
Virtual Ethernet
Enable inter-lpar communications without a physical adapter
IEEE-compliant Ethernet programming model
Implemented through inter-partition, in-memory communication
VLAN splits up groups of network users on a physical network onto
segments of logical networks
Virtual switch provides support for multiple (up to 4K) VLANs
Each partition can connect to multiple networks, through one or more
adapters
VIO server can add VLAN ID tag to the Ethernet frame as appropriate.
Ethernet switch restricts frames to ports that are authorized to receive frames
with specific VLAN ID
Virtual network can connect to physical network through routing"
partitions generally not recommended
7/22/2019 010 VIO Overview
25/40
Copyright IBM Corporation 2006
IBM System p
QPS01 010 VIO Overview 25
SEA Configuration
$ lsdev | grep ent[0-9]ent0 Available Virtual I/O Ethernet Adapter (l-lan)
ent1 Available 10/100 Mbps Ethernet PCI Adapter II (1410ff01)
$ mkvdev sea ent1 vadapter ent0 default ent0 defaultid 1ent2 Available
$ lsdev | grep ent[0-9]ent0 Available Virtual I/O Ethernet Adapter (l-lan)
ent1 Available 10/100 Mbps Ethernet PCI Adapter II (1410ff01)
ent2 Available Shared Ethernet Adapter
$ lsdev dev ent2 -attrpvid 1 PVID to use for the SEA device
pvid_adapter ent0 Default virtual adapter to use for non-VLAN-tagged packetsreal_adapter ent1 Physical adapter associated with the SEA
thread 0 Thread mode enabled (1) or disabled (0)
virt_adapters ent0 List of virtual adapters associated with the SEA
7/22/2019 010 VIO Overview
26/40
Copyright IBM Corporation 2006
IBM System p
QPS01 010 VIO Overview 26
Shared Ethernet Adapter setup
VIO Server after mksysb InstallPhysicalEthernet
Virtual
Ethernet
VIO Server after sea config
mkvdev sea ent1 vadapter ent0 default ent0 defaultid 1ent2 Available
VIO servers IP address may beput on shared adapter interface,or on separate, non-external
access virtual adapter. VIO IPaddress is NOT configured onphysical, nor on SEA virtual adapter.
ent1 ent0
9.19.126.98
Shared Ethernet ent2
ent0ent1
7/22/2019 010 VIO Overview
27/40
Copyright IBM Corporation 2006
IBM System p
QPS01 010 VIO Overview 27
SEA with NIB in VIO client Partitions
At Power5 GA, only way to provide network redundancy
VIO servers each share adapter, each on different internal network (PortVLAN ID)
Client builds NIB EtherChannel across the Port VLAN IDs no support for VLAN tagging both VIOs, all clients same IP subnet
7/22/2019 010 VIO Overview
28/40
Copyright IBM Corporation 2006
IBM System p
QPS01 010 VIO Overview 28
What is SEA Failover? (AIX 5.3 ML3, VIO 1.2.0.0)
Shared Ethernet Adapter Redundancy
- SEA adapters in two VIO servers can back each other up.
- Two VIO servers now allowed to put external accessvirtual adapters on same internal PVID, with differenttrunk priorities
- Administrator dictates which SEA will be the primary basedon an assigned Trunk Priority
- A standby mode is available to force the SEA failover tobackup server, allowing for regular updates or maintenance
- The backup server that became primary will switch to beinga backup when primary resumes operation
7/22/2019 010 VIO Overview
29/40
Copyright IBM Corporation 2006
IBM System p
QPS01 010 VIO Overview 29
What is SEA Failover (AIX 5.3 ML3, VIO 1.2.0.0)
Current limitation is one backup SEA per primary SEA atthis time
The backup SEA is idle until a failure occurs on theprimary SEA
- This is transparent to client LPARs
Failure may be due to:
- Physical adapter failure
- VIOS crash/reboot
- Failure to reach a pre-defined remote address
7/22/2019 010 VIO Overview
30/40
Copyright IBM Corporation 2006
IBM System p
QPS01 010 VIO Overview 30
SEA Failover
Inside P5, we now allowed more than one trunk adapter with the sameVLAN ID; These adapters get different Trunk Priority between 1-15
- A lower priority is the more favored SEA; primary
- If there is more than one virtual Ethernet on the same SEA,each must have the same priority
The SEA failover protocol helps determine which one should behave asprimary, and when a failover should occur
7/22/2019 010 VIO Overview
31/40
Copyright IBM Corporation 2006
IBM System p
QPS01 010 VIO Overview 31
7/22/2019 010 VIO Overview
32/40
Copyright IBM Corporation 2006
IBM System p
QPS01 010 VIO Overview 32
Shared Ethernet Adapter Failover
CLI command to configure SEA failover with a remotemkvdev sea ent0 vadapter ent1 default ent1 defaultid 10 \
attr ha_mode=auto ctl_chan=ent2
IBM S t G
7/22/2019 010 VIO Overview
33/40
33
IBM Systems Group
POWER5 VIO Server Overview - QPS01 010 2006 IBM Corporation
Shared Ethernet Adapter Failover
Recap of these adapters:ent0 physical adapter to be shared it could be an EtherChannel over several physicalsent1 VIO Server 1 virtual, PVID 10, Access External Checkbox, Trunk Priority lowent1 VIO Server 2 virtual, PVID 10, Access External Checkbox, Trunk Priority highent2 virtual, PVID 10, no Access External Checkbox, no Trunk Priority, VIO server IP addrent3 virtual, PVID 99, Control Channelent4 Shared adapter, created over ent0 and ent1
ent3
ent1
en
t4
ent2
vio
ip
addr
en
t4
ent0
ent0
ent1
ent2
vio
ip
addr
ent3
Enterprise network
Power5 internal PVID 10
Power5 internal PVID 99
VIO
Server1VIO
Server2
Client LPAR
ent0
Lower trunkpriority Higher trunk
priority
IBM S t G
7/22/2019 010 VIO Overview
34/40
34
IBM Systems Group
POWER5 VIO Server Overview - QPS01 010 2006 IBM Corporation
Shared Ethernet Adapter Failover
VIO command line to configure SEA failover in each VIO LPAR
mkvdev sea ent0 vadapter ent1 default ent1 defaultid 10 attrha_mode=auto ctl_chan=ent3
Some examples show setting ha_mode and ctl_chan with a separate chdevcommand. Dont.
ent3
ent1
en
t4
ent2
vio
ip
addr
en
t4
ent0
ent0
ent1
ent2
vio
ip
addr
ent3
Enterprise network
Power5 internal PVID 10
Power5 internal PVID 99
VIO
Server1VIO
Server2
Client LPAR
ent0
Lower trunkpriority-favored Higher trunk
priority
IBM Systems Group
7/22/2019 010 VIO Overview
35/40
35
IBM Systems Group
POWER5 VIO Server Overview - QPS01 010 2006 IBM Corporation
Shared Ethernet Adapter Failover
You may put VIO IP addr on en4 (supported). We find stability putting it on anothervirtual adapter en2. ent2 adapter is a virtual adapter on Port VLAN id 10, withoutAccess External Network checkbox
Remove and reconfig of SEA can be done without taking down VIO server ip, default router, etcNIM restore of either VIO thru SEA on the other VIO is more seamless for example,
nim install VIO Server1 into its ent2 adapter, with RECOVER_DEVICES=yes, thru SEA ofVIO Server2
ent3
ent1
en
t4
ent2
vio
ip
addr
en
t4
ent0
ent0
ent1
ent2
vio
ip
addr
ent3
Enterprise network
Power5 internal PVID 10
Power5 internal PVID 99
VIO
Server1VIO
Server2
Client LPAR
ent0
Lower trunkpriority Higher trunk
priority
7/22/2019 010 VIO Overview
36/40
Copyright IBM Corporation 2006
IBM System p
QPS01 010 VIO Overview 36
SEA Failover Pros and Cons
Unexpected outages such as systems crashes are uncommon but may occur
- A backup server detecting that the primary server is down will make itself the primary andwill begin bridging traffic.
No EtherChannel requirement at VIO Client partition- Single virtual adapter in client partition, nim install thru it, keep the configured IP
interface after nim install and boot
- Client does not need rmtcpip ; EtherChannel config ; mktcpip after nim install
Allows use of Enterprise Network VLANs inside Power5 Server
SEA failover can support planned VIO maintenance
- Scheduled maintenance can now be performed with no loss of user connections
- Traffic is transparently sent over another SEApadmin chdev dev -attr ha_mode=standby
Cons?
- More adapters in VIO server a bit more complex to configure
- All client partitions drive IO thru the more favored VIO SEA, until failed over to lessfavored VIO no balancing of client LPARs between VIO servers. Perhaps EtherChannelof physical adapters in VIO is important here.
7/22/2019 010 VIO Overview
37/40
37 IBM Systems GroupPOWER5 VIO Server 2006 IBM Corporation
SEA Failover vs Network Interface Backup
Manual intervention required to force
affected (1/2) client partitions back to
primary VIO Server
Automatic fail-back to primary
VIO Server when network link
becomes available
Failback to Primary
Not SupportedSupported802.1Q VLAN Tagging
No idle connections as client partitions
can be split between the two VIO
Servers
Requires a backup idle
connection on opposite VIO
Server
Physical network connections
(historically limited to 1500)
> 1500 requires AIX V5.3 TL4*
>= 1500MTU Size
Configuration required on all of the
client partitions
Configuration required only on the
VIO Servers
Configuration
VIO Server V1.1VIO Server V1.2
GA6 Firmware/HMC
Software Pre-requisites
Network Interface BackupSEA Failover
Note: IY79375 (AIX V5.3 TL4) provides Jumbo Frame Support for EtherChannel with virtual adapters
7/22/2019 010 VIO Overview
38/40
38 IBM Systems GroupPOWER5 VIO Server 2006 IBM Corporation
Integrated Virtualization Manager (IVM) providesHMC functionality in a partitionat a much lower price point
AIX 5L
V5.3
Linux
SLES 9
Linux
RHEL 4
Virtual LAN
VIOS
IVM
Provides LPAR/virtualization support
without a physical HMC no HMC
hardware translates to lower entry point
Web-based, intuitive/user-friendly interface
Shipped with the Virtual I/O Server (VIOS)
provides tight integration of LPAR
management and virtual storage
configuration
Supports creating/management of I/O and
LPARs within a single physical server
All I/O is virtualized Virtual Console,
Storage, Ethernet, and Optical
Subset of HMC Service functionality
To be available for entry servers up to
p5-550Q (both System p5 and p5
families)
HMC within a partition
POWER Hypervisor
Integrated Virtualization Manager
Integrated Virtualization Mgr
7/22/2019 010 VIO Overview
39/40
Copyright IBM Corporation 2006
IBM System p
QPS01 010 VIO Overview 39
Reference
InfoCenter http://publib.boulder.ibm.com/infocenter/eserver/v1r2s/en_US/index.htm
Virtualizing your compute environment
http://publib.boulder.ibm.com/infocenter/eserver/v1r2s/en_US/info/iphb1/iphb2.pdf
VIO Server and PLM command line reference http://publib.boulder.ibm.com/infocenter/eserver/v1r2s/en_US/info/iphb1/commands/commands.pdf
Redbook
Advanced POWER Virtualization on IBM System p5http://www.redbooks.ibm.com search on SG24-7940
http://www.redbooks.ibm.com/redpieces/pdfs/sg247940.pdf
VIO Server Page from Fix Central Overview, Download, Documentation
http://www14.software.ibm.com/webapp/set2/sas/f/vios/home.html
7/22/2019 010 VIO Overview
40/40
Copyright IBM Corporation 2006
IBM System p
QPS01 010 VIO Overview 40
More references
http://techsupport.services.ibm.com/server/vios/documentation/configuring_mpio_for_the_virtual_client.pdf
http://www14.software.ibm.com/webapp/set2/sas/f/vios/documentation/faq.html
http://www14.software.ibm.com/webapp/set2/sas/f/vios/home.html
http://www-128.ibm.com/developerworks/eserver/library/es-pwr5-virtualvlan/?ca=dnp-439
http://www14.software.ibm.com/webapp/set2/sas/f/vios/documentation/home.html
Top Related