OPENVMS I/O AND STORAGE

download OPENVMS I/O AND STORAGE

of 44

Transcript of OPENVMS I/O AND STORAGE

  • 7/30/2019 OPENVMS I/O AND STORAGE

    1/44

    2010 Hewlett-Packard Development Company, L.P.The information contained herein is subject to change without notice

    2011 Hewlett-Packard Development Company, L.P.The information contained herein is subject to change without notice

    OPENVMS I/O AND

    STORAGETips and Best Practices for good performance

    Rafiq Ahamed K

    OpenVMS Engineering

  • 7/30/2019 OPENVMS I/O AND STORAGE

    2/44

    Agenda

    2

    OpenVMS I/O Facts I/O Evolution on Integrity servers

    What to expect from hardware

    What do you know about multipathing

    Notes on IOPERFORM

    NUMA and I/O

    Storage Tips and Tricks EVA Best Practices

    OpenVMS Connectivity

    Have a sneak into EVA PA

    Q & A

    9/19/2011

  • 7/30/2019 OPENVMS I/O AND STORAGE

    3/44

    Operating System performance is largely dependent

    on the underlying hardware

    ..So know your hardware capabilities

    9/19/20113

  • 7/30/2019 OPENVMS I/O AND STORAGE

    4/44

    Integrity server I/O

    4

    Architecture has evolved drastically for I/O Devices within Integrity,Performance & Scalability - doubling in each new hardware release

    9/19/2011

    PCIInterconnects

    PCI (66MHz,0.5GB/sec)

    PCI-X(133MHz,1GB/sec)PCI-X(266MHz, 2GB/sec)

    PCIe Gen 1(2.5Gb/sec/Lane,250MB/sec)

    PCIe Gen 2(5Gb/sec/Lane,500MB/sec)

    Core I/O Ultra SCSI 160 Ultra SCSI 320 3G SAS (LSI Logic) 6G SAS (p410i)

    SAN I/O 1G FC 2G FC 4G FC 8G FC

    Disk Size 300GB 600GB 1.2TB 7.2TB

    # I/O Device 3 6 8 12/16

  • 7/30/2019 OPENVMS I/O AND STORAGE

    5/44

    Examples of latest speeds andfeeds of I/O on Integrity

    platforms

    5

  • 7/30/2019 OPENVMS I/O AND STORAGE

    6/44

    Each BL890c/rx2800 supports eight SFF SAS HDD, up to 7.2TB capacity

    Leadership in I/O and Storage on i2 architecture

    High performance, reliable and scalable

    9/19/20117

    SAS provides apoint-to-point

    connection to each

    HDD @6G Speeds

    Parallel SCSI with rx7640has a shared busUltra160 SCSI

    Provides four p410RAID controllers

    (one per blade) onBL890c i2,

    One 410 RAID onrx2800

    Configured as RAID 0/1or HBA mode [Future]

    Stripe data withinand across multiple

    p410 RAIDcontrollers

    (OpenVMSShadowing)

    Striping within provideshigh Performance.

    Striping across controllersprovides no SPOF storage

  • 7/30/2019 OPENVMS I/O AND STORAGE

    7/44

    Data shows the impact of p410i Caching and Striping

    Core I/O on i2 servers

    8

    0

    500

    1000

    1500

    2000

    2500

    1 2 4 8 16 32 64 128 256

    IOPS

    Load

    rx2800 i2 - SAS Logical Disk (Striping)

    1 disk with Cache 2 disk with Cache 4 disk with Cache

    0

    100

    200

    300

    400

    500

    600

    700

    1 2 4 8 16 32 64 128 256

    IOPS

    Load

    rx2800 i2 - Core SAS Caching

    W/O Cache Cache

    9/19/2011

    Use p410i Cache Batter Kit for faster response Stripe across multiple disks to maximize the utilization and

    throughput

  • 7/30/2019 OPENVMS I/O AND STORAGE

    8/44

    How is I/O Performance on Integrity Servers?

    How they compare against my existing high end alpha servers?

    After migrating to new platform what should I expect?

    Why is i2 server I/O market differentiator?

    Customer Concerns..

    9/19/20119

  • 7/30/2019 OPENVMS I/O AND STORAGE

    9/44

    Software capabilities;

    multipathing

    10

  • 7/30/2019 OPENVMS I/O AND STORAGE

    10/44

    Multipathing 1(4)

    11

    Multipathing (MP) is a technique to manage multiple paths to storage

    device through failover and failback mechanisms

    It helps user to load balance across multiple paths to storage

    By default multipath will be enabled on OpenVMS

    OpenVMS MP supports ALUA (Asymmetric Logical Unit Access) [> V8.3]

    OpenVMS MP supports FAILOVER and FAILBACK

    OpenVMS MP load balances the storage devices across all the pathsavailable It spreads # of devices evenly across all the paths during boot time

    At any point in time only single path can be active to a device

    Users are recommended to use static load balancing techniques

    9/19/2011

  • 7/30/2019 OPENVMS I/O AND STORAGE

    11/44

    MP Connections Good and Bad

    12

    Alpha

    HSVHSV

    switch switch

    IA64 Alpha

    HSVHSV

    switch switch

    IA64

    IA64

    HSV

    switch

    IA64

    HSV

    switch switch

    IA64

    HSV

    switch

    Single Two Paths Four Paths

    Shows single controllerconfigurations

    9/19/2011

  • 7/30/2019 OPENVMS I/O AND STORAGE

    12/44

    Multipathing 2(4)

    13

    Device discovery initiates the path discovery and forms MP set for each

    device MC SYSMAN IO AUTO, SDA > SHOW DEVICE DGAxx shows MP Set

    First path discovered is considered primary

    The active path is called current

    Automatic path selection algorithms optimized to support Active/Active arrays

    Always active optimized (AO) paths are picked for I/O, if no alternative then active non-optimized (ANO) is picked [ how to fix this will discuss during EVA best practices]

    With latest firmware's on storage, very rare you will get connected toANO

    9/19/2011

  • 7/30/2019 OPENVMS I/O AND STORAGE

    13/44

    VMS switches its path to a LUN when:

    Multipathing 3(4)

    14

    An I/O error triggers mount verification (MV)

    Device is not reachable on current path and another path works

    MOUNT of a device with current path offline

    Manual path switch via SET DEVICE/SWITCH/PATH=

    Some local path becomes available when current path is MSCP

    Path switch from MSCP to local triggered by the poller [not if manual sw]

    Note: Any MV might trigger a path switchMV due to loss of cluster quorum

    MV due to SCSI error flush

    9/19/2011

  • 7/30/2019 OPENVMS I/O AND STORAGE

    14/44

    Multipathing 4(4)

    15

    MPDEV_POLLER is light weight, will poll all paths for availability SET DEVICE device/POLL/NOPOLL

    MV is not bad. It only indicates that OpenVMS validated your device Shadow devices can initiate and complete a MV ; Each shadow member operateindependently on a path switch

    But, MV followed with path switch is indication of failover/failback The operator logs will indicate the details

    SHOW DEV device/full will show details of path switch [time etc]

    SDA> SHOW DEV device logs lot of diagnostic information in MPDEV structure

    9/19/2011

  • 7/30/2019 OPENVMS I/O AND STORAGE

    15/44

    SCSI Error Poller

    9/19/201116

    We have seen customers reporting huge traffic of Unit Attention(UA) in

    SAN resulting in Cluster hangs, slow disk operations, high mountverifications etc!

    These UA are initiated due to any changes in SAN like FirmwareUpgrades, Bus Reset etc

    SCSI_ERROR_POLL is poller responsible for clearing the latchederrors (like SCSI UA) on all the fibre and SCSI devices, which canotherwise cause confusion in SAN

    By default the poller will be enabled SYSGEN>SHOW

    SCSI_ERROR_POLL

  • 7/30/2019 OPENVMS I/O AND STORAGE

    16/44

    After upgrading my SAN components, I see large number of mountverifications, does that indicate problem?

    Does multipathing do load balance? Or policies?

    I see too many mount verification messages in operator log, will it

    impact the volume performance (especially latency)

    How do I know if my paths are well balanced or not?

    How do I know my current path is active optimized or not?

    Does multipathing support Active/Active Arrays, ALUA, third partystorage, SAS device, SSD device

    OpenVMS Multipathing

    Customer/Field Concerns..

    9/19/201117

  • 7/30/2019 OPENVMS I/O AND STORAGE

    17/44

    Did you know QIO is one of the most heavily usedinterface in OpenVMS. We want to put it on diet.What should we do?

    1. Optimize QIO

    2. Replace QIO3. Provide alternative

    9/19/201118

  • 7/30/2019 OPENVMS I/O AND STORAGE

    18/44

    IOPERFORM/FastIO

    19

    Fast I/O is a performance-enhanced alternative to performing QIOs

    Substantially reduces the setup time for an I/O request

    Fast I/O uses buffer objects to eliminate the I/O overhead ofmanipulating I/O buffers locked memory doubly mapped

    Performed using the buffer objects and the following system services:

    sys$io_setup ,sys$io_perform, sys$io_cleanup

    sys$create_bufobj (jacket) /sys$create_bufobj_64

    $ dir sys$examples:io_perform.c

    System management considerations:

    SYSGEN parameter MAXBOBMEM limits memory usage (defaults to 100)

    VMS$BUFFER_OBJECT_USER identifier is required for process

    Creating buffer objects once and using them for the time of application is faster

    9/19/2011

  • 7/30/2019 OPENVMS I/O AND STORAGE

    19/44

    Impact of IOPERFORM/FASTIO

    20

    Resource usage is reduced by20-70% depending on load andsystem Small size random workloads doublesthe throughput with increased loads

    Larger size sequential workloadsoperate same

    0

    50100

    150200

    250300350400

    450

    1 2 3 4 5 6 7Threads

    I/O Data Rate (MB/sec)

    128K_READ

    128K_READ_QIO

    128K_WRITE

    128K_WRITE_QIO

    05000

    100001500020000

    2500030000

    350004000045000

    1 2 4 8 16 32 64

    Threads

    Throughput (IOPS)

    8K_READ

    8K_READ_QIO

    9/19/2011

  • 7/30/2019 OPENVMS I/O AND STORAGE

    20/44

    What you should know?

    NUMA/RAD Impact

    22 9/19/2011

    DGA100

    P1

    BL890c i2 (architecturally 4 Blades conjoined)

  • 7/30/2019 OPENVMS I/O AND STORAGE

    21/44

    NUMA/RAD Impact

    23

    In RAD based system, each RAD is made of CPU, Memory and I/O

    devices

    The accessibility of I/O devices from local to remote domains results inaccessing remote memory and remote interrupt latency

    3000

    3100

    3200

    3300

    3400

    3500

    3600

    3700

    0 1 2 3 4 5 6 7 8

    IO

    Rate

    RAD #

    Impact of RAD on I/O Device

    Opt/sec9/19/2011

    10-15 % overheadOptimized Performance

  • 7/30/2019 OPENVMS I/O AND STORAGE

    22/44

    RAD Guidelines for I/O

    9/19/201124

    Keep I/O Devices close to process which is heavily accessing it

    Make use of FASTPATHING efficiently Make sure to FASTPATH the Fibre Devices close to the process which is initiating the I/O

    The overhead involved in handling the remote I/O can impact the throughput [chart]

    FASTPATH algorithms assign the CPU on round robin basis

    Statically load balance the devices across multiple RADs

    Make use to SET PROC/AFFINITY to bind processes with high I/O

    Use SET DEVICE device/PREFERRED_CPUS

  • 7/30/2019 OPENVMS I/O AND STORAGE

    23/44

    STORAGE BEST PRACTICES

    25

  • 7/30/2019 OPENVMS I/O AND STORAGE

    24/44

    EVA4400 EVA6400 EVA8400 P6300 P6500

    Model HSV300 HSV400 HSV450 HSV340 HSV360

    Memory /controller pair 4GBytes 8GBytes 14/22GBytes 4GBytes 8GBytes

    Host Ports / controllerpair

    4 FC

    20 w switches

    8 FC 8 FC 8 FC, 0 GbE

    4 FC, 8 1GbE

    4 FC, 4 10GbE

    8 FC, 0 GbE

    4 FC, 8 1GbE

    4 FC, 4 10GbE

    Host Port speed 4Gb/s FC4Gb/s FC 4Gb/s FC 8Gb/s FC

    1Gb/s iSCSI10Gb/s iSCSI/FCoE

    8Gb/s FC1Gb/s iSCSI

    10Gb/siSCSI/FCoE

    Device Ports, # 4 8 12 8 16

    Device Ports, Speed 4Gb/s FC 4Gb/s FC 4Gb/s FC 6Gb/s 6Gb/s

    # 3-1/2 drives 96 216 324 120 240

    # 2-1/2 drives 0 0 0 250 500Max. Vdisk 1024 2048 2048 1024 2048

    I/O Read Bandwidth 780 MB/s 1,250 MB/s 1,545 MB/s 1,700 MB/s 1,700 MB/s

    I/O Write Bandwidth 590 MB/s 510 MB/s 515 MB/s 600 MB/s 780 MB/s

    Random Read I/O 26,000 IOPs 54,000 IOPs 54,000 IOPs 45,000 IOPs 55,000 IOPs

    Speeds and Feeds

    EVA Differences

    9/19/201127

  • 7/30/2019 OPENVMS I/O AND STORAGE

    25/44

    After upgrading the OS or applying the patch, I/O response has

    become slowerWe see 5-6 millisecond delay in completion of I/O compared to yesterday

    After moving to new blade in same SAN environment, our CRTL FSYNCis running slow

    After upgrading, we see additional CPUs ticks for copy, delete andrename

    Our database is suddenly responding slow

    Some nodes in cluster see high I/O latency after mid-night

    Customer wants to know if this storage is enough for next 5 years

    Customer is migrating from older version of EVA to newer version, canyou advise

    General I/O issues reported

    9/19/201128

  • 7/30/2019 OPENVMS I/O AND STORAGE

    26/44

    Maximum number of storage performance issuesreported are due to mis-configuration of SANcomponents

    9/19/201129

  • 7/30/2019 OPENVMS I/O AND STORAGE

    27/44

    Best Practices..1(6)

    30

    Number of disks influences performance - YesFill the EVA with as many disk drives as possible.

    Tests have shown linear growth in throughput (small random)

    Number of disk groups influences performance No

    In mixed load environments, it would be ok to have random vs.sequential application disk groups

    Vraid level influences performance YesVraid1 best performance over the widest range of workloads; however,

    Vraid5 is better for some sequential-write workloadsVraid0 provides the best random write workload performance , but noprotection Use for non-critical storage needs

    9/19/2011

  • 7/30/2019 OPENVMS I/O AND STORAGE

    28/44

    Best Practices..2(6)

    31

    Fibre channel disk speed - Yes10K vs 15K Speed 15K rpm disks provide highest performance

    Large-block sequential I/O, speed doesnt matter, but capacity matters

    Small-block random I/O, 30-40% gains in request rates is seen

    Best price-performance

    For the equivalent cost of using 15k rpm disks, consider using more 10k rpmdisks.

    Combine disks with different performance characteristics inthe same disk groupDo not create separate disk groups to enhance performance

    9/19/2011

  • 7/30/2019 OPENVMS I/O AND STORAGE

    29/44

    Best Practices..4(6)

    32

    Mixing disk capacities Yes

    The EVA stripes LUN capacity across all the disks in a disk group. The larger disk willhave more LUN capacity leading into imbalanced density

    No control over the demand to the larger disks.

    Use disks with equal capacity in a disk group.

    Read cache management influences performance and always ENABLE

    LUN count Yes, No Good to have few LUNs per controller

    Depends on Host Requests and Queue Depths

    Monitor OpenVMS Queue depth

    Transfer size Yes Impacts SEQUENTIAL Workloads Tune the write transfer size to be a multiple of 8K and no greater than 128K.

    OpenVMS Max Transfer Size is 128K for Disk and 64K for Tapes!DEVICE_MAX_IO_SIZE

    9/19/2011

  • 7/30/2019 OPENVMS I/O AND STORAGE

    30/44

    Best Practices..3(6)

    33

    SSD performance - Yes, Yes, YesSSDs are about 12 times better

    OpenVMS performed 10x times better than FC [next slide details]

    Workloads like transaction processing, data mining, and databasessuch as Oracle are ideal

    Spread SSDs evenly across all available back end loopsSSDs and HDDs may be mixed in the same drive enclosure

    Monitor your application and EVA, accordingly can assign SSD orHDD to individual Controllers, or enable write through mode for SSDs

    can help [Experiment!!]Customers use SSD drives to keep the critical path data, where theresponse time is un-compromised

    9/19/2011

  • 7/30/2019 OPENVMS I/O AND STORAGE

    31/44

    10xFaster

    SSD Drive Through EVA

    Mixed Load, 8 Disks SSD/FC DG on EVA4400 Smaller IOs(4K/8K) showed 9-10x times sustained increased in IOPS and MB/sec

    with increase in load for SSD carved LUNs compared to FC

    With 10 times faster response time, SSD carved LUN was able to deliver 10 timemore performances and bandwidth for smaller size IOs

    OpenVMS 8.4 Performance Results

    0

    2000

    4000

    6000

    8000

    10000

    12000

    14000

    1 2 4 8 16 32 64 128 256

    IOPS

    Threads

    4K Mixed QIO - FC vs SSD

    QIO_FC QIO_SSD

    0

    50

    100

    150

    200

    250

    0 50 100 150 200 250 300

    ResTime(msec)

    Threads

    4K Mixed QIO FC vs SSD

    FC SSD

    10x Faster

    OpenVMS V8.4

    9/19/201134

  • 7/30/2019 OPENVMS I/O AND STORAGE

    32/44

    Best Practices..5(6)

    35

    Controller balancing Yes, Yes, Yes

    Maximize the utilizationActive/Active present LUNs simultaneously through both controllers, Ownership is only one

    Manually load balance LUN ownership to both controllers (use EVAPerf) either through usingCV EVA or using OpenVMS SET DEVICE/SWITCH/PATH='PATH_NAME' 'DEV_NAME'command

    Preferred Path During the initial boot of the EVA the preferred path parameter is read and

    determines the managing controller [see below figure for options]Verify LUN ownership is reassigned after a failed controller has been repaired

    Balance the workload as evenly as possible across all the Host Ports

    HSVxxxHSVxxx

    DGA99

    DGA99 does I/O on these ports

    DGA99 answers Inquiry on these ports

    Command View EVA Preferred Path Settings9/19/2011

  • 7/30/2019 OPENVMS I/O AND STORAGE

    33/44

    Controller Load Imbalance & No Equal Port Load Distribution

    Customer Scenario

    36 9/19/2011

  • 7/30/2019 OPENVMS I/O AND STORAGE

    34/44

    Best Practices..6(6)

    37

    Ensure there are no HW issues Specially Battery failure (Cache Battery

    failure causes to change to Write-Through mode, hence Writeperformance becomes an issue), Device loop failure etc. Drive reportingtimeouts

    Deploy the array only in supported configurations

    Stay latest on EVA firmware!! BC and CA have different best practices and beyond the scope of this

    discussion

    9/19/2011

  • 7/30/2019 OPENVMS I/O AND STORAGE

    35/44

    Some Points to Remember

    38

    Large latencies may be quite natural in some contexts, such as an array

    processing large I/O requests

    9/19/2011

    Array processor utilizations tend to be high under intense, small-block,transaction oriented workloads but low under intense large-blockworkloads

  • 7/30/2019 OPENVMS I/O AND STORAGE

    36/44

    8K workloadspushing close to

    EVA MaxThroughputs!

    P6500 36G RAID 5 Volume 4G FC Infrastructure

    OpenVMS IO Data

    40

    0

    51015202530354045

    138 286 367 383 404 411 412 412

    resptime(msec)

    MB/sec

    Sequential Read (MB/sec)

    0

    1

    2

    3

    4

    5

    6

    6562 12935 22287 31888 38136 40052 43452 45504 46202

    resptime(msec)

    IOPS

    Random Read (IO/sec)

    Higher throughputs can beobtained with smaller blocks.

    Usually smaller blocks need alot of processing power

    Higher bandwidth can be obtainedwith larger blocks.Larger blocks can drain theinterconnects faster due large datatransfers.

    9/19/2011

    128K IOs pushing4G FC line speeds!

  • 7/30/2019 OPENVMS I/O AND STORAGE

    37/44

    STORAGE PERFORMANCEANALYSIS TOOLS &

    REFERENCES

    9/19/201141

  • 7/30/2019 OPENVMS I/O AND STORAGE

    38/44

    Storage Performance Tools

    9/19/201142

    OpenVMS Host Utilities

    T4, TLViz Disk, FCP, VEVAMON (older EVAs) SDA > FC [for fibre devices], PKR/PKC [for SAS devices]

    SYS$ETC: SCSI_INFO.EXE, SCSI_MODE.EXE, FIBRE_SCAN.EXE

    Many more..

    EVAPerf Command Line EVA Performance Data Collector EVA Performance Advisor [ Year End Release]

    XP Performance Advisor

    Storage Essential Performance Pack

  • 7/30/2019 OPENVMS I/O AND STORAGE

    39/44

    Slated to release soon Can participate in early adaptor program

    Salient aspects of HP P6000 Performance Advisor

    Integrated with Command View 10.0 in single pane of glass

    User centric design compliance

    Features:Dashboard

    Threshold monitoring & notificationKey metric

    Chart

    Report

    Quick setup Events

    Database

    9/19/201143

  • 7/30/2019 OPENVMS I/O AND STORAGE

    40/44

    HP StorageWorks Sizing Tool

    Sizing EVA

    44 9/19/2011

  • 7/30/2019 OPENVMS I/O AND STORAGE

    41/44

    References EVA Performance

    45

    HP StorageWorks Enterprise Virtual Array A tactical approach to

    performance problem diagnosis. HP Document Library

    9/19/2011

    http://h20195.www2.hp.com/v2/GetPDF.aspx/4AA2-0994ENW.pdfhttp://h20195.www2.hp.com/v2/GetPDF.aspx/4AA2-0994ENW.pdfhttp://h20195.www2.hp.com/v2/GetPDF.aspx/4AA2-0994ENW.pdfhttp://h20195.www2.hp.com/v2/GetPDF.aspx/4AA2-0994ENW.pdfhttp://h20195.www2.hp.com/v2/erl.aspx?cc=us&lc=enhttp://h20195.www2.hp.com/v2/erl.aspx?cc=us&lc=enhttp://h20195.www2.hp.com/v2/GetPDF.aspx/4AA2-0994ENW.pdfhttp://h20195.www2.hp.com/v2/GetPDF.aspx/4AA2-0994ENW.pdfhttp://h20195.www2.hp.com/v2/GetPDF.aspx/4AA2-0994ENW.pdfhttp://h20195.www2.hp.com/v2/GetPDF.aspx/4AA2-0994ENW.pdf
  • 7/30/2019 OPENVMS I/O AND STORAGE

    42/44

    Business Manager (Rohini Madhavan) [email protected]

    Office of Customer Programs [email protected]

    Questions/Comments

    9/19/201146

    mailto:[email protected]?subject=From%202011%20BC%20Storage%20Updatesmailto:[email protected]?subject=From%202011%20BC%20Storage%20Updatesmailto:[email protected]?subject=From%202011%20BC%20Storage%20Updatesmailto:[email protected]?subject=From%202011%20BC%20Storage%20Updates
  • 7/30/2019 OPENVMS I/O AND STORAGE

    43/44

    THANK YOU

  • 7/30/2019 OPENVMS I/O AND STORAGE

    44/44

    EVAModel Controller Model Firmware

    EVA3000 HSV1001.XXX, 2.XXX or 3.XXX (Latest 3110)

    EVA5000 HSV110

    EVA3000 HSV1014.XXX (Latest 4100)

    EVA5000 HSV111

    EVA4000 HSV200 or HSV200-A

    5.XXX or 6.XXX (Latest 6220)

    EVA6000 HSV200 or HSV200-A

    EVA4100 HSV200-B

    EVA6100 HSV200-B

    EVA8000 HSV210 or HSV210-A

    EVA8100 HSV210-BEVA4400 HSV300 09XXXXXX or 10000000

    EVA6400 HSV400095XXXXX, or 10000000

    EVA8400 HSV450

    P6300 HSV34010000090

    P6500 HSV360

    EVA Models - Reference