IBM Fs9500 - VMware Implementation - Sg248505
IBM Fs9500 - VMware Implementation - Sg248505
Redbooks
IBM Redbooks
October 2022
SG24-8505-01
Note: Before using this information and the product it supports, read the information in “Notices” on
page xv.
This edition applies to IBM Spectrum Virtualize 8.5 and VMware vSphere ESXi Version 7.0.
© Copyright International Business Machines Corporation 2021, 2022. All rights reserved.
Note to U.S. Government Users Restricted Rights -- Use, duplication or disclosure restricted by GSA ADP Schedule
Contract with IBM Corp.
Contents
Figures . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vii
Tables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xi
Examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xiii
Notices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xv
Trademarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xvi
Preface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xvii
Authors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xvii
Now you can become a published author, too! . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xix
Comments welcome. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xx
Stay connected to IBM Redbooks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xx
Chapter 1. Introduction. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.1 IBM and VMware. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.2 Overview of IBM Spectrum Virtualize and IBM FlashSystem . . . . . . . . . . . . . . . . . . . . . 2
1.2.1 IBM Spectrum Virtualize . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.2.2 IBM FlashSystem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
1.2.3 Key IBM Spectrum Virtualize terminology. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
1.2.4 Key VMware terminology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
1.3 Overview of IBM FlashSystem with VMware. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
Contents v
Abbreviations and acronyms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 185
viii IBM FlashSystem and VMware Implementation and Best Practices Guide
5-7 Storage Replication Adapters view . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 114
5-8 Site Recovery view . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 115
5-9 IBM HyperSwap . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 117
5-10 Read operations from hosts on either site are serviced by the local I/O group. . . . . 118
5-11 Write operations from hosts on either site are serviced by the local I/O group. . . . . 120
5-12 Write operations from hosts on primary are replicated . . . . . . . . . . . . . . . . . . . . . . . 121
5-13 Write operations from hosts on Secondary are forwarded and replicated . . . . . . . . 121
5-14 Non-Uniform host access vMSC . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 123
5-15 Uniform host access vMSC. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 124
5-16 Three data store architectures . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 125
6-1 vVols prerequisites . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 132
6-2 NTP time zone . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 133
6-3 Reviewing the certificate information . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 134
6-4 Reviewing the Subject Alternative Name value . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 134
6-5 Secure Communications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 135
6-6 Update Certificate . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 135
6-7 Update Certificate . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 136
6-8 Create Host Cluster . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 137
6-9 Create Host Cluster . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 138
6-10 Make Host Cluster . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 138
6-11 Add Host . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 139
6-12 Entering the details . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 140
6-13 Selecting the Host Cluster . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 141
6-14 Hosts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 142
6-15 Selecting Modify Host Types . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 142
6-16 Clicking Modify . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 143
6-17 Rescanning the storage adapters . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 143
6-18 Enable vVol toggle becomes available . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 144
6-19 Required objects for vVol support are created . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 145
6-20 Copy the following URL: option. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 147
6-21 Selecting Storage Providers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 148
6-22 New Storage Provider . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 148
6-23 Operation failed message . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 149
6-24 Newly added Storage Provider is showing online and active . . . . . . . . . . . . . . . . . . 149
6-25 Selecting New Datastore . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 150
6-26 Selecting vVol . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 150
6-27 Defining the name of the new vVol data store . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 151
6-28 Selecting the hosts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 151
6-29 Summary window . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 152
6-30 Reviewing the Datastores tab . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 152
6-31 Selecting Unmount Datastore . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 155
6-32 Unmount Datastore option . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 156
6-33 Policies and Profiles view . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 157
6-34 Selecting the policy to remove . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 157
6-35 Removing the IBM Spectrum Connect Storage Provider . . . . . . . . . . . . . . . . . . . . . 158
6-36 Deleting the child pools that were allocated to any vVol Storage Spaces . . . . . . . . 159
6-37 Storage credentials . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 160
6-38 User account that is used by IBM Spectrum Connect . . . . . . . . . . . . . . . . . . . . . . . 160
6-39 Selecting Migrate . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 162
6-40 Identifying the newly created vVol data store . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 163
6-41 Reviewing the tasks to ensure that the VMs successfully migrated . . . . . . . . . . . . . 163
6-42 Volumes by Pool view . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 164
6-43 Selecting the vVol-enabled child pool . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 165
Figures ix
6-44 Displaying the Name column . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 166
7-1 IBM Storage Insights architecture . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 171
7-2 IBM Storage Insights System overview for block storage . . . . . . . . . . . . . . . . . . . . . . 172
7-3 Adding vCenter Server . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 173
7-4 ESXi host details with daily response time, IO rate, and most active volumes on IBM
Storage Insights . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 174
7-5 VMDK level performance monitoring on IBM Storage Insights Pro . . . . . . . . . . . . . . 175
8-1 Collecting a support package in the GUI . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 178
8-2 Collecting IBM Spectrum Connect logs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 180
2-1 Creating a claim rule for an IBM Spectrum Virtualize system to set the path selection
limit to 1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
2-2 Output of using the esxcli storage hpp device list command . . . . . . . . . . . . . . . . . . . . 20
3-1 XCOPY transfer size 4096 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35
3-2 Verifying device VAAI support where “naa.xxx” stands for device identifier . . . . . . . . . 38
3-3 PowerCLI command that is used to evaluate the VAAI status . . . . . . . . . . . . . . . . . . . 39
4-1 The svcinfo lsadminlun command. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66
4-2 The svcinfo lsmetadatavdisk command . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 78
4-3 Verifyin gthat the underlying VDisk is in an operational state . . . . . . . . . . . . . . . . . . . . 79
4-4 The vvold.log file . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 80
6-1 The lsownershipgroup command . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 153
6-2 The lsprovisioningpolicy command . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 153
6-3 The mkmdiskgrp command. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 153
6-4 Identifying the user account that is used by IBM Spectrum Connect by using the
command-line interface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 161
6-5 The lsusergrp command . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 161
6-6 The lsmetadatavdisk command . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 161
8-1 Using svc_livedump to manually generate statesaves . . . . . . . . . . . . . . . . . . . . . . . . 179
8-2 Using svc_snap to generate a support package in the CLI . . . . . . . . . . . . . . . . . . . . 179
8-3 ESXi All Paths Down log signature . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 181
8-4 ESXi All Paths Down timeout . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 182
8-5 ESXi permanent device loss log signature . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 182
8-6 VMware Log Storage vMotion timeout . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 183
This information was developed for products and services offered in the US. This material might be available
from IBM in other languages. However, you may be required to own a copy of the product or product version in
that language in order to access it.
IBM may not offer the products, services, or features discussed in this document in other countries. Consult
your local IBM representative for information on the products and services currently available in your area. Any
reference to an IBM product, program, or service is not intended to state or imply that only that IBM product,
program, or service may be used. Any functionally equivalent product, program, or service that does not
infringe any IBM intellectual property right may be used instead. However, it is the user’s responsibility to
evaluate and verify the operation of any non-IBM product, program, or service.
IBM may have patents or pending patent applications covering subject matter described in this document. The
furnishing of this document does not grant you any license to these patents. You can send license inquiries, in
writing, to:
IBM Director of Licensing, IBM Corporation, North Castle Drive, MD-NC119, Armonk, NY 10504-1785, US
This information could include technical inaccuracies or typographical errors. Changes are periodically made
to the information herein; these changes will be incorporated in new editions of the publication. IBM may make
improvements and/or changes in the product(s) and/or the program(s) described in this publication at any time
without notice.
Any references in this information to non-IBM websites are provided for convenience only and do not in any
manner serve as an endorsement of those websites. The materials at those websites are not part of the
materials for this IBM product and use of those websites is at your own risk.
IBM may use or distribute any of the information you provide in any way it believes appropriate without
incurring any obligation to you.
The performance data and client examples cited are presented for illustrative purposes only. Actual
performance results may vary depending on specific configurations and operating conditions.
Information concerning non-IBM products was obtained from the suppliers of those products, their published
announcements or other publicly available sources. IBM has not tested those products and cannot confirm the
accuracy of performance, compatibility or any other claims related to non-IBM products. Questions on the
capabilities of non-IBM products should be addressed to the suppliers of those products.
Statements regarding IBM’s future direction or intent are subject to change or withdrawal without notice, and
represent goals and objectives only.
This information contains examples of data and reports used in daily business operations. To illustrate them
as completely as possible, the examples include the names of individuals, companies, brands, and products.
All of these names are fictitious and any similarity to actual people or business enterprises is entirely
coincidental.
COPYRIGHT LICENSE:
This information contains sample application programs in source language, which illustrate programming
techniques on various operating platforms. You may copy, modify, and distribute these sample programs in
any form without payment to IBM, for the purposes of developing, using, marketing or distributing application
programs conforming to the application programming interface for the operating platform for which the sample
programs are written. These examples have not been thoroughly tested under all conditions. IBM, therefore,
cannot guarantee or imply reliability, serviceability, or function of these programs. The sample programs are
provided “AS IS”, without warranty of any kind. IBM shall not be liable for any damages arising out of your use
of the sample programs.
The following terms are trademarks or registered trademarks of International Business Machines Corporation,
and might also be trademarks or registered trademarks in other countries.
AIX® IBM Cloud® Redbooks®
Easy Tier® IBM FlashCore® Redbooks (logo) ®
FlashCopy® IBM FlashSystem® Storwize®
HyperSwap® IBM Garage™ Tivoli®
IBM® IBM Spectrum®
The registered trademark Linux® is used pursuant to a sublicense from the Linux Foundation, the exclusive
licensee of Linus Torvalds, owner of the mark on a worldwide basis.
Microsoft, Windows, and the Windows logo are trademarks of Microsoft Corporation in the United States,
other countries, or both.
Red Hat, are trademarks or registered trademarks of Red Hat, Inc. or its subsidiaries in the United States and
other countries.
VMware, VMware vSphere, and the VMware logo are registered trademarks or trademarks of VMware, Inc. or
its subsidiaries in the United States and/or other jurisdictions.
Other company, product, or service names may be trademarks or service marks of others.
xvi IBM FlashSystem and VMware Implementation and Best Practices Guide
Preface
This IBM® Redbooks® publication details the configuration and best practices for using the
IBM FlashSystem® family of storage products within a VMware environment. The first version
of this book was published in 2021 and specifically addressed IBM Spectrum® Virtualize
Version 8.4 with VMware vSphere 7.0. This second version of this book includes all the
enhancements that are available with IBM Spectrum Virtualize 8.5.
Topics illustrate planning, configuring, operations, and preferred practices that include
integration of IBM FlashSystem storage systems with the VMware vCloud suite of
applications:
VMware vSphere Web Client (vWC)
vSphere Storage APIs - Storage Awareness (VASA)
vSphere Storage APIs – Array Integration (VAAI)
VMware Site Recovery Manager (SRM)
VMware vSphere Metro Storage Cluster (vMSC)
Embedded VASA Provider for VMware vSphere Virtual Volumes (vVols)
This book is intended for presales consulting engineers, sales engineers, and IBM clients who
want to deploy IBM FlashSystem storage systems in virtualized data centers that are based
on VMware vSphere.
Authors
This book was produced by a team of specialists from around the world.
xviii IBM FlashSystem and VMware Implementation and Best Practices Guide
Leandro Torolho is a Storage Client Technical Specialist for
US Public Market (West). Before joining the technical sales
team in 2015, he worked as a SAN/Storage subject matter
expert (SME) for several international clients. Leandro is an
IBM Certified IT Specialist and holds a bachelor’s degree in
computer science, and a post-graduate degree in computer
networks. He has 13 years of experience in storage services
and support, and is also a Certified Distinguished IT Specialist
by The Open Group.
Thanks to the following people for their contributions that made this book possible:
Pierre Sabloniere
IBM France
Matthew Smith
IBM UK
Markus Oscheka
IBM Germany
Erica Wazevski
IBM US
Rivka Pollack
IBM Israel
The authors team would also like to thank the following people from VMware for reviewing the
VMware-related technical content in the book:
Weiguo He|, Peter Toro, Paul Turner, Ken Werneburg
Find out more about the residency program, browse the residency index, and apply online at:
ibm.com/redbooks/residencies.html
Preface xix
Comments welcome
Your comments are important to us!
We want our books to be as helpful as possible. Send us your comments about this book or
other IBM Redbooks publications in one of the following ways:
Use the online Contact us review Redbooks form found at:
ibm.com/redbooks
Send your comments in an email to:
[email protected]
Mail your comments to:
IBM Corporation, IBM Redbooks
Dept. HYTD Mail Station P099
2455 South Road
Poughkeepsie, NY 12601-5400
This section describes the technical changes that are made in this edition of this
IBM Redbooks publication and in previous editions. This edition also might include minor
corrections and editorial changes that are not identified.
This book is the second edition of the IBM FlashSystem and, SG24-8505 that was originally
published on July 15, 2021. The new information that is included in this revision is described
next.
New information
Chapter 6, “Embedded VASA Provider for Virtual Volumes” on page 129 is a new chapter.
“Abbreviations and acronyms” on page 185 added.
VMware key terminology.
Claim rules to manage storage devices.
Various performance best practices updates.
Non-Volatile Memory Express (NVMe) volumes protocol types.
Small Computer System Interface (SCSI) and NVMe zones.
SCSI UNMAP feature.
Virtual machine (VM) level performance screens.
Changed information
Revised section on IBM Storage Insights integration
Chapter 1. Introduction
This IBM Redbooks publication describes the configuration and best practices for using
IBM Spectrum Virtualize based storage systems within a VMware environment. The first
version of this book was published in 2021 and addressed IBM Spectrum Virtualize 8.4 with
VMware Elastic Sky X integrated (ESXi) 7.0. This book is the second version, and it includes
all enhancements that are available with IBM Spectrum Virtualize 8.5.
This publication is intended for storage and VMware administrators. The reader is expected to
have a working knowledge of IBM Spectrum Virtualize and VMware. Initial storage and server
setup is not covered.
IBM is a VMware Technology Alliance Partner. IBM storage is deployed in the VMware
Reference Architectures Lab. Therefore, VMware products run well on IBM storage.
For this book, we mostly refer to the storage as IBM FlashSystem. However, other
IBM Spectrum Virtualize storage products work in similar fashion.
The primary function of IBM Spectrum Virtualize is block-level storage virtualization. IBM
defines storage virtualization as a technology that makes one set of resources resemble
another set of resources, preferably with more desirable characteristics.
The storage that is presented to the host is virtual and does not correspond to a specific
back-end storage resource so that IBM Spectrum Virtualize can perform many
enterprise-class features without impacting the hosts.
IBM Spectrum Virtualize first came to market in 2003 in the form of the IBM SVC. In 2003, the
SVC was a cluster of commodity servers attached to a storage area network (SAN). The SVC
did not contain its own storage. Instead, SVC used back-end storage that was provided from
other storage systems. At the time of writing, IBM Spectrum Virtualize supports up to 500+
different storage controllers.
IBM Spectrum Virtualize is software, so new features and functions are added regularly. Over
the years, IBM Spectrum Virtualize acquired a number of technologies that integrate it well
with VMware.
On the high end is Storage Class Memory (SCM), which is built around Intel's 3D-Xpoint and
Samsung Z-NAND technologies. SCM offers unprecedented throughput and low latency, with
limited capacities. The IBM FlashSystem 9500 control enclosure supports NVMe SCM/FCM
and solid-state drives (SSDs), and serial-attached SCSI (SAS) SSDs are supported only by
using expansion enclosures.
Figure 1-1 shows the front of an IBM FlashSystem 9200 enclosure. It is configured with 24
dual-ported NVMe drive slots.
Each system can be scaled up to include addition control enclosures or scaled out to include
more expansion enclosures.
Chapter 1. Introduction 3
Figure 1-2 shows the back of an IBM FlashSystem 9200 enclosure. In the center are two
canisters that run the IBM Spectrum Virtualize software and perform I/O. The canisters are
identical, except that the upper canister is flipped upside down. The far left and right sides
contain redundant power supplies.
IBM FlashSystem 9200 can be configured with various I/O ports that support various
protocols. The most common configuration is 8 or 12 Fibre Channel (FC) ports. These ports
can support the Fibre Channel protocol (FCP) or Non-Volatile Memory Express over Fibre
Channel (FC-NVMe).
Canister The IBM FlashSystem hardware that runs the IBM Spectrum
Virtualize software.
I/O group A pair of nodes or canisters that work together to service I/O.
Managed disk (MDisk) Either an array of internal storage or a logical unit number (LUN)
provided by an external storage controller.
Host The server that uses the storage. An ESXi server in this case.
IBM HyperSwap® A business continuity solution that copies data across two sites.
Data store SAN-based shared storage resources for ESXi clusters. Local
disk-based data stores cannot be shared by multiple servers.
Native multipathing (NMP) NMP plug-in. Most of the FC SAN-based storage systems are
controlled by NMP.
Claim rules Claim rules determine which multipathing module owns the paths to
a storage device. They also define the type of multipathing support
that the host provides to the device.
VMware vSphere Virtual vVols are virtual machine disk (VMDK) granular storage entities that
Volumes (vVols) are exported by storage arrays. vVols are exported to the ESXi host
through a small set of Protocol Endpoints (PEs). PEs are part of the
physical storage fabric, and they establish a data path from virtual
machines (VMs) to their respective vVols on demand. Storage
systems enable data services on vVols. The results of these data
services are newer vVols. Data services configuration and
management of virtual volume systems are exclusively done
out-of-band regarding the data path.
Chapter 1. Introduction 5
1.3 Overview of IBM FlashSystem with VMware
Figure 1-3 summarizes the various of VMware and IBM software components that are
discussed in this publication.
Two key components of integrating IBM Spectrum Virtualize with VMware are as follows:
IBM Spectrum Connect, which is a no additional charge software solution that is available
with all IBM storage. For more information about IBM Spectrum Connect, see Chapter 4,
“Integrating with VMware by using IBM Spectrum Connect” on page 41.
Integration with VMware Site Recovery Manager, which is not part of IBM Spectrum
Connect. For more information about Integration with VMware Site Recovery Manager,
see Chapter 5, “VMware and IBM Spectrum Virtualize multi-site guidelines” on page 105.
Figure 2-1 Configuration and connectivity for the test environment that is used in this book
Volumes that are mapped to a host cluster are assigned to all members of the host cluster
that use the same Small Computer System Interface (SCSI) ID. Before this feature was
implemented in IBM Spectrum Virtualize, as an example, an Elastic Sky X (ESX) cluster
would be created as a single host object, containing all the worldwide node names (WWNNs)
for the hosts in the cluster, up to 32.
The minimum size of a cluster is two nodes for vSphere high availability (HA) to protect
workloads if one host stops functioning. However, in most use cases, a 3-node cluster is more
appropriate because you have the option of running maintenance tasks on an ESXi server
without having to disable HA.
Configuring large clusters has benefits too. You typically have a higher consolidation ratio, but
there might be a downside if you do not have enterprise-class or correctly sized storage in the
infrastructure. If a data store is presented to a 32-node or a 64-node cluster and the virtual
machines (VMs) on that data store are spread across the cluster, there is a chance that you
will run into SCSI-locking contention issues. Using a VMware vSphere Storage APIs – Array
Integration (VAAI) aware array helps reduce this problem with Atomic Test and Set (ATS).
However, if possible, consider starting small and gradually growing the cluster size to verify
that your storage behavior is not impacted.
Figure 2-2 shows one of the VMware host clusters that was used in the test configuration for
this book. There are two hosts that are defined in the VMware host cluster.
Figure 2-2 VMware host clusters that were used in the test configuration
Note: Do not add Non-Volatile Memory Express (NVMe) hosts and SCSI hosts to the
same host cluster.
Figure 2-4 shows one of the hosts in the host cluster with the volumes that are connected to
it. The volumes were assigned to the host cluster, and not directly to the host. Any hosts that
are added to a host cluster have all of the volumes mapped to the host cluster automatically
assigned to the hosts.
Note: Private mappings can still be provisioned to individual hosts within a host cluster, for
example for storage area network (SAN) Boot configurations.
Figure 2-4 One of the hosts in the host cluster with the volumes that are connected to it
IBM Spectrum Virtualize supports commands that are used for SCSI offload and VMware
VAAI.
SCSI offload enables the host to offload some data operations to the storage system.
VAAI enables VMware hosts to also offload some operations to supported storage
systems.
Both technologies reduce traffic on the storage network, and load on the host. Hosts use
these offload commands to perform tasks such as formatting new file systems or performing
data copy operations without a host needing to read and write data. Examples are the WRITE
SAME and XCOPY commands. IBM Spectrum Virtualize 8.1.0.0 introduced support for WRITE
SAME when UNMAP is enabled. WRITE SAME is a SCSI command that tells the storage system to
write the same pattern to a volume or an area of a volume.
When SCSI UNMAP is enabled on IBM FlashSystem storage, it advertises this situation to
hosts. At versions 8.1.0.0 and later, some hosts respond to the UNMAP command by issuing a
WRITE SAME command, which can generate large amounts of I/O. If the back-end storage
system cannot handle the amount of I/O, volume performance can be impacted. IBM
Spectrum Virtualize offload throttling can limit the concurrent I/O that is generated by the
WRITE SAME or XCOPY commands.
IBM FlashSystem storage systems implement data reduction by using data reduction pools
(DRPs). A DRP can contain thin-provisioned or compressed volumes. DRPs also provide
more capacity to volumes in the pool by supporting data deduplication.
With a log-structured pool implementation, DRPs help to deliver more consistent performance
from compressed volumes. DRPs also support compression of all volumes in a system,
potentially extending the benefits of compression to all data in a system. Traditional storage
pools have a fixed allocation unit of an extent, and that does not change with DRPs. However,
features like Thin Provisioning and IBM Real-time Compression (RtC) use smaller allocation
units and manage this allocation with their own metadata structures. These features are
described as Binary Trees or Log Structured Arrays (LSAs).
Figure 2-5 shows the types of volumes that can be created in a DRP.
DRP fully allocated volumes provide the best performance for the IBM FlashSystem
products, but storage efficiency and space savings are not realized.
Thin-compressed volumes provide storage-space efficiency with the best performance of
the four options for space-efficient volumes.
Figure 2-5 Types of volumes that can be created in a data reduction pool
For more information about data reduction pools, see Implementation Guide for IBM
Spectrum Virtualize Version 8.5, SG24-8520.
2.2.1 iSCSI
iSCSI connectivity is a software feature that is provided by the SAN Volume Controller (SVC)
code. The iSCSI protocol is a block-level protocol that encapsulates SCSI commands into
Transmission Control Protocol/Internet Protocol (TCP/IP) packets. Therefore, iSCSI uses an
IP network rather than requiring the Fibre Channel (FC) infrastructure. The iSCSI standard is
defined by Request for Comment (RFC) 3720.
You can use the following types of iSCSI initiators in host systems:
Software initiator: Available for most operating systems (OSs), including IBM AIX, Linux,
and Windows.
Hardware initiator: Implemented as a network adapter with an integrated iSCSI processing
unit, which is also known as an iSCSI host bus adapter (HBA).
Ensure that the iSCSI initiators and targets that you plan to use are supported. Use the
following sites for reference:
IBM FlashSystem 8.5 Support Matrix
IBM FlashSystem 9x00 8.5
IBM System Storage Interoperation Center (SSIC)
An alias string can also be associated with an iSCSI node. The alias enables an organization
to associate a string with the iSCSI name. However, the alias string is not a substitute for the
iSCSI name.
Important: The cluster name and node name form part of the IQN. Changing any of them
might require reconfiguration of all iSCSI nodes that communicate with IBM FlashSystem.
2.2.2 iSER
IBM FlashSystem that run IBM Spectrum Virtualize v8.2.1 or greater support iSER for host
attachment, which is implemented by using RDMA over Converged Ethernet (RoCE) or
Internet Wide-Area RDMA Protocol (iWARP). This feature supports a fully Ethernet-based
infrastructure (and not Fibre Channel) in your data center:
IBM FlashSystem internode communication with 2 or more IBM FlashSystem in a cluster.
HyperSwap.
Using iSER requires that an Ethernet adapter is installed in each node, and that dedicated
Remote Direct Memory Access (RDMA) ports are used for internode communication. RDMA
enables the Ethernet adapter to transfer data directly between nodes. The direct transfer of
data bypasses the central processing unit (CPU) and cache and makes transfers faster.
2.2.3 FC-NVMe
The NVMe transport protocol provides enhanced performance on high-demand
IBM FlashSystem drives. NVMe is a logical device interface specification for accessing
non-volatile storage media. Host hardware and software use NVMe to fully leverage the levels
of parallelism possible in modern solid-state drives (SSDs).
Compared to the SCSI protocol, NVMe improves I/O and brings performance improvements
such as multiple, long command queues, and reduced latency. SCSI has one queue for
commands, unless multi-queue support such as blk_mq is enabled on the operating system,
and you are limited to the number of cores in the CPUs on the host.
NVMe is designed to have up to 64 thousand queues. In turn, each of those queues can have
up to 64 thousand commands that are processed simultaneously. This queue depth is much
larger than SCSI typically has. NVMe also streamlines the list of commands to only the basic
commands that Flash technologies need.
IBM FlashSystem implements NVMe by using the FC-NVMe protocol. FC-NVMe uses the
Fibre Channel protocol as the transport so that data can be transferred from host memory to
the target, which is similar to RDMA. For more information about NVMe, see IBM Storage and
the NVM Express Revolution, REDP-5437.
Every physical FC port on IBM FlashSystem storage supports four virtual ports: one for SCSI
host connectivity, one for FC-NVMe host connectivity, one for SCSI host failover, and one for
FC-NVMe host failover. Every NVMe virtual port supports the functions of NVMe discovery
controllers and NVMe I/O controllers. Hosts create associations (NVMe logins) to the
discovery controllers to discover volumes or to I/O controllers to complete I/O operations on
NVMe volumes. Up to 128 discovery associations are allowed per node, and up to 128 I/O
associations are allowed per node. An extra 128 discovery associations and 128 I/O
associations per node are allowed during N_Port ID virtualization (NPIV) failover.
At the time of this writing, IBM Spectrum Virtualize 8.5 supports a maximum of 64 NVMe
hosts. For more information, see IBM Support.
If FC-NVMe is enabled on the IBM FlashSystem, each physical WWPN reports up to four
virtual WWPNs. Table 2-1 lists the NPIV ports and port usage when FC-NVMe is enabled.
Table 2-1 NPIV ports and port usage when FC-NVMe is enabled
NPIV port Port description
Primary Port The WWPN that communicates with back-end storage if the IBM FlashSystem is virtualizing any
external storage.
SCSI Host Attach The virtual WWPN that is used for SCSI attachment to hosts. This WWPN is a target port only.
Port
Failover SCSI The standby WWPN that is brought online only if the partner node in an I/O group goes offline. This
Host Port WWPN is the same WWPN as the primary host WWPN of the partner node.
NVMe Host The WWPN that communicates with hosts for FC-NVMe. This WWPN is a target port only.
Attach Port
Failover NVMe The standby WWPN that is brought online only if the partner node in an I/O group goes offline. This
Host Attach Port WWPN is the same WWPN as the primary host WWPN of the partner node.
For more information about FC-NVMe and configuring hosts to connect to IBM FlashSystem
storage systems by using FC-NVMe, see VMware ESXi installation and configuration for
NVMe over Fibre Channel hosts
Figure 2-6 shows how to add a FC-NVMe host to an IBM FlashSystem from the Add Host
window.
The window that is shown in Figure 2-8 opens and shows the available SCSI and NVMe
devices and data stores.
NVMe devices are managed by the VMware high-performance plug-in (HPP). To see the
NVMe devices, run the esxcli storage hpp device list command on esxcli.
The IBM Spectrum Virtualize software that runs on IBM FlashSystem storage uses the SCSI
protocol to communicate with its clients, and presents storage space in form of SCSI logical
units (LUs) identified by SCSI logical unit numbers (LUNs).
Note: In formal practice, LUs and LUNs are different entities. In practice, the term LUN is
often used to refer to a logical disk or LU.
Since most applications do not directly access storage, but work with files or records, the OS
of a host must convert these abstractions to the language of storage, which are vectors of
storage blocks that are identified by logical block addresses within an LU. In IBM Spectrum
Virtualize, each of the externally visible LUs is internally represented by a volume, which is an
amount of storage that is taken out of a storage pool. Hosts use the SCSI protocol to send I/O
commands to IBM FlashSystem storage to read and write data to these LUNs.
As with FC-NVMe host attachment, if NPIV is enabled on the IBM FlashSystem storage
system, hosts attach to a virtual WWPN. Table 2-1 on page 14 lists the SCSI and Failover
Host Attach Ports.
VMware V7.0u2 and later supports RoCE v2 as host connectivity for IBM Spectrum Virtualize
8.5 storage systems.
RNICs can use RDMA over Ethernet through RoCE encapsulation. RoCE wraps standard
InfiniBand payloads with Ethernet or IP over Ethernet frames, which is sometimes called
InfiniBand over Ethernet. There are two main RoCE encapsulation types:
RoCE v1
Uses dedicated Ethernet Protocol Encapsulation to send Ethernet packets between
source and destination MAC addresses by using Ethertype 0x8915.
RoCE v2
Uses dedicated UDP over Ethernet Protocol Encapsulation to send IP UDP packets by
using port 4791 between source and destination IP addresses. UDP packets are sent over
Ethernet by using source and destination MAC addresses.
RoCE v2 is not compatible with other Ethernet options, such as RoCE v1.
For more information about configuring the VMware ESXi for NVMe over RDMA on
IBM FlashSystem storage systems, see Configuring the VMware ESXi operating system for
NVMe over RDMA host.
Tip: The VMware preferred flag can be set on a path. This flag is not applicable if the path
selection policy is set to Most Recently Used.
Fixed
The Fixed policy uses the designated preferred path flag if it is configured. Otherwise, it uses
the first working path that is discovered at system start time. If the ESXi host cannot use the
preferred path or it becomes unavailable, the ESXi host selects an alternative available path.
The host automatically returns to the previously defined preferred path when it becomes
available. This policy is the default for LUNs that are presented from an active/active storage
array.
Round-Robin
The Round-Robin policy is the recommended policy for IBM FlashSystem products. This path
selection policy uses a round-robin algorithm to load balance paths across all LUNs when
connecting to a storage array. This policy is the default for VMware starting with ESXi 5.5. You
must explicitly set Round-Robin for versions earlier than ESXi 5.5.
With Asymmetric Logical Unit Access (ALUA) in an active/active storage array, such as the
IBM FlashSystem 9200 and 9500 systems, only the optimized paths to the preferred control
enclosure node are used for transferring data. Round-Robin cycles through only those
optimized paths. You should configure pathing so that half the LUNs are preferred by one
control enclosure node, and the other half are preferred by the other control enclosure node.
The default path selection limit is IOPS, and the default value is 1000 IOPS before the path
changes. In some cases, a host can experience latency to storage with no latency seen on
the SAN. In these cases, the load of 1000 IOPS saturates the bandwidth of the path.
Lowering this value can increase storage performance and help prevent this cause of latency.
The recommended path-selection limit setting for IBM FlashSystem is to use IOPS and set
the value to 1. For more information about the IOPS limit, see Adjusting Round Robin IOPS
limit from default 1000 to 1 (2069356).
Example 2-1 Creating a claim rule for an IBM Spectrum Virtualize system to set the path selection limit to 1
esxcli storage nmp satp rule add -s VMW_SATP_ALUA -V IBM -M "2145" -c tpgs_on --psp="VMW_PSP_RR"
-e "IBM arrays with ALUA support" -O "iops=1"
To configure the claim rule, use vSphere Host Profile window, as shown in Figure 2-9.
Figure 2-9 Configuring a claim rule by using the Host Profile window
Note: Existing and previously presented devices must be manually set to Round-Robin
with an IOPS limit of 1. Optionally, the ESXi host can be restarted so that it can inherit the
multipathing configuration that is set by the new rule.
In vSphere 7.0 Update 1 and earlier, NMP remains the default plug-in for local NVMe devices,
but you can replace it with HPP. Starting with vSphere 7.0 Update 2, HPP becomes the
default plug-in for local NVMe and SCSI devices, but you can replace it with NMP.
By default, ESXi passes every I/O through the I/O scheduler. However, using the scheduler
might create internal queuing, which is not efficient with the high-speed storage devices.
You can configure the latency sensitive threshold and enable the direct submission
mechanism that helps I/O to bypass the scheduler. With this mechanism enabled, the I/O
passes directly from Pluggable Storage Architecture (PSA) through the HPP to the device
driver.
For the direct submission to work properly, the observed average I/O latency must be lower
than the latency threshold that you specify. If the I/O latency exceeds the latency threshold,
the system stops the direct submission, and temporarily reverts to using the I/O scheduler.
The direct submission is resumed when the average I/O latency drops below the latency
threshold again.
Note: HPP does not benefit when the systems perform lower than 200,000 IOPS.
Example 2-2 shows how to list the devices that are controlled by the HPP.
Example 2-2 Output of using the esxcli storage hpp device list command
[root@localhost:~] esxcli storage hpp device list
eui.70000000000004b5005076081280000c
Device Display Name: NVMe Fibre Channel Disk (eui.70000000000004b5005076081280000c)
Path Selection Scheme: LB-RR
Path Selection Scheme Config: {iops=1,bytes=10485760;}
Current Path: vmhba64:C0:T0:L3
Working Path Set: vmhba64:C0:T0:L3, vmhba65:C0:T0:L3
Is SSD: true
Is Local: false
Paths: vmhba64:C0:T0:L3, vmhba65:C0:T1:L3, vmhba65:C0:T0:L3, vmhba64:C0:T1:L3
Use ANO: false
Fixed
With this scheme, a designated preferred path is used for I/O requests. If the preferred path is
not assigned, the host selects the first working path that is discovered at start time. If the
preferred path becomes unavailable, the host selects an alternative available path. The host
returns to the previously defined preferred path when it becomes available again.
When you configure FIXED as a path selection mechanism, select the preferred path.
The preferred method is to use only WWPN zoning. You should not mix zoning types.
WWPN-based zoning is more flexible than Switchport-based and is required if the IBM
Spectrum Virtualize NPIV feature is enabled. Switch-port based zoning can cause failover of
the NPIV ports to not work correctly, and in certain configurations can cause a host to be
connected to the IBM FlashSystem on both the physical and virtual WWPNs.
For more information about the NPIV feature and switch-port zoning, see Using Switch
Port-Based Zoning with the IBM Spectrum Virtualize NPIV Feature.
Naming convention
When you create and maintain a Storage Network zoning configuration, you must have a
defined naming convention and zoning scheme. If you do not define a naming convention and
zoning scheme, your zoning configuration can be difficult to understand and maintain.
Environments have different requirements, which means that the level of detailing in the
zoning scheme varies among environments of various sizes. Therefore, ensure that you have
an understandable scheme with an appropriate level of detailing for your environment. Then,
use it consistently whenever you change the environment.
Aliases
Use zoning aliases when you create your IBM FlashSystem zones. Aliases make your zoning
easier to configure and understand and minimize errors. Define aliases for the
IBM FlashSystem physical WWPNs, the SCSI-FC WWPNs, and the FC-NVMe WWPNs if you
have that feature enabled.
Tip: If you have enough IBM FlashSystem ports available and you have many hosts that
you are connecting to an IBM FlashSystem, you should use a scheme to balance the hosts
across the ports on the IBM FlashSystem. You can use a simple round-robin scheme, or
you can use another scheme, such as numbering the hosts with the even-numbered hosts
zoned to the even-numbered ports and the odd-numbered hosts zoned to the
odd-numbered ports. Whichever load-balancing scheme that you choose to use, you
should ensure that the maximum number of paths from each host to each volume is four
paths. The maximum supported number is eight paths. The recommended number is four
paths per volume.
Note: Do not add NVMe and SCSI ports to the same zone.
For SAN zoning best practices, see the IBM San Zoning Best Practices at Support page.
In recent years, Brocade released the Peer Zoning feature. Cisco released a similar feature
that is called Smart Zoning. Both features allow multiple initiators to be in the same zone, but
prevent them from connecting to each other. They can connect only to target ports in the
zone, which allows multiple hosts to be in the same zone, but prevents the issue of a
malfunctioning host port from affecting the other ports.
For VMware clusters, the preferred zoning configuration is to have the ports for all of the hosts
in the cluster in a zone with the IBM FlashSystem virtual WWPN.
Brocade Peer zoning must be enabled for this zone on Brocade fabrics. Brocade Peer
Zoning was introduced in FOS v7.4.x.
For more information about Brocade Peer zoning, see Brocade Fabric OS Administration
Guide 9.0.x.
Cisco Smart Zoning must be enabled for this zone on Cisco fabrics. Cisco Smart Zoning
was introduced in NX-OS v5.2.x.
For more information about Cisco Smart Zoning, see “Configuring and Managing Zones”
in Cisco MDS Family 9000 NX-OS Fabric Configuration Guide.
Large-scale workloads with intensive I/O patterns require adapter queue depths greater than
the PVSCSI default values. At the time of writing, the PVSCSI queue depth default values are
64 (for device) and 254 (for adapter). You can increase PVSCSI queue depths to 254 (for
device) and 1024 (for adapter) inside a Windows or Linux VM.
When you work with VMFS data stores, consider the following items:
Data store extents: Do not span more than one extent in a data store. The
recommendation is to have a 1:1 ratio between the data store and the volume.
Block size: The block size on a VMFS data store defines the maximum file size and the
amount of space a file occupies. VMFS5 and VMFS6 data stores support the block size of
1 MB.
Storage vMotion: Storage vMotion supports migration across VMFS, virtual storage area
network (VSAN), and VMware vSphere Virtual Volume (vVol) data stores. A vCenter
Server performs compatibility checks to validate Storage vMotion across different types of
data stores.
Storage Distributed Resource Scheduler (SDRS): VMFS5 and VMFS6 can coexist in the
same data store cluster. However, all data stores in the cluster must use homogeneous
storage devices. Do not mix devices of different formats within the same data store cluster.
Device Partition Formats: A new VMFS5 or VMFS6 data store uses a globally unique
identifier (GUID) partition table (GPT) to format the storage device. The GPT format
enables you to create data stores larger than 2 TB. If your VMFS5 data store was
previously upgraded from VMFS3, it continues to use the master boot record (MBR)
partition format, which is characteristic for VMFS3. Conversion to GPT happens only after
you expand the data store to a size larger than 2 TB.
Starting from vSphere 6 VMware introduced IO reservations with SIOC. When reservations
are used, the same I/O injector that is used for checking latency also samples the input/output
operations per second (IOPS) capabilities of a data store. When the configured IOPS
reservation that is set on the VMs exceeds the capabilities of the observed IOPS capabilities
of that data store, IOPS is distributed to the VMs proportionally to their percentage of the
number of set reservations.
For critical systems, the usual recommendation is to not employ limits or throttling on the VMs
resources. Even though SIOC falls into the throttling category, it also provides a great fail-safe
for unavoidable and unpredictable contention. This function might be helpful when there are
multiple VMDKs that share a data store for manageability reasons.
For more information about SIOC on data stores used by production databases, see Storage
I/O control for critical apps is a great idea.
Note: The goal of Distributed Resource Scheduler (DRS) I/O load-balancing is to fix
long-term prolonged I/O imbalances, VMware vSphere SIOC addresses short-term burst
and loads.
You do not need to adjust the threshold setting in most environments. If you change the
congestion threshold setting, set the value based on the following considerations:
A higher value typically results in higher aggregate throughput and weaker isolation.
Throttling does not occur unless the overall average latency is higher than the threshold.
If throughput is more critical than latency, do not set the value too low. For example, for
Fibre Channel disks, a value below 20 ms might lower peak disk throughput. A high value
(above 50 ms) might allow high latency without significant gain in overall throughput.
A lower value results in lower device latency and stronger VM I/O performance isolation.
Stronger isolation means that the shares controls are enforced more often. Lower device
latency translates into lower I/O latency for the VMs with the highest shares, at the cost of
higher I/O latency experienced by the VMs with fewer shares.
A low value (lower than 20 ms) results in lower device latency and isolation among I/Os at
the potential cost of a decrease in aggregat4e data store throughput.
Setting the value high or low results in poor isolation.
SDRS consists of features that can be used to balance storage space and load between data
stores by using Storage vMotion to migrate VMs. Depending on your environment, you can
automate these tasks or decide to be notified and implement actions yourself.
If the latency for a data store exceeds the threshold (default: 15 ms) over a percentage of
time, SDRS migrates VMs to other data stores within the data store cluster until the latency is
below the threshold limit. SDRS might migrate a single VM or multiple VMs to reduce the
latency for each data store below the threshold limit. If SDRS is unsuccessful in reducing
latency for a data store, it (at a minimum) tries to balance the latency among all data stores
within a data store cluster.
When I/O metrics are enabled, SIOC is enabled on all data stores in the data store cluster.
Note: The I/O latency threshold for SDRS should be lower than or equal to the SIOC
congestion threshold.
Powered-on VMs with snapshots are not considered for space balancing.
Tip: As a general recommendation, consider using a data store cluster or SDRS whenever
possible. However, make sure to disable latency-based rules in Easy Tier environment.
SDRS simplifies VM placement when creating, deploying, or cloning VMs. SRDS provides
recommendations for balancing on space and I/O. In manual mode, recommendations can
be applied on a case-by-case basis.
For answers to frequently asked questions about SDRS, see SDRS FAQ (2149938).
Figure 3-1 shows an illustration of the RDM. RDM is a symbolic link from a VMDK file on a
VMFS to a raw LUN.
An RDM offers a number of benefits, but it should not be used in every situation. In general,
virtual-disk files are preferred over RDMs for manageability purpose.
The most common use case for RDM is Microsoft Cluster Server (MSCS). In an MSCS
clustering scenario that spans multiple hosts, which can be a mix of virtual and physical
clusters, the cluster data and quorum disk are to be configured as RDMs.
The VMFS data store is hosted by a single volume on a storage system, such as the IBM
FlashSystem 9200. A single VMFS data store can have hundreds or even thousands of
VMDKs.
The integration of vVols with IBM Spectrum Virtualize storage systems is dependent upon the
vSphere application programming interfaces (APIs) for Storage Awareness (VASA). These
APIs facilitate VM-related tasks that are initiated at the vSphere level to be communicated
down to the storage system.
IBM support for VASA is provided by IBM Spectrum Connect. IBM Spectrum Connect is an
out-of-band VASA Provider, which enables the communication between vSphere and the
storage system along the control plane.
IBM FlashSystem manages vVols at the storage level and enables the flexible and dynamic
provisioning of VM storage that is required of a truly software-defined storage environment.
For information about how to implement vVols with IBM FlashSystem, see Chapter 4,
“Integrating with VMware by using IBM Spectrum Connect” on page 41.
Because the storage for vVols is allocated as a child pool, it is important to consider the
structure of the parent pools from which these child pools are allocated. The following
sections describe the two contrasting approaches to defining parent pools and a description
of how their usage might influence the vVols environment.
We can use vSphere policies (for example, Gold, Silver, and Bronze) to select the appropriate
class of storage when we provision VMs.
Figure 3-2 A vVols environment where parent and child pools are segregated by drive class
With the introduction of vVols, by defining a range of storage services on IBM Spectrum
Connect, policies can become far more interesting and useful than the simple Gold, Silver,
and Bronze model.
Each of the policies (gold, silver, and bronze) can be further subdivided. For example, we
might divide our solid-state drive (SSD) parent pool into two distinct child pools. One child
pool is linked to an encrypted storage service, and the other is associated with an
unencrypted storage service. This approach provides the vSphere administrators with the
flexibility to provision VMs on storage that matches the requirements of the application, on a
per-VM basis.
Because vVols are a special volume, Easy Tier can manage their extents in an identical
fashion.
A hot (frequently used) extent of a vVol is promoted to faster storage, such as SSD.
A cold (infrequently used) extent of a vVol is moved onto slower drives.
A vVols implementation that takes advantage of Easy Tier can provide greater simplicity for
the storage administrator. By defining a child pool within an Easy Tier enabled parent pool,
the storage system is enabled to flexibly manage the extents of any vVols created therein.
This flexibility removes the requirement for a choice of storage class when the vSphere
administrator initially provisions the VM. Such an approach can also minimize the need for
Storage vMotion tasks because Easy Tier eliminates the requirement to manually migrate
vVols onto faster or slower storage as the needs of an application change.
Figure 3-3 on page 33 demonstrates a vVols configuration, based on a single parent pool,
with Easy Tier enabled.
Figure 3-3 The simplified approach to vVols provisioning that can be implemented by enabling Easy
Tier
VAAI helps the ESXi performance because the storage area network (SAN) fabric is not used
and fewer central processing unit (CPU) cycles are needed because the copy does not need
to be handled by the host.
In vSphere 5.x and later releases, these extensions (VAAI operations) are implemented as
T10 Small Computer System Interface (SCSI) commands. As a result, with the devices that
support the T10 SCSI standard (such as the IBM FlashSystem) your ESXi host can
communicate directly and does not require the VAAI plug-ins.
The following types of operations are supported by the VAAI hardware acceleration for
IBM FlashSystem:
Atomic Test and Set (ATS), which is used during creation and locking of files on the VMFS
volume
Clone Blocks, Full Copy, and extended copy (XCOPY), which is used to copy or migrate
data within the same physical array
Zero Blocks/Write Same, which is used when creating VMDKs with an eager-zeroed thick
provisioning profile
SCSI UNMAP, which is used to reclaim storage space
ATS is a standard T10 SCSI command with opcode 0x89 (SCSI Compare and Write (CAW)).
The ATS primitive has the following advantages where LUNs are used by multiple applications
or processes at one time:
Significantly reduces SCSI reservation contentions by locking a range of blocks within an
LUN rather than issuing a SCSI reservation on the entire LUN.
Enables parallel storage processing.
Reduces latency for multiple ESXi hosts accessing the same LUN during common.
Increases cluster scalability by greatly extending the number of ESXi hosts and VMs that
can viably reside simultaneously on a VMFS data store.
Note: All newly formatted VMFS5 and VMFS6 data stores use the ATS-only mechanism if
the underlying storage supports it. SCSI reservations are never used.
VMFS3 volumes that are upgraded to VMFS5 must be manually upgraded to ATS-only so
that it is easier to redeploy the data store and migrate the VMs to the data store.
Note: The use of ATS heartbeating is not supported on the following platforms:
ESXi hosts that run version 5.5 update 2 or later
ESXi version 6.0 before update 3
During high-latency events, ATS heartbeats might timeout, which results in ATS miscompare
errors. If multiple heartbeat attempts fail, the ESXi host might lose access to the data store in
which timeouts are observed.
ATS heartbeating increases the load on the system and can lead to access issues on busy
systems, particularly during maintenance procedures. To reduce this load, ATS heartbeats
can be disabled.
For VMware vSphere versions 5.5 and 6.0, the recommendation is to disable ATS
heartbeating because of host-disconnect issues.
To disable ATS heartbeats, run the following command:
# esxcli system settings advanced set -i 0 -o /VMFS3/UseATSForHBOnVMFS5
For VMware vSphere versions 6.5, 6.7 and 7.0, the recommendation is to enable ATS
heartbeating.
Similarly, XCOPY reduces the volume of traffic moving through the SAN when a VM is
deployed. It does so by synchronizing individual VM-level or file system operations (including
clone and migration activities) with the physical storage-level operations at the granularity of
individual blocks on the devices. The potential scope in the context of the storage is both
within and across LUNs.
The SCSI opcode for XCOPY is 0x83. As a best practice, set the XCOPY transfer size to 4096,
as shown in Example 3-1.
3.2.3 WRITE_SAME
Block Zeroing, Write_Same (Zero), or hardware-accelerated initialization use the
WRITE_SAME 0x93 SCSI command to issue a chain of identical write transactions to the
storage system. This command almost entirely eliminates server processor and memory use
by eliminating the need for the host to run repetitive identical write transactions. It also
reduces the volume of host HBA and SAN traffic when repetitive block-level write operations
are performed within VM disks to the IBM FlashSystem.
The WRITE_SAME 0x93 SCSI command allows the IBM FlashSystem to minimize internal
bandwidth consumption. For example, when provisioning a VMDK file with the
eagerzeroedthick specification, the Zero Block’s primitive issues a single WRITE_SAME
command that replicates zeros across the capacity range that is represented by the
difference between the provisioned capacity of the VMDK and the capacity that is consumed
by actual data. The alternative to using the WRITE_SAME command requires the ESXi host to
issue individual writes to fill the VMDK file with zeros. The same applies when cloning or
performing storage vMotion of a VM with eager-zeroed thick VMDKs.
When an IBM FlashSystem receives a SCSI UNMAP command, it overwrites the relevant region
of the volume with all-zero data, which allows thin-provisioned storage controllers (such as
the IBM FlashSystem) to reclaim physical capacity through garbage collection.
The main benefit is that this action helps prevent a thin-provisioning storage controller from
running out of free capacity for write I/O requests, which means that when thin-provisioned
storage controllers are used, SCSI Unmap should normally be left enabled.
With lower-performing storage (such as nearline arrays), extra I/O workload can be
generated, which can increase response times.
Enabling SCSI UNMAP does not affect data on older volumes that are created before using this
command. You must create data stores on newly created volumes, migrate data through
storage vMotion, and delete old volumes.
Host UNMAP commands can increase the free capacity that is reported by the data reduction
pool (DRP) when received by thin-provisioned or compressed volumes. SCSI UNMAP
commands also are sent to internal FlashCore Modules (FCMs) to free physical capacity.
For more information, see SCSI Unmap support in IBM Spectrum Virtualize systems.
Inside the VM, storage space is freed when you delete files on the thin virtual disk. Storage
space that is left by deleting or removing files from a VMFS data store can be freed up within
the file system. This free space is allocated to a storage device until the file system releases
or unmaps it. This operation helps the storage array to reclaim unused free space.
On VMFS6 data stores, ESXi supports the automatic asynchronous reclamation of free
space. VMFS6 can run the UNMAP command to release free storage space in the background
on thin-provisioned storage arrays that support unmap operations. Asynchronous unmap
processing has several advantages:
Unmap requests are sent at a rate (that can be throttled in vSphere), which helps to avoid
any instant load on the backing array.
Freed regions are batched and unmapped together.
I/O performance of other workloads is not impacted by the UNMAP command.
For information about the space-reclamation parameters for VMFS6 data stores, see Space
Reclamation Requests from VMFS Datastores.
The guest operating system notifies VMFS about freed space by sending the UNMAP
command. The UNMAP command that is sent from the guest operating system releases space
within the VMFS data store. The command proceeds to the array so that the array can reclaim
the freed blocks of space.
VMs on top of VMFS5 typically cannot pass the UNMAP command directly to the array; you
must run the esxcli storage vmfs unmap command to trigger unmaps from the IBM
FlashSystem. However, for a limited number of the guest operating systems, VMFS5
supports the automatic space reclamation requests.
To send the unmap requests from the guest operating system to the array, the VM must meet
the following prerequisites:
The virtual disk must be thin-provisioned.
VM hardware must be version 11 (ESXi 6.0) or later.
The advanced EnableBlockDelete setting must be set to 1.
The guest operating system must be able to identify the virtual disk as thin.
The following considerations apply when you use space reclamation with VMFS6:
VMFS6 processes the unmap request from the guest operating system (OS) only when
the space to reclaim equals 1 MB or is a multiple of 1 MB. If the space is less than 1 MB or
is not aligned to 1 MB, the unmap requests are not processed.
For VMs with snapshots in the default SEsparse format, VMFS6 supports the automatic
space reclamation only on ESXi hosts version 6.7 or later.
Space reclamation affects only the top snapshot and works when the VM is powered on.
Example 3-2 Verifying device VAAI support where “naa.xxx” stands for device identifier
[root@ESX1-ITSO:~] esxcli storage core device vaai status get -d naa.xxx
naa.xxx
VAAI Plugin Name:
ATS Status: supported
Clone Status: supported
Zero Status: supported
Delete Status: unsupported
You can verify and change your VAAI settings in host Advanced System Settings (Figure 3-4).
A value of 1 means that the feature is enabled. If the setting is host-wide, it is enabled if the
connected storage supports it.
Example 3-3 shows how to evaluate the VAAI status by using the PowerCLI command.
Example 3-3 PowerCLI command that is used to evaluate the VAAI status
# Get-VMHost | Get-AdvancedSetting -name *HardwareAccelerated* | select Name, value
Name Value
---- -----
DataMover.HardwareAcceleratedMove 1
VMFS3.HardwareAcceleratedLocking 1
DataMover.HardwareAcceleratedInit 1
Note: You might decide to increase a block size (MaxHWTransferSize for XCOPY), which is
processed by storage globally. Although we do not recommend changing default values,
you might notice a small improvement in the performance during Data Mover operations
(typically around 10%). This change is global and affects all your VAAI-enabled storage
devices that are connected to the ESXi. Therefore, changing the default values can have
unpredictable impact on different storage arrays.
IBM Spectrum Connect provides a web-based user interface (UI) that makes this entire
administration easy and straightforward. Use of the UI saves a significant amount of time in
setting up, connecting, and integrating the required storage resources into your cloud
environment.
Through its user credential, storage system, storage space, and service management
options, IBM Spectrum Connect facilitates the integration of IBM storage system resources
with the supported virtualization, cloud, and container platforms.
The following storage services, which can be considered as storage profiles, are defined in
IBM Spectrum Connect and delegated for use in VMware for simplified profile-based volume
provisioning:
VMware vSphere Web Client (vWC)
VMware vSphere Storage APIs - Storage Awareness (VASA)
VMware vRealize Operations (vROps) Manager
VMware vRealize Automation, and VMware vRealize Orchestrator (vRO)
Microsoft PowerShell
Note: The VASA for VMware vSphere Virtual Volumes (vVols)) function is also available
through the Embedded VASA Provider in IBM Spectrum Virtualize 8.5.1.0 or later. For
more information about using VASA or vVols with the Embedded VASA Provider, see
Chapter 6, “Embedded VASA Provider for Virtual Volumes” on page 129.
As shown in Figure 4-1, the IBM Spectrum Connect application communicates with IBM
storage systems by using command-line interface (CLI) commands over Secure Shell (SSH).
VMware also issues application programming interface (API) calls directly to IBM Spectrum
Connect over Transmission Control Protocol/Internet Protocol (TCP/IP). Therefore, IP
connectivity must exist between the IBM Spectrum Connect server and the Management IP
address (sometimes referred to as a Cluster IP) of the storage system. For security, some
network infrastructures might be segmented into virtual local area networks (VLANs) or have
isolation preventing the management of the storage system from being accessible from virtual
machines (VMs) within a vSphere environment. Check with your network administrator to
ensure that IP connectivity exists between the different components.
Communication between VMware vRO, vSphere, or vCenter and the IBM storage system by
using IBM Spectrum Connect is “out-of-band” and is therefore separate from I/O traffic
between a host (for example an Elastic Sky X integrated (ESXi) host) and the storage system.
If network connectivity issues occur, which prevent IBM Spectrum Connect from
communicating with either the cloud interface or the storage system, the I/O workload that is
running on the hosts is unaffected.
VM resource requirements might depend on the roles that are performed by the IBM
Spectrum Connect Server. The IBM Spectrum Connect server periodically (by default, it is
configured to run in 10-minute intervals) queries the storage system for an inventory of
objects such as VDisks, hosts, and FlashCopy mappings. This query populates a local cache
of the configuration, which can be used for cache can be used by the following services:
vROps for performance data
The IBM Storage Enhancements plug-in, which maintains constant awareness of host
objects that are defined on the storage system
Depending on the size and complexity of the IBM Spectrum Virtualize configuration, the
population task might take some time.
Tip: In large environments, consider increasing the population interval to allow sufficient
time for the task to complete. For more information about IBM Spectrum Connect, see IBM
Documentation at Working with multiple storage systems.
Note: Before you install the IBM Spectrum Connect application, it is a best practice to
configure Network Time Protocol (NTP) client on the Linux operating system (OS). Given
the multiple components in the end-to-end infrastructure any time-skew between
IBM Spectrum Connect, IBM Spectrum Virtualize, and the VMware platforms can
complicate debugging issues when reviewing logs.
Make note of the installation summary screen (Figure 4-2 on page 45), which provides details
about additional steps to configure firewall rules and SELinux, if required.
Notes:
When specifying the storage-system credentials, it is a best practice to create a
dedicated user account on the storage-system, which is easily identifiable as being
from an IBM Spectrum Connect server. This account is the User account that is used
when IBM Spectrum Connect issues CLI commands to the storage system. Having an
easily recognizable username assists with tasks (for examples, reviewing audit logs
within IBM Spectrum Virtualize) and make it clear that CLI commands were issued by
IBM Spectrum Connect.
The storage-system credentials are global and apply to all storage systems being
registered in IBM Spectrum Connect. When possible, consider using Lightweight
Directory Access Protocol (LDAP) authentication on the storage system to simplify user
account management.
Ensure that the associated user account is created on every storage system that is to be
registered and that the account was granted a suitable role.
4. Enter a suitable username and select the VASAUsers group created previously. Enter and
confirm the password for this user account then click Create, as shown in Figure 4-7 on
page 49.
5. When the initial configuration wizard is complete, in the IBM Spectrum Connect
management interface, you see a “Guided tour” that provides a brief overview of the
interface. When the tour is completed, empty inventory of IBM Spectrum Connect is
displayed (Figure 4-8).
6. Click the Interfaces pane, on the left side of the screen, to view and configure the Cloud
Interfaces. You can configure items such as a vCenter server for the IBM Storage
Enhancements or the connector for vRO.
Alternatively, click the Monitoring pane, on the right side of the screen, to configure the
vROps Manager integration.
You are prompted to enter only the IP or hostname of the storage system because the
Storage Credentials were defined in the initial configuration wizard (Figure 4-10).
Note: If you use VMware vSphere Virtual Volumes, ensure that the Storage Credentials
are configured with the “VASA Provider” role.
7. Click Add. The Storage System is now represented in the IBM Spectrum Connect UI
(Figure 4-11 on page 51).
IBM Spectrum Connect uses the concept of a Storage Service for simpler and more flexible
storage management.A Storage Service can also be described as a Storage Provisioning
Profile. If a storage administrator defines the capabilities on the service (or profile), all
volumes that are created in that service are created the same way and inherit the same
attributes.
However, a storage administrator might be reluctant to grant full access to a storage system
because if too many volumes are created, the capacity is difficult to manage. In this scenario,
the storage administrator can create child pools to present ring-fenced allocations of storage
capacity to a Storage Service. A vSphere administrator can then create volumes on-demand
within that storage allocation.
Tips:
When a Storage Service is dedicated to use with VMware vSphere Virtual Volumes,
ensure that the vVol Service checkbox is selected.
Consider a mix of configuration options, such as Space Efficiency, Encryption, Tier, and
Data Reduction.
Note: When you create the Storage Service and select the capabilities, Storage Systems
that are not compatible with the current selection are unavailable, such as:
When Synchronous Replication is selected, but the Storage System that is registered
in IBM Spectrum Connect is not yet configured for HyperSwap.
When Flash storage is requested, but storage pools do not exist in the Flash devices.
2. After you define the required capabilities of the storage profile, click Create.
The newly created Storage Service is now listed. Notice that allocated storage capacity
does not exist for this Storage Service. A storage resource needs to be associated to the
Storage Service.
Note: A Storage Resource can be either an existing parent pool within the IBM Spectrum
Virtualize storage system, or a new or existing child pool. A child pool is a ring-fenced
allocation of storage capacity that is taken from an existing parent pool.
This step creates a child pool of a defined name and capacity within a specified parent
pool.
Note: When using a HyperSwap configuration, you are asked to specify the parent pool
at both sites (Figure 4-17). This specification creates a paired child pool with the same
capacity that is defined at each site.
3. Verify that the Storage Resource allocation for the Storage Service is correct and click
Close (Figure 4-18 on page 57).
Tip: When you use vVol-enabled Storage Services, this step is optional. However it is a
best practice because it provides more visibility of the individual vVol-to-VM relationships.
Note the changes that occur when the Storage Service is delegated:
– The color changes to yellow on the Storage Service (Figure 4-19).
– The Delegate icon changes from to (Figure 4-20).
– The Allocated Storage capacity, which is described just under the selected vCenter
interface, is updated to reflect the new Storage Service delegation (Figure 4-20).
2. If you want to remove the delegation, click the Delegate icon ( ) icon to disassociate the
Storage Service mapping (Figure 4-20).
Note: The creation and removal of delegation of Storage Service to Interface does not
impact the existing volumes or host mappings on the storage system.
With the introduction of IBM Spectrum Virtualize 8.5.1.0 (or later), the vVol function can be
provided in either of two ways:
IBM Spectrum Connect: An isolated application that runs on a separate VM within the
vSphere environment that converts API calls from vSphere into CLI commands that are
issued to the IBM FlashSystem array.
Embedded VASA Provider: This application runs natively as a service on the
IBM FlashSystem configuration node, communicates directly with vSphere, and is not
dependent on any other external application or server.
The two models cannot be run in parallel, so you must choose which one to implement in your
infrastructure.
To learn more about the Embedded VASA Provider feature and to see whether it is
appropriate for your environment, see Chapter 6, “Embedded VASA Provider for Virtual
Volumes” on page 129.
The vVol architecture, introduced in VMware vSphere 6.0 with VASA 2.0, preserves the
concept of a traditional data store, maintaining familiarity and functions and also offers some
benefits of legacy Raw Device Mappings.
Also, in this scenario there are several VMs, distributed across many ESXi hosts, all
performing I/O operations to that volume generate a high workload. When investigating these
performance issues, the storage administrator might observe high IOPS measurements, and
high response times against that volume. In this example, the storage system would have
limited knowledge of where the I/O workload was coming from or the cause of the
performance issues. Similarly, if multiple VMs are performing sequential workloads to their
VMDK disks, the ESXi layer confuses the stream of I/O and cannot use the advanced caching
features of IBM Spectrum Virtualize.
Finally, when performing logical unit number (LUN)-level operations on the storage system
(for example, volume-level snapshots that use a FlashCopy map), all VMs that are on the
VMFS data store that is backed by that volume are affected.
However, RDMs do not offer the same level of functions when compared with VMFS data
stores, specifically regarding VM cloning and snapshots. RDMs also require a storage
administrator to present dedicated volumes for each VMDK file, which increases
storage-provisioning management overhead.
With vVol, the IBM storage systems become aware of individual VMDKs, which allows data
operations, such as snapshots, to be performed directly by the storage system at the VM
level. The storage system uses IBM Spectrum Connect (the VASA provider) to present vVols
to the ESXi host and inform the vCenter of the availability of vVol-aware storage by using
Protocol Endpoints (PEs).
One PE (also known as Administrative LUN or PE) is presented at each node in the
IBM Spectrum Virtualize cluster. A PE presents itself as a traditional storage device and is
detected by an ESXi server like a normal volume mapping. However, PEs offer a more
efficient method for vVols to be mapped and detected by ESXi hosts when compared to
traditional volume mappings and do not require a rescan of host bus adapter (HBA).
Storage services are configured on the IBM Spectrum Connect server by the storage
administrator. Storage services are then used to configure various storage containers with
specific capabilities that can later be configured as vVol data stores in the vSphere Client.
Note: To keep time synchronized between all components of the vVol infrastructure,
ensure that an NTP server is defined on the storage system, IBM Spectrum Connect
server, vCenter, and ESXi hosts.
3. Select a parent pool to store the Utility Volume. This volume is configured as a
space-efficient 2 TB volume, which stores the associated metadata database that is
required by IBM Spectrum Connect to store information that is required to manage the
vVol environment.
Although the volume is created with a capacity of 2 TB, it is unlikely to require more than
2 GB of capacity.
If possible, consider creating a mirrored copy of this Utility Volume by selecting a second
pool to store the additional copy.
4. Define credentials for a newly dedicated user account, which enables the IBM Spectrum
Control server to connect to the CLI of the storage system.
This process initially creates a User Group within the IBM Spectrum Virtualize Storage
system that is assigned with the VASAProvider role, and then creates the User Account by
using the specified credentials within that user group.
5. Click Enable (Figure 4-22).
Hosts with a generic host-type allow volumes to be mapped with a Small Computer
System Interface (SCSI) ID of 0 - 2048. However, ESXi detects only SCSI IDs 0 - 1023.
Hosts with a vVol (or adminlun) host type are limited to SCSI IDs 0 - 511. This warning
refers to existing volumes that might be mapped with a SCSI ID above 511 that are lost
when changing the host type to support vVols. Normally, the lowest available SCSI ID is
used when a volume is mapped to a host. Therefore, it is only in rare circumstances where
either:
– More than 511 volumes are mapped to a host or host cluster.
– A SCSI ID above 511 is specified when a previous mapping was created.
7. Verify the existing SCSI IDs by reviewing the existing host mappings for a host or host
cluster, as follows:
a. Select Hosts → Mappings in the navigation menu, and select All host mappings.
Sort the mappings in descending order by the SCSI ID column to show the highest
SCSI IDs in use (Figure 4-29).
b. Ensure that SCSI IDs above 511 are not in use before you change the host type.
id SCSI_id UID
0 0300000000000000 600507680CD00000DC000000C0000000
1 0301000000000000 600507680CD00000DC000000C0000001
e. Sort the LUN column in descending order and confirm that a PE exists for each node in
the storage system.
In a standard two-node (single I/O-group) cluster, two PEs are presented. SCSI IDs that
are used for the first PEs start at 768-769, and increase for each additional node in the
storage system.
2. Define the credentials that vCenter uses to connect to IBM Spectrum Connect when the
storage provider is registered (Figure 4-32).
Note: The URL must be in the format of https://<IP or FQDN of Spectrum Connect
server>:8440/services/vasa.
4. Click OK. The Storage Providers list now show the newly registered storage provider
(Figure 4-35).
In this example, we created and delegated various vVol-enabled storage services, such as:
Production systems and applications.
Development or test environments.
An application environment that requires encryption or extra security considerations and
must be isolated from other applications.
Each storage service can be configured with unique capabilities that are associated to Space
Efficiency, Encryption, or Storage Tier (Figure 4-36).
4. Select the hosts that require access to the data store (Figure 4-41) and click NEXT.
Review the summary and click FINISH.
Figure 4-42 Repeating the process for any additional vVol data stores
Therefore, it is a best practice to protect the IBM Spectrum Connect server to enable optimal
functions:
Where possible, ensure that VMware vSphere Fault Tolerance (FT) is used to ensure that
if there is an outage to the ESXi host that is running the IBM Spectrum Connect server,
then the wider infrastructure is still able to access the IBM Spectrum Connect integrations.
If FT is not available, then ensure that vSphere HA is used to minimize downtime of the
server hosting the IBM Spectrum Connect application.
Allocate appropriate compute resources to the IBM Spectrum Connect VM. Because the
applications run as a conduit between the multiple VMware interfaces and multiple storage
systems, performance of the end-to-end system is impacted if resources are limited.
Most importantly, ensure that all VMs associated with providing the vVol infrastructure are
not stored on a VMware vSphere Virtual Volume data store, which include the vCenter
Service Appliance (vCSA), the IBM Spectrum Connect server, and other servers that
these applications require to function.
If an outage occurs, the IBM Spectrum Connect server must start before any vVol
operations function.
If the IBM Spectrum Connect server depends on an external LDAP or NFS server that is
on a vVol data store, then it will fail to successfully start IBM Spectrum Connect services,
which means that vVol data stores will be inaccessible and VMs on the vVol data store will
be unavailable. If this situation occurs, contact IBM Support.
By creating the association between the vVol-enable storage service and the vCenter
interface in IBM Spectrum Connect, the IBM Storage Enhancements plug-in can assimilate
information from both systems.
The relationships between vVols and VMs can then be displayed for the vSphere
administrator.
2. Click IBM Storage vVols to list the detected vVols on the storage array and the vVols size
and type.
3. If required, export the list as a CSV file to generate a report.
For more information about the IBM Storage Enhancements plug-in, see 4.5, “IBM
Storage Enhancements for VMware vSphere Web Client” on page 82.
Therefore, VM-management tasks (for example, powering off a VM) fail when an upgrade is
running on the storage system. Automated services, such as VMware HA and Distributed
Resource Scheduler (DRS), also are affected because they send system commands by using
IBM Spectrum Connect.
Tip: After you perform an upgrade, it is possible that the vSphere Web Client will mark cold
VMs as inaccessible, which means that ESXi hosts were unable to start a new binding to
these VMs (expected during a code upgrade) and should not cause alarm.
To recover management of these VMs, the vSphere administrator should remove the
affected VMs from the inventory and then add them again.
To prevent an accidental action in the IBM Spectrum Virtualize from impact the IBM Spectrum
Connect objects, the storage system prevents manipulation of these objects unless the
manipulation is performed by a user with the VASA-Provider role. When a “superuser” or
similar account is used to change objects that IBM Spectrum Connect created, the following
error is reported:
CMMVC8652E The command failed as a volume is owned and has restricted use.
Warning: Do not make manual changes unless advised by a member of IBM Support
because it might cause more issues.
When configured in an HA configuration, both Active and Standby IBM Spectrum Connect
servers use the same database to ensure infrastructure consistency.
Tip: For more information about the HA configuration, see the IBM Spectrum Connect
User Guide available at IBM Fix Central.
If the IBM Spectrum Connect server is permanently unavailable (and no backup exists), a
new IBM Spectrum Connect server can be commissioned to recover the metadata database
on the metadata VDisk, and the environment can be resurrected. For more information about
this recovery process, contact IBM Support.
To query the active state of the metadata VDisk, connect to the Storage System by using the
CLI management interface and run the following command (Example 4-2).
If the status of the metadatavdisk reports corrupt, verify that the underlying VDisk is in an
operational state by checking the detailed view of the specific vdisk_id (vdisk_id 16 in
Example 4-3 on page 79). Verify that the underlying VDisk is in an operational state.
In rare circumstances, the output from the lsmetadatavdisk command shows corrupt, and
there are messages in the IBM Spectrum Connect hsgsvr.log file reporting the following
error:
CMMVC8580E The action failed because the metadata utility volume is corrupt.
This error might occur if the configuration node experienced issues when attempting to
access the internal mounted volume. To resolve this issue, put the configuration node into
service state by using the service assistant. After a few minutes, bring the node back in to the
cluster, and retry the lsmetadatavdisk command again. The metadata VDisk should now
report online status.
Note: If you are unsure of the status of the metadata VDisk, contact IBM Support.
4.4.9 Certificates
Certificates are used by vSphere, IBM Spectrum Connect, and IBM Spectrum Virtualize to
secure communication between the separate components. Therefore, ensure that the
IBM Spectrum Connect certificate is configured correctly.
If this certificate-issue occurs, regenerate the certificate in IBM Spectrum Connect with the
correct common name and fully qualified domain name, as follows:
1. Go to the IBM Spectrum Connect management interface window and click Server
Certificate (Figure 4-45).
2. In the Server Certificate window, click Generate (Figure 4-46 on page 81).
3. After the certificate regenerates, you must remove and re-register the Storage Provider.
Prerequisites
The following items are the prerequisites:
NTP: NTP client must be configured on the base Linux OS under the IBM Spectrum
Connect application, the IBM Spectrum Virtualize storage system, and vSphere (vCenter
and ESXi hosts). Given the multiple components in the end-to-end infrastructure, any
time-skew between IBM Spectrum Connect, IBM Spectrum Virtualize, and the VMware
platforms can complicate debugging issues when logs are reviewed.
Supported versions: Check IBM Documentation for the interoperability matrix for
supported versions of IBM Spectrum Connect, IBM Spectrum Virtualize, and vSphere. For
IBM Spectrum Connect V3.7, which was the latest version at the time of writing, see
Compatibility and requirements.
Restrictions
The following are the restrictions:
Array-based replication for vVol is not currently supported by IBM Spectrum Connect or
IBM Spectrum Virtualize. This items is planned for a future release.
vVols are not supported in a HyperSwap configuration.
Note: For more information about the types of vVols, see Virtual Volume Objects.
If more VMDKs are configured on the VM, then associated more vVols are created. If a
VM-level snapshot is taken of the VM, then an additional vVol is created for each VMDK
configured on the VM.
For the specific version of IBM Spectrum Virtualize, see the “Configuration Limits and
Restrictions for IBM Spectrum Virtualize” page in IBM Documentation. For example, see
V8.4.0.x Configuration Limits and Restrictions for IBM FlashSystem 9200.
The storage administrator can use the Storage Spaces and Storage Services objects in IBM
Spectrum Connect to complete the following actions:
Create a preset of specific storage capabilities.
Allocate pools of storage capacity (either parent pools or child pools) in which volumes
that are created by using the IBM Storage Enhancements vSphere plug-in will be located.
Delegate those presets to vCenter so they can be consumed by a vSphere administrator
as either VMFS data stores or RDMs.
The IBM Storage Enhancements for VMware vSphere Web Client plug-in is automatically
deployed and enabled for each vCenter server that is registered in the Interfaces window of
IBM Spectrum Connect.
The storage services that you attach on the IBM Spectrum Connect side become visible on
vSphere Client, and can be used for volume creation by using the IBM Storage
Enhancements for vSphere Client.
Before you begin, log out of vSphere Client browser windows on the vCenter server to which
you want to add IBM Spectrum Connect. If you stay logged in, you will be able to use the
extension only after you log out and log in to vCenter after IBM Spectrum Connect is added?
You need to add the vCenter servers for which you can later attach storage services that
would be visible and accessible on the vSphere Client side. You can add a single vCenter
server at a time.
When entering the vCenter credentials on the IBM Spectrum Connect side, verify that the
vCenter user has sufficient access level in vCenter to complete this procedure.
If the provided IP address and credentials are accepted by the vCenter server, it is added
to the list of servers on the Interfaces window. The yellow frame and the exclamation mark
in Figure 4-48 indicate that storage services are not yet delegated to interface.
Notes:
If you want to use the vSphere Web Client extension on all vCenter servers that
operate in linked mode, each server instance must be added to IBM Spectrum
Connect, which ensures that the extension is registered on all linked servers
properly.
The same vCenter server cannot be added to more than one IBM Spectrum
Connect instance. Any attempt to add an already registered vCenter server to
another IBM Spectrum Connect overrides the primary connection.
3. If you need to update the vCenter credentials that are being used by IBM Spectrum
Connect or to remove the vCenter interface, right-click the vCenter as displayed in the
Interfaces window, and click either Modify or Remove (Figure 4-49).
To provision volumes by using the IBM Storage Enhancements plug-in, complete the following
steps:
1. In the vSphere Client, select the Hosts & Clusters tab. Right-click the host or cluster to
which you want to provision volumes and select IBM Storage → Create new IBM
Storage volume (Figure 4-50).
2. In the Hosts Mapping field, click the arrow and select the hosts to which you want to map
the volumes. If you select a vSphere Cluster from the list, the volumes are mapped by
using the Host Cluster feature in IBM Spectrum Virtualize (see 2.1.1, “IBM FlashSystem
host clusters” on page 8). This mapping ensures that if more hosts are added to the Host
Cluster on the IBM Spectrum Virtualize system, they automatically inherit existing Host
Cluster mappings.
3. If a custom-volume mapping is required, click Custom, and select the boxes for each ESXi
host or cluster to which you want to map the volumes and click OK (Figure 4-52).
5. Select the Storage Service in which you want to create the volumes. If multiple Storage
Services exist, they are included in the list.
– Storage capabilities that were defined on the Storage Service are listed in the green
area of the window.
– A summary of the task is shown in the blue area of the window.
6. The Storage LUN value defines the SCSI ID to be used when mapping the volume to the
host or host cluster. Unless for a specific requirement, select Auto for the Storage LUN
value so that the SCSI ID can be automatically determined by the system.
7. Click OK to begin the storage provisioning task.
The status of the operation is displayed in the Recent Tasks tab of the vSphere Client
window (Figure 4-54).
8. Verify that the commands were issued correctly by checking the Audit Log (Figure 4-56) in
the IBM Spectrum Virtualize storage system. To access the Audit Log, log in to the web
interface of the Storage System and select Access → Audit Log. You can also use the
CLI to run the catauditlog command when you are logged in using SSH.
The volumes are created and mapped to the selected Hosts or Clusters, and the vSphere
administrator is now able to create VMFS data stores or RDMs from these volumes by
using the normal vSphere workflow (Figure 4-58).
For each vCenter server, the following IBM Storage categories are available to view for IBM
Spectrum Virtualize platforms:
Storage services
Storage spaces
Storage volumes
Storage vVols
Important: You might notice references to IBM consistency groups. However, this
integration applies only to IBM FlashSystem A9000/R storage systems.
To view additional information about the storage objects, complete the following steps:
1. Go to the vCenter Server under the Resources list from the Global Inventory Lists view in
the vSphere Client, and open an IBM Storage category to view additional information
about the objects that are currently delegated to the selected vCenter server
(Figure 4-59).
Figure 4-60 Viewing the capabilities that are defined on a Storage Service
3. To find specific information about a particular volume, select IBM Storage Volumes from
the menu on the left. Beneath the menu, select the specific storage volume (Figure 4-61).
4.6.1 Considerations
Volume protection is an IBM Spectrum Virtualize feature that prevents volumes from being
inadvertently deleted or unmapped from a host or host cluster. When attempting to delete a
volume or remove existing host mappings, this task might fail if volume protection is enabled
on the storage system, and the volume recently processed I/O operations. When this setting
is enabled, volumes must be idle before they can be deleted from the system or unmapped
from a host or host cluster. Volumes can be deleted only if they have been idle for the
specified interval (by default this interval is set to 15 minutes).
When volume protection is disabled, volumes can be deleted even if they recently processed
I/O operations. In the management GUI, select Settings → System → Volume Protection to
manage volume protection values on the system and on specific pools. You must have the
SecurityAdmin or Administrator role to change volume protection settings.
The Extend volume size task might fail if a thick-provisioned (fully allocated) volume is created
on the storage system, and a fast-format task is still in progress. If this situation occurs, wait
for the fast-formatting process to complete, and then run the command again.
When the Extend volume size task successfully completes, the LUN that is backing the data
store increases in size but the VMFS file system does not change. You must rescan the HBA
for each host that accesses the data store, so that the host can detect the change in volume
size. Then, you must expand the VMFS file system by right-clicking a data store and selecting
Increase Datastore Capacity to take advantage of the additional capacity.
To access the vRO management options, go to the Interfaces window of the IBM Spectrum
Connect server and add the vRO server interface, as shown in Figure 4-62.
The yellow frame and the exclamation mark (Figure 4-62) indicate that Storage Services are
not yet delegated to the interface. You can then manage the integration with vRO as
described in the following sections.
To download and install the IBM Storage plug-in package, complete the following steps:
1. On the Interfaces window, right-click the vRO server, and then select Modify.
2. On the bottom of the vRO Settings dialog, click Download plug-in package to save the
package to your local computer (Figure 4-63).
Alternatively, you can download the package from Downloading and installing the plug-in
package for vRO.
Installation is completed and the IBM Storage plug-in is displayed in the list of vRO
plug-ins.
8. Start the VRO Client and go to the Workflows tab.
10.Locate the Set Server and Token workflow and click Run.
11.Enter the following information in the correct fields in Figure 4-67:
– The server field: Enter the FQDN.
– The port field: Enter the port of the IBM Spectrum Connect server.
– The token field: Paste the token from step 3 on page 94, and click Run.
Tip: If you experience issues running the Set Server and Token workflow, retry the
procedure with a web browser in Incognito or Private Browsing modes.
After a Storage Service is created and allocated, complete the following steps:
1. Run the Create and Map a Volume workflow (Figure 4-69).
2. Click in the “Service on which the volume should be created” window and search for a
Storage Service (Figure 4-70 on page 97).
The VMware plug-in is available through the VMware Marketplace and is published as
“vRealize Operations Management Pack for IBM SAN Volume Controller and IBM Storwize
4.0.0”. For more information, see Management Pack for IBM SVC and Storwize.
When using IBM Spectrum Connect for vROps integration, the management pack can be
downloaded from the IBM Spectrum Connect GUI and then deployed on the vROps Manager
server. After a VMware vROps Manager server is registered on an instance of IBM Spectrum
Connect that is configured with storage systems, storage spaces, services, and vRealize
servers, the storage-related data is pushed to the vROps Managerserver in 5-minute intervals
by default.
The dedicated IBM storage system adapter that is deployed on the vROps Manager server
enables monitoring of the supported IBM storage system by using the vROps Manager. This
adapter reports the storage-related information, such as monitoring data of all logical and
physical elements, covering storage systems, storage domains, storage pools, volumes,
hosts, modules, target ports, disks, health status, events, thresholds, and performance. It also
provides the dashboards that display detailed status, statistics, metrics, and analytics data
alongside hierarchical flowcharts with graphic representation of IBM storage system
elements.
To download the PAK file from IBM Spectrum Connect, complete the following steps:
1. Go to the Monitoring window of the IBM Spectrum Connect GUI. The Set vROps Server
dialog is displayed (Figure 4-74).
To deploy the management package on the vROps, complete the following steps:
1. Access the vROps Manager administrative web console by using https://<hostname or
IP address of the vROps UI>.
2. Select Administration → Solutions → Repository.
3. In the Repository window, click ADD/UPGRADE to add a management package. The Add
Solution dialog is displayed.
100 IBM FlashSystem and VMware Implementation and Best Practices Guide
4. In the Add Solution dialog, click Browse and select the management package that is
downloaded from IBM Spectrum Connect (Figure 4-75). Click UPLOAD to start
deployment.
102 IBM FlashSystem and VMware Implementation and Best Practices Guide
To add the VROps Manager server to IBM Spectrum Connect, complete the following steps:
1. Go to the Monitoring window of the IBM Spectrum Connect GUI (Figure 4-77).
2. Enter the following information:
– IP/Hostname: IP address or FQDN of the vROps Manager server
– Username
– Password
3. Select the checkbox to confirm you installed the PAK file on the vROps Manager server.
After a vROps server connection is defined and storage systems are associated with the
vROps server, detailed monitoring information for these storage systems becomes available
in vROps.
104 IBM FlashSystem and VMware Implementation and Best Practices Guide
5
Copy services are a collection of functions that provide capabilities for business continuity,
disaster recovery, data migration, and data duplication solutions. This chapter provides an
overview and the preferred best practices guide for VMware and IBM Spectrum Virtualize for
copy services capabilities. These capabilities include FlashCopy, Metro Mirror, Global Mirror,
and VMware Site Recovery Manager.
This chapter describes some of the solutions that can help you prepare your environment to
recover from a disruption and includes the following sections:
“Copy Services overview” on page 106
“Storage Replication Adapter with VMware Site Recovery Manager” on page 109
“IBM HyperSwap with VMware vSphere Metro Storage Cluster” on page 116
The following copy services (relationships) are supported by IBM Spectrum Virtualize storage
system:
FlashCopy, for point-in-time copy
Metro Mirror, for synchronous remote copy
Global Mirror, for asynchronous remote copy
Global Mirror with Change Volumes (GMCV), for asynchronous remote copy for a
low-bandwidth connection
Note: All these services are supported by VMware SRM when using IBM Spectrum
Virtualize Family SRA.
5.1.1 FlashCopy
FlashCopy is known as a point-in-time copy. It makes a copy of the blocks from a source
volume and duplicates them to the target volumes.
106 IBM FlashSystem and VMware Implementation and Best Practices Guide
Tip: Metro Mirror can impact write latency. For best performance, use shorter distances
and create Metro Mirror relationships only between systems with similar performance.
Using SRM with IBM Spectrum Virtualize storage system can help you protect your virtual
environment.
SRM automates the failover processes and the ability to test failover processes or DR without
having a negative impact on the live environment, which helps you meet your recovery time
objectives (RTOs).
VMware SRM requires one vCenter server in each site with the respective licenses. Also, if
you are using SRM with IBM FlashSystem storage, you are required to use an IBM Spectrum
Virtualize SRA, which is described in 5.1.6, “Storage Replication Adapter” on page 108.
For more information about SRM, see the VMware Site Recovery Manager Documentation.
The adapter is used to enable the management of Advanced Copy Services (ACS) on
IBM FlashSystem Storage, such as Metro Mirror and Global Mirror (including changed
volumes).
The combination of SRM and SRA enables the automated failover of VMs from one location
to another, connected by either Metro Mirror or Global Mirror technology.
By using the IBM Spectrum Virtualize Family Storage Replication Adapter, VMware
administrators can automate the failover of an IBM FlashSystem 9200 at the primary SRM
site to a compatible system, such as another IBM FlashSystem 9200, 7200, or IBM SAN
Volume Controller at a recovery (secondary) SRM site.
108 IBM FlashSystem and VMware Implementation and Best Practices Guide
In a failover, the ESXi servers at the secondary SRM site mount the replicated data stores on
the mirrored volumes of the auxiliary storage system. When the primary site is back online,
perform fallback from the recovery site to the primary site by clicking Reprotect in the SRM.
For more information, see the IBM Spectrum Virtualize Family Storage Replication Adapter
documentation.
Figure 5-2 SRA and VMware SRM with IBM Spectrum Virtualize integrated solution
SRA configuration might vary depending on the specific site configuration. Consider the
following preparatory steps and configuration when using SRA with SRM.
110 IBM FlashSystem and VMware Implementation and Best Practices Guide
Ensure that for non-preconfigured environments, the recovery volumes remain unmapped.
Make sure that the recovery VMware ESX or ESXi hosts are defined as hosts at the
recovery site and report as online.
As a best practice, stop your currently installed version of the SRA container before running a
different version. Ensure that you satisfy all of the prerequisites that are listed in before you
run the SRA container. Follow the below steps to run the IBM Spectrum Virtualize storage
system SRA container on the SRM server.
Note: The IBM Spectrum Virtualize Family SRA installation creates a shortcut that is
named IBM Spectrum Virtualize Family SRA Configuration Utility.exe on the desktop.
The configuration utility must be run on both the protected (primary) site and the recovery
(secondary) site SRM host.
112 IBM FlashSystem and VMware Implementation and Best Practices Guide
Figure 5-5 SRM Appliance Management interface login
3. Click Upload. Go to the directory where you saved the SRA file, and double-click it. The
SRA upload process begins.
114 IBM FlashSystem and VMware Implementation and Best Practices Guide
Figure 5-8 Site Recovery view
Account for the following practices when working with SRA and VMware ARM:
Create Metro or Global Mirror relationships between the source and target VDisks and
add them to consistency groups, as explained in “Preparing the SRA environment” on
page 110.
Before you use the SRA, make sure that the Metro or Global Mirror relationships and
consistency groups are in a consistent synchronized state.
For Stretched Cluster, make sure that the two copies of a stretched volume are at different
sites and that both copies are online.
For IBM HyperSwap, make sure that the primary volume and secondary volume of a
HyperSwap volume are online.
Consider the following tips when you add an Array-air to the VMware SRM are as follows:
– Enter the same Common Information Model Object Manager (CIMOM) address and
Common Information Model (CIM) port as the primary IBM Spectrum Virtualize
Cluster, if you are using IBM Stretched Cluster or IBM HyperSwap.
– If M:N topology is being used, enter any one of the IBM Spectrum Virtualize Clusters
(N) on the remote site for the CIM address of the remote IBM Spectrum Virtualize
Cluster. The term M:N refers to the number of IBM Spectrum Virtualize Clusters on the
local site (M) and remote site (N). Calculations are not derived from this term. M:N is
used only to denote the replication process.
In this document, the focus is on solutions that rely both on VMware vSphere Metro Storage
Cluster (vMSC) and VMware SRM in relation to IBM Spectrum Virtualize.
116 IBM FlashSystem and VMware Implementation and Best Practices Guide
5.3.1 IBM HyperSwap
IBM Spectrum Virtualize HyperSwap is a dual-site solution that provides continuous
availability of data during planned and unplanned outages. If storage at either site goes
offline, HyperSwap automatically fails over storage access to the system at the surviving site.
When you configure a system with a HyperSwap topology, the system is split between two
sites for data recovery, migration, or high availability use cases. When a HyperSwap topology
is configured, each node or enclosure, external storage system, and host in the system
configuration must be assigned to one of the sites in the topology. Both node canisters of an
I/O group must be at the same site. This site must be the same site of any external storage
systems that provide the managed disks to that I/O group. When managed disks are added to
storage pools, their site attributes must match. This requirement ensures that each copy in a
HyperSwap volume is fully independent and spans multiple failure domains (Figure 5-9).
HyperSwap Volume is a group of volumes and remote copy relationship all working together
to provide the active-active solution, and ensure that data is synchronized between sites.
An active-active HyperSwap relationship exists between the Master and Auxiliary volumes to
facilitate the data synchronization and replication between sites.
Like a traditional Metro Mirror relationship, the active-active relationship attempts to keep the
Master Volume and Auxiliary Volume synchronized while also servicing application I/O
requests. The relationship uses the CVs as journaling volumes during the resynchronization
process (Figure 5-10).
Figure 5-10 Read operations from hosts on either site are serviced by the local I/O group
The HyperSwap Volume always uses the unique identifier (UID) of the Master Volume. The
HyperSwap Volume is assigned to the host by mapping only the Master Volume even though
access to the Auxiliary Volume is ensured by the HyperSwap function. For each HyperSwap
volume, hosts across both sites see a single volume that is presented from the storage
system with the UID of the Master Volume.
Cluster considerations
Consider the following tips when you work with HyperSwap and VMware vSphere Metro
Storage Cluster (vMSC):
One IBM Spectrum Virtualize-based storage system, which consists of at least two I/O
groups. Each I/O group is at a different site. Both nodes of an I/O group are at the same
site.
HyperSwap-protected hosts on IBM Spectrum Virtualize must be connected to both
storage nodes by using Internet Small Computer System Interface (iSCSI) or Fibre
Channel.
In addition to the two sites that are defined as failure domain, a third site is needed to
house a quorum disk or IP quorum application.
More system resources are used to support a fully independent cache on each site. This
allows full performance even if one site is lost.
118 IBM FlashSystem and VMware Implementation and Best Practices Guide
HyperSwap relationships
One site is considered as the Primary for each HyperSwap Volume or Consistency Group.
This site is dynamically chosen according to the site that writes more data (more than 75% of
write I/Os) to the volume or consistency group over a 20-minute period.
Note: Low write-throughput rates do not trigger a direction switch to protect against
unnecessary direction changes when experiencing a trivial workload.
While the I/O group on each site processes all reads from the hosts on that local site, any
write requests must be replicated across the inter-site link, which incurs added latency. Writes
to the primary site experience a latency of 1x the round-trip time (RTT) between the sites,
however due to the initial forwarding process, writes to the non-primary site will experience 2x
the round-trip time. This additional performance overhead should be considered when you
provision storage for latency-sensitive applications.
These relationships automatically run and switch direction according to which copy or copies
are online and up-to-date.
Relationships can be grouped into consistency groups, in the same way as other types of
remote-copy relationships. The consistency groups fail over consistently as a group based on
the state of all copies in the group. An image that can be used for disaster recovery is
maintained at each site.
Figure 5-11 Write operations from hosts on either site are serviced by the local I/O group
Read I/O is facilitated by whichever I/O group is local to the requesting host, which prevents
the I/O from having to transfer the long-distance link and incurring unnecessary latency.
120 IBM FlashSystem and VMware Implementation and Best Practices Guide
Figure 5-12 Write operations from hosts on primary are replicated
Figure 5-13 Write operations from hosts on Secondary are forwarded and replicated
Note: Whenever the direction of a HyperSwap relationship changes, there is a brief pause
to all I/O requests to that volume. In most situations, this pause is less than 1 second.
Where possible, consider how application workload to a single HyperSwap volume (or
HyperSwap consistency group) across sites can reduce the likelihood of repeated direction
changes.
IBM Spectrum Virtualize facilitates vMSC with the ability to create single storage cluster that
spans both sites, such that a data store must be accessible in both locations. In other words,
the data store must be able to read and be written to simultaneously from both sites by using
the HyperSwap feature of IBM Spectrum Virtualize. In a failure, the vSphere hosts are able to
continue read and write access to the data store from either location seamlessly and with no
impact on ongoing storage operations.
For every volume presented from the IBM FlashSystem, a preferred node is automatically
elected. To evenly distribute the workload across the IBM FlashSystem upon volume creation,
the preferred node usually alternates between each node in the I/O group.
When you map a volume to a host object and rescan the host bus adapter (HBA) on the ESXi
host, ESXi automatically identifies the available paths to both nodes in the I/O group, as
follows:
The paths to the preferred node for each volume are identified as the “Active/Optimized
paths”.
The paths to the non-preferred node are identified as “Active/Non-Optimized paths”.
By default, ESXi uses the Round-Robin path selection policy (PSP) to distribute I/O as
follows:
Over any available “Active/Optimized paths” to the preferred node.
Only fail over to “Active/Non-Optimized paths” if available paths to the preferred node do
not exist.
122 IBM FlashSystem and VMware Implementation and Best Practices Guide
Non-Uniform configuration
In Non-Uniform vMSC implementations, ESXi hosts use Small Computer System Interface
(SCSI) Asymmetric Logical Unit Access (ALUA) states to identify Active/Optimized paths to
the preferred node in the local I/O group and Active/Non-Optimized paths to the partner node.
The host has no visibility of the storage system at the remote site (Figure 5-14).
With Non-Uniform vMSC environments, if a storage failure occurs at the local site, the ESXi
hosts lose access to the storage because paths are not available to the storage system at the
remote site. However, this architecture might be useful when you run clustered applications
like Microsoft SQL or Microsoft Exchange with servers that are located at each site. It might
be preferable to have a clustered application fail over so that an application can continue to
run with locally available storage.
Each of these models applies at a single data store level. Therefore, it is possible to
incorporate all three methods in an IBM FlashSystem HyperSwap architecture.
124 IBM FlashSystem and VMware Implementation and Best Practices Guide
Figure 5-16 Three data store architectures
The negative aspect of this configuration is that there can be many ESXi servers and storage
resource at Site 2 that are idle.
In addition, if the workload is unregulated, the direction of the HyperSwap relationship can be
swapped repeatedly, which generates more I/O pauses for each change of direction.
Consider creating multiple VMFS data stores where the primary associated site alternates
between Site 1 and Site 2, and keep VM-storage resources local to hosts that are running the
VMs.
As detailed in “Write operations from ESXi hosts at remote site” on page 121, Alternative
Access Method ensures that instead of any write I/Os at either site having to be forwarded
over the inter-site link before being replicated to the remote site, they will be serviced by the
I/O group at the site where the I/O originated. This method reduces the overall latency and
increases the performance and throughput.
The intention is for a given HyperSwap volume or consistency group to keep VM I/O
workloads local to the host running the VM, which minimizes the workloads being driven from
a host at the non-primary site.
In a site outage at either site, vSphere high availability (HA) automatically recover the VMs on
the surviving site.
For more information about DRS Host/VM groups and rules, see Create a Host DRS Group.
In an ideal HyperSwap environment, you do not want VMs to move to the other site. Rather,
you want VMs to move to the other site only in a site failure, or intentionally to balance
workloads and achieve optimal performance.
ESX hostnames
Create a logical naming convention that allows you to quickly identify which site a host is in.
For example, the site can be included in the chosen naming convention or you can choose a
numbering system that reflects the location (for example odd hosts are in site one). The
naming convention makes the designing and the day-to-to running of your system easier.
126 IBM FlashSystem and VMware Implementation and Best Practices Guide
Logically name the LUNs with their home sites
This task is not a must and some might argue that they want the flexibility to move LUNs
between data centers, but it makes it easier for business as usual (BAU) staff to track which
are local data stores.
VMware vSphere host multipathing ensures that VMs that are running continue to operate
during various failure scenarios. Table 5-1 outlines the tested and supported failure scenarios
when using SVC or IBM FlashSystem Family HyperSwap function, and VMware vSphere
Metro Storage Cluster (vMSC).
Table 5-1 IBM Spectrum Virtualize HyperSwap and VMware vSphere Metro Storage Cluster supported failure scenarios
Failure scenario HyperSwap behavior VMware HA impact
Path failure: SVC or IBM Single path failure between SVC or IBM No impact.
FlashSystem Family Back-End FlashSystem Family control enclosure and
(BE) Port flash enclosure. No impact on HyperSwap.
Path failure: SVC or IBM Single path failure between SVC or IBM No impact.
FlashSystem Family Front-End FlashSystem Family control enclosure and
(FE) Port vSphere host. vSphere host uses
alternative paths.
Complete Site-1 failure (The SVC or IBM FlashSystem Family continues VMs running on vSphere hosts at the
failure includes all vSphere to provide access to all volumes through the failed site are impacted. VMware HA
hosts and SVC or IBM control enclosures at Site 2. When the automatically restarts them on
FlashSystem Family controllers control enclosures at Site-1 are restored, vSphere hosts at Site-2.
at Site-1) the volume copies are synchronized.
Complete site 2 failure Same behavior as a failure of Site-1. Same behavior as a failure of Site-1.
Multiple vSphere host failures, No impact. VMware HA continues to use the data
network disconnect store heartbeat to exchange cluster
heartbeats. No impact.
SVC or IBM FlashSystem SVC or IBM FlashSystem Family active vSphere hosts continue to access
Family inter-site link failure, quorum is used to prevent a split-brain volumes through the remaining I/O
vSphere cluster management scenario by coordinating one I/O group to group. No impact.
network failure remain servicing I/O to the volumes. The
other I/O group goes offline.
128 IBM FlashSystem and VMware Implementation and Best Practices Guide
6
Starting with IBM Spectrum Virtualize firmware 8.5.1.0 or later, the VASA Provider function
has been incorporated natively in to the configuration node of the cluster to simplify the
overall architecture of a vVol environment. This feature is referred to as the Embedded VASA
Provider.
130 IBM FlashSystem and VMware Implementation and Best Practices Guide
Platform name Supports Embedded VASA
Provider
6.1.2 Feature comparison between the Embedded VASA Provider and IBM
Spectrum Connect
In the initial release of the Embedded VASA Provider, there are several limitations that might
restrict functions when compared to existing vVol support that uses IBM Spectrum Connect.
Evaluate the requirements of your environment before selecting a VASA Provider.
Table 6-2 Feature comparison between the Embedded VASA Provider and IBM Spectrum Connect
Item IBM Spectrum Connect Embedded VASA Provider
The system must have a standard pool with storage capacity that is allocated.
Note: Data reduction pools (DRPs) are not supported for either the metadata volume disk
(VDisk) or individual vVols.
The system must be configured with a Network Time Protocol (NTP) server to ensure that
the date and time are consistent with the VMware infrastructure.
The system must be configured with a certificate with a defined Subject Alternative Name
value.
Note: Hosts that require access to the vVol data store must be configured with the vVol
host type.
132 IBM FlashSystem and VMware Implementation and Best Practices Guide
6.2.2 Configuring the NTP server
To configure the NTP server on the system, complete the following steps:
1. Go to the Settings → System window in the GUI, and select Date and time, as shown in
Figure 6-2.
Note: If you use an FQDN or DNS name for the NTP server, you must ensure that a DNS
server is configured in the system. To configure DNS servers, select Settings → Network
and select DNS.
When you use a self-signed certificate, you must update the Subject Alternative Name field in
the certificate before registering the Embedded VASA Provider within vCenter. When you use
a signed certificate, this value is likely defined.
2. Expand the details and review the Subject Alternative Name value, as shown in
Figure 6-4.
134 IBM FlashSystem and VMware Implementation and Best Practices Guide
3. Alternatively, run the lssystemcert command, which shows the following output:
IBM_2145:vvolsftw-sv1:superuser>lssystemcert | grep -A 1
"Subject Alternative Name"
X509v3 Subject Alternative Name:
IP Address:9.71.20.20
4. If no Subject Alternative Name field is defined, update the self-signed certificate. To do this
task, select Settings → Security and select Secure Communications, as shown in
Figure 6-5.
Note: If you use an FQDN or DNS name, you must ensure that a DNS server is configured
in the system. To configure DNS servers, select Settings → Network and select DNS.
During this step, the cluster IP is unavailable for a few minutes while the new security
settings are applied. After a few minutes, you might need to refresh your browser window,
and then you are prompted to accept the new self-signed certificate.
136 IBM FlashSystem and VMware Implementation and Best Practices Guide
6.2.4 Preparing Elastic Sky X integrated hosts for vVol connectivity
Any Elastic Sky X integrated (ESXi) hosts that require access to a vVol data store must be
defined as a vVol host type in the storage system.
Note: As a best practice, create a single host object in IBM Spectrum Virtualize to
represent each physical ESXi server in the configuration. When you use clustered host
environments, for example, when multiple ESXi hosts are part of a vSphere cluster, you
should use IBM Spectrum Virtualize Host Cluster objects.
3. Review the summary window, and click Make Host Cluster, as shown in Figure 6-10.
138 IBM FlashSystem and VMware Implementation and Best Practices Guide
4. When creating a host object, ensure that it is defined with the Host Type of vVol. To do this
task, access the Hosts view in the GUI by selecting Hosts → Hosts, and clicking Add
Host, as shown in Figure 6-11.
6. If the host object will be a member of a Host Cluster, expand the advanced view and select
the Host Cluster from the list, as shown in Figure 6-13 on page 141.
140 IBM FlashSystem and VMware Implementation and Best Practices Guide
Figure 6-13 Selecting the Host Cluster
Note: When creating the host object by using the CLI, use the host type adminlun:
IBM_2145:vvolsftw-sv1:superuser>mkhost -fcwwpn
2100000E1EC249F8:2100000E1EC249F9 -name vvolsftw-02 -hostcluster vvolsftw -type
adminlun
7. Repeat the process for each additional host that you want to create.
9. You can ensure consistency across all members of the host cluster by defining the host
type at the host cluster level. To do this task, select Hosts → Host Clusters. Right-click
the host cluster and select Modify Host Types, as shown in Figure 6-15.
10.Select the vVol host type and click Modify, as shown in Figure 6-16 on page 143.
142 IBM FlashSystem and VMware Implementation and Best Practices Guide
Figure 6-16 Clicking Modify
By configuring the vVol host type on the host or host cluster object, the system
automatically presents the Protocol Endpoints to the ESXi hosts.
11.Before continuing with the Embedded VASA Provider configuration, verify that all hosts in
the vSphere cluster correctly detected the Protocol Endpoints from the storage system. To
do this task, rescan the storage adapters on each ESXi host and verify that there is a
Small Computer System Interface (SCSI) device with SCSI ID 768 and 769, as shown in
Figure 6-17.
Note: A Protocol Endpoint is presented from each node in the IBM Spectrum Virtualize
cluster. Ensure that all ESXi hosts correctly identified all Protocol Endpoints (PEs) before
continuing.
After all the values are defined, the GUI creates all the necessary objects within
IBM Spectrum Virtualize to facilitate vVol support, as shown in Figure 6-19 on page 145.
144 IBM FlashSystem and VMware Implementation and Best Practices Guide
Figure 6-19 Required objects for vVol support are created
Note: Even though it has been allocated a maximum capacity of 2 TB, this volume is only
intended to store system metadata, so it is likely that the used capacity will never grow
beyond a few gigabytes in size.
Although the initial vVol configuration allows only for the creation of a single child pool, more
vVol-enabled child pools can be created later through the storage system CLI if increased
capacity or different tiers of storage are required.
Note: Swap vVols always are created as fully allocated volumes within IBM Spectrum
Virtualize regardless of the specified provisioning policy.
The storage credentials that are defined in this window are required when registering the
Storage Provider within vSphere, and they are initially used to authenticate the vSphere
environment against the IBM Spectrum Virtualize storage system.
The system automatically creates a user group that is assigned with a specific role within
IBM Spectrum Virtualize. Then, the user account that is specified here is created as a
member of this group and granted specific access rights to allow manipulation of vVol objects
within the storage system.
However, upon successful registration of the storage provider within vCenter, the password is
removed from the user account within IBM Spectrum Virtualize, and instead the vSphere
certificate is used to authenticate the user account.
146 IBM FlashSystem and VMware Implementation and Best Practices Guide
Note: The password that is defined in the window is used once, and it is required only in
the initial Storage Provider registration process.
Now, the Storage Provider can be reregistered in vCenter by using the new password.
3. Enter an identifiable name, and paste the URL into the URL field. Add the user credentials
that were defined earlier and click OK, as shown in Figure 6-22.
148 IBM FlashSystem and VMware Implementation and Best Practices Guide
4. You might see an error or warning saying that the operation failed, but this message is
related only to the initial certificate thumbprint warning, so it can be ignored (Figure 6-23).
5. Verify that the newly added Storage Provider is showing online and active in the Storage
Providers list (Figure 6-24).
Figure 6-24 Newly added Storage Provider is showing online and active
150 IBM FlashSystem and VMware Implementation and Best Practices Guide
3. Select the backing storage container in the list and define the name of the new vVol data
store. Click NEXT (Figure 6-27).
Figure 6-27 Defining the name of the new vVol data store
4. Select the hosts that will access the vVol data store and click NEXT (Figure 6-28).
6. Review the Datastores tab and ensure that the capacity and accessibility are correctly
reported, as shown in Figure 6-30.
152 IBM FlashSystem and VMware Implementation and Best Practices Guide
6.3.7 Provisioning more vVol data stores
In this section, we cover the steps for provisioning more vVol data stores.
Identify the target parent pool in which to create the child pool and note the mdiskgrp ID or
name.
By default, the name that is associated with the VASA ownership group is VASA.
Identify the provisioning policy that is required for the new vVol child pool by running the
lsprovisioningpolicy command, as shown in Example 6-2.
To perform a migration between vVol solutions, you must complete the following steps:
1. Provision Virtual Machine File System (VMFS) data stores with sufficient capacity to store
all VMs that are on the vVol storage.
2. Use Storage vMotion to migrate existing VMs or templates from vVol storage to VMFS
data stores.
3. Remove the vVol IBM Spectrum Connect configuration from vCenter.
4. Disable or decommission the vVol function on the storage system.
5. Enable vVol by using the Embedded VASA Provider and create vVol data stores.
6. Use Storage vMotion to migrate VMs and templates from VMFS storage to the new vVol
data stores.
7. Remove the temporary VMFS data stores if they are no longer needed.
After the new volumes are created and mapped to the appropriate hosts or host cluster in the
storage system, create the VMFS data stores.
Identify the VMs that are on the vVol data stores that are presented by IBM Spectrum
Connect and perform a storage migration to move them to the new VMFS data stores.
Depending on the number of VMs, consider using, for example, PowerCLI to automate bulk
VM migrations.
Note: VM templates might exist on the vVol data stores that require conversion into a VM
before they can be migrated. After the migration completes, they can be converted back
into a VM template.
154 IBM FlashSystem and VMware Implementation and Best Practices Guide
6.4.3 Removing the vVol IBM Spectrum Connect configuration from vCenter
In this section, we describe removing the vVol IBM Spectrum Connect configuration from
vCenter.
3. After the data store is unmounted from all hosts, it automatically is removed from vCenter.
4. Repeat these steps for all the vVol data stores that are presented by IBM Spectrum
Connect.
156 IBM FlashSystem and VMware Implementation and Best Practices Guide
Figure 6-33 Policies and Profiles view
2. Select VM Storage Policies and identify any policies that were created by using
IBM Spectrum Connect.
3. Select the policy to remove, and click DELETE, as shown in Figure 6-34.
4. Repeat these steps for any remaining policies that are associated with IBM Spectrum
Connect.
6.5.1 Identifying and removing the vVol child pools for IBM Spectrum Connect
Identify the child pools that were allocated to any vVol Storage Spaces within the
IBM Spectrum Connect GUI and delete them, as shown in Figure 6-36 on page 159.
158 IBM FlashSystem and VMware Implementation and Best Practices Guide
Figure 6-36 Deleting the child pools that were allocated to any vVol Storage Spaces
Alternatively, log on to the storage system CLI with a user account that has the VASA
Provider role and run the following command to identify any existing vVol child pools that are
used by IBM Spectrum Connect:
lsmdiskgrp -filtervalue owner_type=vvol_child_pool
Note the mdiskgrp name or ID, and then verify that the vVol pool is no longer required and
that the name and ID are correct because there is no way to recover the pool after it is
deleted. Once you are sure, run the following command to remove the child pool:
rmmdiskgrp <name or id>
Warning: Removing the pool might fail if any vVols are in the pool. You might need to
manually remove any vVols in the pool before removing the pool itself. To identify any vVols
that are in the pool to be deleted, run the following command:
lsvdisk -filtervalue mdisk_grp_name=<child pool name>
For each vVol, identify the VDisk ID or name and run the following command to delete the
vVol.
Warning: Verify that the vVol is no longer required and that the name and ID are correct
because there is no way to recover the data after the volume is deleted.
After any lingering vVols are deleted, retry the pool removal command until all IBM Spectrum
Connect vVol pools are removed.
Warning: If other integration interfaces are configured, for example, vCenter or vRealize
Orchestration, do not remove the user account because its removal will cause future
integration commands to fail.
Identify the user account that is used by IBM Spectrum Connect by either reviewing the
Storage System Credentials window in the IBM Spectrum Connect GUI or by using the CLI.
Using the GUI, select Menu → Storage credentials, as shown in Figure 6-37.
You see the user account that is used by IBM Spectrum Connect, as shown in Figure 6-38.
You also can use the command in Example 6-4 on page 161 on the storage system CLI.
160 IBM FlashSystem and VMware Implementation and Best Practices Guide
Example 6-4 Identifying the user account that is used by IBM Spectrum Connect by using the command-line interface
IBM_2145:vvolsftw-sv1:superuser>lsuser
id name password ssh_key remote usergrp_id usergrp_name owner_id owner_name locked password_change_required
0 superuser yes yes no 0 SecurityAdmin no no
1 vasa_server yes no no 6 VASAUsers no no
Remove the user account that is used by IBM Spectrum Connect, and then run the following
command:
rmuser <user_id or name>
To identify the User Group that is associated with the VASA Provider role, run the command
in Example 6-5 on the storage system CLI.
Assuming no other user accounts are in the user group, remove the VASA Provider user
group by running the following command:
rmusergrp <usergrp_id or name>
To identify the location of the metadata VDisk, run the command in Example 6-6 on the
storage system CLI.
Remove the metadata VDisk by running the following command on the storage system CLI.
Warning: The metadata VDisk contains all metadata that is associated with the vVol
environment. This operation cannot be undone.
rmmetadatavdisk
After the new vVol data store is online, do the migration by completing the following steps:
1. Identify the VMs that will be migrated, select them, right-click them, and click Migrate, as
shown in Figure 6-39.
2. In the Select Storage window, identify the newly created vVol data store and click NEXT,
as shown in Figure 6-40 on page 163.
162 IBM FlashSystem and VMware Implementation and Best Practices Guide
Figure 6-40 Identifying the newly created vVol data store
3. Complete the Storage vMotion workflow and review the tasks to ensure that the VMs
successfully migrated (Figure 6-41).
Figure 6-41 Reviewing the tasks to ensure that the VMs successfully migrated
During the migration operation, vVols automatically are created on the storage system
within the child pool that was configured as a vVol storage container.
164 IBM FlashSystem and VMware Implementation and Best Practices Guide
5. Select the vVol-enabled child pool in the list by identifying the vVol logo underneath the
pool information on the left side. When you select the vVol pool, the individual vVol objects
appear, as shown in Figure 6-43.
Note: By default, the Name column is not displayed in the GUI table view, but it is an
optional field that can be added by right-clicking the column headers and selecting the
Name checkbox.
166 IBM FlashSystem and VMware Implementation and Best Practices Guide
7
IBM Storage Insights is an IBM Cloud® software as a service (SaaS) offering that can help
you monitor and optimize the storage resources in the system and across your data center.
IBM Storage Insights provides cognitive support capabilities, monitoring, and reporting for
storage systems, switches and fabrics, and VMware Elastic Sky X integrated (ESXi) hosts in
a single dashboard.
Monitoring Inventory management IBM block storage, IBM and non IBM block
switches, fabrics, storage, file storage,
and VMware ESXi object storage, switches,
hosts fabrics, and VMware ESXi
hosts
168 IBM FlashSystem and VMware Implementation and Best Practices Guide
Resource Functions IBM Storage IBM Storage Insights
management Insights Pro (subscription)
Drill-down performance
workflows to enable
deep troubleshooting
Explore virtualization
relationships.
Explore replication
relationships.
Exporting performance
data to a file
Hassle-free log
collection
Simplified ticketing
Performance planning
Capacity planning
Business impact
analysis (applications,
departments, and
groups)
Optimize data
placement with tiering.
Restriction: You must have a current warranty or maintenance agreement for an IBM
block storage system to open tickets and send log packages.
As an on-premises application, IBM Spectrum Control does not send the metadata about
monitored devices off-site, which is ideal for dark shops and sites that do not want to open
ports to the cloud. However, if your organization allows for communication between its
network and the cloud, you can use IBM Storage Insights for IBM Spectrum Control to
transform your support experience for IBM block storage.
170 IBM FlashSystem and VMware Implementation and Best Practices Guide
IBM Storage Insights for IBM Spectrum Control is like IBM Storage Insights Pro in capability,
and it is available for no additional cost if you have an active license with a current
subscription and support agreement for IBM Virtual Storage Center, IBM Spectrum Storage
Suite, or any edition of IBM Spectrum Control.
The data collector streams performance, capacity, asset, and configuration metadata to your
IBM Cloud instance.
The metadata flows in one direction, that is, from your data center to IBM Cloud over HTTPS.
In the IBM Cloud, your metadata is protected by physical, organizational, access, and security
controls. IBM Storage Insights is ISO/IEC 27001 Information Security Management certified.
Figure 7-1 shows the architecture of the IBM Storage Insights application, the supported
products, and the three main teams who can benefit from using the tool.
For more information about IBM Storage Insights and to sign up and register for the
no-charge service, see the following resources:
Fact sheet
Demonstration
Security guide
Figure 7-2 IBM Storage Insights System overview for block storage
The block storage dashboard is a default one that is shown when you go to the Operations
dashboard. To take a close look at the storage systems that are being monitored, select
Dashboards → Operations. Then, click the storage system in which you are interested in
the list at the left of the dashboard. For example, you can see the health, capacity, and
performance information for a block storage system in the Operations dashboard. The
storage system is colored red because there are problems with nodes and logical
components.
IBM Storage Insights supports Brocade and Cisco switches and fabrics so that you can detect
and investigate performance issues throughout your storage environment. You can follow the
trail of storage requests through the components in the SAN fabric to the target storage
systems.
For more information about IBM Storage Insights, see Key features on IBM Storage Insights
Documentation.
172 IBM FlashSystem and VMware Implementation and Best Practices Guide
7.4 IBM Storage Insights VMware integration
You can add VMware ESXi hosts and their VMs for monitoring with IBM Storage Insights by
specifying connection information for a vCenter Server. When you add VMware ESXi hosts,
then you can collect data, generate reports, and manage storage that is related to hosts and
VMs.
Identify the IP addresses and user credentials of the vCenter Servers that manage the
VMware ESXi hosts and VMs that you want to monitor by completing the following steps.
IBM Storage Insights uses this information to connect to the vCenter Servers and discover
their managed VMware ESXi hosts and VMs. You can add multiple vCenter Servers
concurrently if they share the same user credentials.
1. In the menu bar, select Resources → Hosts.
2. Click Add vCenter Server.
3. Enter the IP addresses or hostnames that you use to connect to the vCenter Server, and
then enter the authentication credentials such as username and password that are shared
by all the vCenter Servers that you are adding, as shown in Figure 7-3.
After the VCenter server initial discovery starts, all ESXi hosts and VMs that are managed by
VCenter are discovered by IBM Storage Insights. You can see details of each ESXi host on
the IBM Storage Insights Pro edition, as shown in Figure 7-4.
Figure 7-4 ESXi host details with daily response time, IO rate, and most active volumes on IBM Storage Insights
End-to-end SAN connectivity is visible when IBM Storage systems, SAN switches, and
VMware ESXi servers are added to IBM Storage Insights.
174 IBM FlashSystem and VMware Implementation and Best Practices Guide
Most of the time, it is a complex process to find which VM creates heavy I/O on a data store
volume. IBM Storage Insights Pro can monitor virtual machine disk (VMDK) level I/O
performance, and you can find the heavy loaded VM, as shown in Figure 7-5.
Figure 7-5 VMDK level performance monitoring on IBM Storage Insights Pro
Chapter 8. Troubleshooting
This chapter provides information to troubleshoot common problems that can occur in an IBM
FlashSystem and VMware Elastic Sky X integrated (ESXi) environment. It also explains how
to collect the necessary problem determination data.
8.1.1 Data collection guidelines for SAN Volume Controller and IBM
FlashSystem
On SAN Volume Controller (SVC) and IBM FlashSystem, system logs can be collected in the
product GUI by selecting Settings → Support Package (Figure 8-1).
For more information about the level of logs to collect for various issues, see What Data
Should You Collect for a Problem on IBM Spectrum Virtualize systems?
For the topics covered in the scope of this document, you typically need to gather a snap
(option) 4, which contains standard logs plus new statesaves. Because this data often takes a
long time to collect, it might be advantageous to manually create the statesaves, and then
collect the standard logs afterward. This task can be done by using the svc_livedump
command-line interface (CLI) utility, which is available in the product command-line interface
(Example 8-1 on page 179).
178 IBM FlashSystem and VMware Implementation and Best Practices Guide
Example 8-1 Using svc_livedump to manually generate statesaves
IBM_FlashSystem:Cluster_9.42.162.160:superuser>svc_livedump -nodes all -y
Livedump - Fetching Node Configuration
Livedump - Checking for dependent VDisks
Livedump - Check Node status
Livedump - Preparing specified nodes - this may take some time...
Livedump - Prepare node 1
Livedump - Prepare node 2
Livedump - Trigger specified nodes
Livedump - Triggering livedump on node 1
Livedump - Triggering livedump on node 2
Livedump - Waiting for livedumps to complete dumping on nodes 1,2
Livedump - Waiting for livedumps to complete dumping on nodes 2
Livedump - Successfully captured livedumps on nodes 1,2
After you generate the necessary statesaves, collect standard logs and the latest statesaves
(option 3), and use the GUI to create a support package including the manually generated
livedumps. Alternatively, you can create the support package by using the CLI (Example 8-2).
When the support package is generated by using the command line, you can download it by
using the GUI or using a Secure Copy Protocol (SCP) client.
When downloading a package for an ESXi host, the default settings provide the information
that is needed to analyze most problems.
Important: When collecting data for problems that are related to SRM, make sure to
collect data from all sites associated with the problem.
8.1.4 Data collection guidelines for IBM Spectrum Connect (VASA or vVols)
For troubleshooting issues associated with VASA or VMware vSphere Virtual Volume (vVol),
the following sets of data are required for troubleshooting:
1. A support package from the storage system as shown in 8.1.1, “Data collection guidelines
for SAN Volume Controller and IBM FlashSystem” on page 178.
2. A support package from IBM Spectrum Connect.
3. A support package from the management application interfacing with IBM Spectrum
Connect.
4. If the problem includes access to the data, ESXi logs as shown in 8.1.2, “Data collection
guidelines for VMware ESXi” on page 179.
180 IBM FlashSystem and VMware Implementation and Best Practices Guide
Collecting data for VMware vCenter
vCenter logs can be collected by using the same process as ESXi hosts, as described in
8.1.2, “Data collection guidelines for VMware ESXi” on page 179. The difference is when
selecting resources for which to collect logs, select the vCenter server instead of (or in
addition to) an ESXi host.
The three general categories of storage loss of access events in VMware products are:
All paths down (APD)
Permanent device loss (PDL)
Virtual machine (VM) crash
These issues are typically the result of errors in path recovery. Corrective actions include:
Validating the best practice multipathing configuration is in use as shown in 2.3,
“Multi-path considerations” on page 18.
Validate all server driver and firmware levels are at the latest supported level.
Validate the network infrastructure connecting the host and storage is operating correctly.
For a list of I/O errors that trigger PDL, see Permanent Device Loss (PDL) and
All-Paths-Down (APD) in vSphere 6.x and 7.x (2004684).
These types of events are often the result of a hardware failure or low-level protocol error in
the server host bus adapter (HBA), storage area network (SAN), or the storage array. If
hardware errors are found that match the time in which the PDL happens, PDF is likely the
cause.
182 IBM FlashSystem and VMware Implementation and Best Practices Guide
8.2.2 VMware migration task failures
Two types of migrate tasks are as follows:
vMotion is a migrate task that is used to move the running state of the VM (for example,
memory and compute resource) between ESXi hosts.
Storage vMotion is a migrate task that is used to move the VM storage resources between
data stores, for example VMDK files data stores.
vMotion tasks
vMotion tasks are largely dependent on the Ethernet infrastructure between the ESXi hosts.
The only real storage interaction is at the end when file locks must move from one host to
another. In this phase, it is possible for Small Computer System Interface (SCSI) Reservation
Conflicts or file lock contention to result in the failing of the migration task. The following
articles describe the most frequent issues:
Investigating virtual machine file locks on ESXi hosts (10051)
Resolving SCSI reservation conflicts (1002293)
IBM Spectrum Virtualize APAR HU01894
The default timeout for the task to complete is 100 seconds. If the migration takes longer than
100 seconds to complete, then the task fails with a timeout, as shown in Example 8-6.
The task is generic by nature and the root cause behind the timeout typically requires
performance analysis of the storage arrays that are involved and a detailed review of the ESXi
logs for the host performing the task. In some circumstances, it might be appropriate to
increase the default timeout, as described at Using Storage Motion to migrate a virtual
machine with many disks fails without timeout (1010045).
186 IBM FlashSystem and VMware Implementation and Best Practices Guide
Related publications
The publications that are listed in this section are considered suitable for a more detailed
description of the topics that are covered in this book.
IBM Redbooks
The following IBM Redbooks publications provide more information about the topics in this
document. Some publications that are referenced in this list might be available in softcopy
only.
IBM FlashSystem Best Practices and Performance Guidelines, SG24-8503
IBM Spectrum Virtualize HyperSwap SAN Implementation and Design Best Practices,
REDP-5597
IBM Storage and the NVM Express Revolution, REDP-5437
Implementing IBM FlashSystem with IBM Spectrum Virtualize V8.4, SG24-8492
Introduction and Implementation of Data Reduction Pools and Deduplication, SG24-8430
You can search for, view, download, or order these documents and other Redbooks,
Redpapers, web docs, drafts, and additional materials, at the following website:
ibm.com/redbooks
Online resources
These websites are also relevant as further information sources:
Brocade Peer Zoning:
https://docs.broadcom.com/doc/FOS-90x-Admin-AG
Cisco MDS 9000 NX-OS Configuration Guide:
https://www.cisco.com/c/en/us/td/docs/switches/datacenter/mds9000/sw/5_2/config
uration/guides/fabric/nx-os/nx_os_fabric/zone.html
Data collection instructions for the VMware vRealize Operations Manager:
https://kb.vmware.com/s/article/2074601
Data collection instructions for the VMware vRealize Orchestrator:
https://kb.vmware.com/s/article/2150664
Frequently asked questions about Storage Distributed Resource Scheduler (DRS):
https://kb.vmware.com/s/article/2149938
IBM Documentation for IBM FlashSystem 9100 and 9200:
https://core.vmware.com/reference-architectures
IBM FlashSystem v8.4 Support Matrix:
https://www.ibm.com/support/pages/node/3543675
188 IBM FlashSystem and VMware Implementation and Best Practices Guide
IBM FlashSystem and VMware Implementation and Best Practices Guide
(0.2”spine)
0.17”<->0.473”
90<->249 pages
Back cover
SG24-8505-01
ISBN 0738460869
Printed in U.S.A.
®
ibm.com/redbooks