emc® vplex™ san connectivity · d best pract n 02 ights reserve lication is ac out notice. vided...
TRANSCRIPT
Implementat
EMC®Implem
Fro ESX VPL
Xtre
tion Planning
® VPLentationnt-end CoXi Path LoLEX Back
emIO Con
Abs
ThisVPL
Dec
Ver
g and Best Pr
LEX™ n Plannionnectivoss Hand-end Arra
nnectivity
stract
s document dLEX to storage
cember 2014
sion 02
ractices Tech
SAN ng and ity
dling ay Conne
y
describes SAe array back-
4
nical Notes
Conn Best Pra
ectivity
AN connectivi-end.
ectiviactices
ity for both H
ty
Host to VPLEXX front-end and
2 EMC VPLEX SImplementat
SAN Connecttion Planning
Cop
Pub
EMCThe
Thereppubfitndes
EMCCorher
For Trad
EMCImp
Part
ivity g and Best Pr
pyright © 201
blished Dec.
C believes the information
e informationresentations
blication, andess for a par
scribed in thi
C2, EMC, andrporation in tein are the p
the most updemarks on E
C VPLEX SANplementation
t Number H1
ractices Tech
14 EMC Corp
2014
he informatio is subject to
in this publis or warrantied specificallyticular purpos publication
d the EMC loghe United Stroperty of th
-to-date listin EMC.com.
N Connectivityn Planning an
3546 Versio
nical Notes
poration. All r
n in this pubo change with
ication is proes of any kindy disclaims imose. Use, copn requires an
go are registeates and otheir respectiv
ng of EMC pr
y nd Best Pract
n 02
rights reserve
blication is achout notice.
ovided as is. d with respecmplied warrapying, and disn applicable s
ered trademaher countries.e owners.
roduct names
tices
ed. Published
ccurate as of
EMC Corporact to the info
anties of mercstribution of
software lice
arks or tradem. All other tra
s, see EMC C
d in the USA.
f its publicati
ation makes rmation in thchantability any EMC sofnse.
marks of EMCademarks use
Corporation
.
ion date.
no his or ftware
C ed
ChaAud
Doc
SAN
ChaFron
Hos
VBL
ChaPat
P
v
v
A
vSp
D
C
PDL
ChaBac
A
A
S
A
ChaXtre
VPL
S
S
S
D
D
apter 1 Indience ..........
cument Orga
N Connectivit
apter 2 Fntend/Host I
st Cross-Clus
LOCK and VP
apter 3 Eh loss handl
Persistent de
vSphere Stor
vSphere Stor
APD Handling
phere Storage
Disable Stora
Change Time
L/APD Refere
apter 4 Vck-end/Stora
Active/Active
Active/Passiv
Significant B
Additional Ar
apter 5 XemIO Overvie
LEX Connectiv
Single Engine
Single Engine
Single Engine
Dual Engine
Dual Engine
Imp
ntroduction....................
nization .......
ty Overview ..
Frontend Co nitiator Port
ter Connect .
LEX Front End
ESXi Path Loing semantic
evice loss (PD
rage for vSph
rage for vSph
g ..................
e for vSphere
age APD Han
eout Limits fo
ences ............
VPLEX Backege Array Con
e Arrays ........
ve and ALUA
it .................
rray Conside
XtremIO Conew ................
vity to Xtrem
e / Single X-B
e / Dual X-Br
e / Quad X-B
/ Single X-Br
/ Dual X-Bric
plementation
n ....................
....................
....................
nnectivity Connectivity
....................
d Connectivit
oss Handlingcs (PDL and A
DL) ...............
here 5.0 U1 ..
here 5.5 ........
....................
e 5.1 ............
dling ...........
or Storage AP
....................
end Array Connectivity ......
....................
A Arrays .........
....................
rations .........
nnectivity ....................
IO ................
Brick ............
rick...............
Brick .............
rick ..............
ck .................
n Planning an
....................
....................
....................
y ...................
....................
ty Rules .......
g APD) .............
....................
....................
....................
....................
....................
....................
PD ................
....................
onnectivity ....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
EMC VPLnd Best Pract
....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
y ....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
LEX SAN Contices Technic
Conte
....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
Contents
nectivity cal Notes
ents
5 ............ 6
............ 6
............ 7
8 ............ 9
.......... 13
.......... 14
16 .......... 17
.......... 17
.......... 18
.......... 18
.......... 19
.......... 20
.......... 20
.......... 20
.......... 21
22 .......... 23
.......... 23
.......... 27
.......... 32
.......... 33
35 .......... 36
.......... 37
.......... 38
.......... 39
.......... 40
.......... 41
.......... 42
3
4
Contents
EMC VPLEX SImplementat
SAN Connecttion Planning
D
Q
Q
Q
ivity g and Best Pr
Dual Engine
Quad Engine
Quad Engine
Quad Engine
ractices Tech
/ Quad X-Bri
e / Single X-B
e / Dual X-Bri
e / Quad X-Br
nical Notes
ck ................
Brick .............
ck ................
rick...............
....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
....................
.......... 43
.......... 44
.......... 45
.......... 46
This
Aud
Doc
SAN
s chapter pre
dience .........
cument Orga
N Connectivit
Imp
esents the fo
....................
nization ......
ty Overview .
plementation
llowing topic
....................
....................
....................
n Planning an
Chapte
cs:
...................
...................
...................
C
EMC VPLnd Best Pract
er 1 In
....................
....................
....................
Chapter 1: Intr
LEX SAN Contices Technic
ntroduc
...................
...................
...................
roduction
nectivity cal Notes
ction
........... 6
........... 6
........... 7
5
6
A
Chapter 1: Int
EMC VPLEX SImplementat
Audience
Documen
roduction
SAN Connecttion Planning
e
Thewilltech
nt Organ
ThisImpdocper
All d
Themad
ivity g and Best Pr
ese technical be configurihnical notes
SAN te
Fiber C
VPLEX
ization
s technical nplementationcument that h-topic inform
documents a
e following lisde available:
EMC VPLE
EMC VPLE
EMC VPLE
EMC VPLE
EMC VPLE
EMC VPLE
EMC VPLE
EMC VPLE
EMC VPLE
ractices Tech
notes are foing, installing requires an u
echnology an
Channel bloc
concepts an
ote is one ofn Planning an had previousmation that w
are available
st represents:
EX Overview a
EX SAN Conne
EX Host Multi
EX IP Network
EX and Recov
EX Data Mobi
EX REST API
EX Upgrades
EX Virtual Edi
nical Notes
r EMC field pg, and suppo understandin
nd network de
k storage co
nd componen
f a set of docnd Best Practsly been avaiill be easier t
on http://su
s the technica
and General
ectivity
pathing
king
verPoint Inter
lity and Migr
tion
personnel anorting VPLEX.ng of the foll
esign
ncepts
nts
uments that tices for EMClable. It is in
to maintain a
upport.emc.c
al note best p
Best Practice
roperability
rations
d partners an. An understaowing:
supersede t VPLEX Techntended to pr
and keep up-
com.
practice docu
es
nd customeranding of the
he monolithinical Notes rovide more c-to-date.
uments that
s who ese
ic
concise,
will be
SAN Con
nectivity
ThisVPLarra
Hosreqreqheathe envenvGeo
y Overvie
s document iLEX Virtual Voays.
st connectivituirements wuirements ar
alth checks a system from
vironments ovironments. SoSynchrony C
Imp
ew
is designed tolumes and V
ty to VPLEX, rhich differ frore based on and ndu prech
m being upgrar Proof of Con
System limitCode level.
plementation
to address fro VPLEX backe
referred to asom the backe
architectural hecks. Connaded. Bypasncept but it is are availab
n Planning an
ontend SAN nd SAN conn
s frontend coend or VPLEX design limit
nectivity violasing these re
is not supporble in the Rele
C
EMC VPLnd Best Pract
connectivity nectivity for a
onnectivity, hX to array conations and a
ations will beequirements rted for prodease Notes f
Chapter 1: Intr
LEX SAN Contices Technic
for host acceaccess to sto
has very specnnectivity. Thare checked de flagged and is allowed fouction
for each
roduction
nectivity cal Notes
ess to rage
cific hese during d prevent or test
7
8
Chapter 2: Fro
EMC VPLEX SImplementat
ontend Connec
SAN Connecttion Planning
This
Fron
Hos
VBL
ctivity
ivity g and Best Pr
s chapter pre
ntend/Host I
st Cross-Clus
LOCK and VP
ractices Tech
C
esents the fo
Initiator Port
ster Connect.
LEX Front En
nical Notes
hapter 2
llowing topic
t Connectivity
....................
d Connectivi
2 Fro
cs:
y .................
...................
ity Rules ......
ntend C
....................
....................
....................
Connecti
...................
...................
...................
ivity
........... 9
......... 13
......... 14
Frontend
d/Host In
Noteduaaddno bmulto anewQuamaythe
Com
UseA0-
nitiator Po
Dual fabri
The front-physical c
Each hostdirector o
Multipaththe dual fa
Each hostLUN from
Four paths
Dual and connectiv
Observe Ddirectory C
GUI Perfor
Skipping tthan four
o
e: For cluster ual to a quad en
ed engines. A benefit. The Nlti engine VPLE dual, the ndu
w engine whichad upgrade wiy choose to reb pair of new en
mplete physi
e the same FEFC00. Please
Imp
ort Conn
c designs are
end I/O modconnections o
t should haven each fabric
ing or path faabrics
t should have a minimum o
s are require
Quad enginevity on each f
Director CPU CPU utilizatio
rmance Dash
the NDU pre- paths and is
o This confPOC
upgrades whengine cluster y Adding additioNDU pre-checkEX cluster as au pre-check mah is why the holl not flag an isbalance the wngines.
cal connectio
E/BE ports one refer to har
plementation
ectivity
e considered
dules on each one to each f
e at least onec for a total o
ailover softw
e fabric zonin of an A and B
d for NDU
e VPLEX Clustabric
utilization anon is below 5
hboard in Geo
-checks woul not conside
figuration wo
en going from aou must rebal
onal engines ak will flag host a configurationay have passeost paths musssue provided
workload acros
ons to the VP
n each directrdware diagra
n Planning an
d a best pract
h director sho fabric (requir
e path to an Aof four logical
ware is requir
ng that proviB director fro
ters require s
nd schedule 50%
oSynchrony
ld be requireered a best pr
ould be supp
a single enginlance the hostand then not c connectivity tn issue. Whened initially butst be rebalancd there were nss the new eng
PLEX before c
tor to avoid cams for port
Chapter 2:
EMC VPLnd Best Pract
tice
ould have a mred)
A director anl paths (requ
red at the hos
des redundam each fabri
spanning eng
NDU for time
5.1 or newer
ed for host coractice
orted for a te
ne to a dual ent connectivity
connecting hosthat does not scaling up a s
t will fail after ed across boto issues priorgines or add a
commissioni
confusion, th layout.
: Frontend Con
LEX SAN Contices Technic
minimum of
nd one path tuired for NDU
st for access
ant access to c.
gines for hos
es when aver
r
onnectivity w
est environm
ngine cluster o across the nest paths to the span engines single engine the addition oth engines. Dr to the upgradadditional hos
ng/setup.
at is, B0-FC0
nnectivity
nectivity cal Notes
two
o a B ).
across
each
st
rage
ith less
ent or
or from a wly em is of in a cluster
of the ual to
de. You ts to
00 and
9
10
Chapter 2: Fro
EMC VPLEX SImplementat
ontend Connec
SAN Connecttion Planning
Figche
Thistwomin
Plea
Note
ctivity
ivity g and Best Pr
gure 1 Host eck requirem
s illustration o VPLEX direcnimum config
ase refer to t
e: Each Initiat
ractices Tech
connectivitments
shows dual ctors on the sguration that
he Release N
or / Target con
nical Notes
ty for Sing
HBAs connecsame engine would meet
Notes for the
nnection is ca
gle Engine C
cted to two F in the single NDU require
total FE port
lled an IT Nex
Cluster mee
Fabrics with ee engine clusements.
IT Nexus lim
xus.
eting NDU pr
each connectter. This is t
mit.
re-
ting to he
Figqua
TheVPLNDUClusiden
gure 2 Host ad engine
e previous illuLEX directorsU pre-checksster s. This cntified in the
Imp
connectivit
ustration sho. This config
s. This confi configuratione Release Not
plementation
ty for HA r
ows host conguration offeriguration app
n still only cotes for that ve
n Planning an
requirements
nectivity witrs increased plies to both
ounts as four ersion of Geo
Chapter 2:
EMC VPLnd Best Pract
s for NDU p
h dual HBAs levels of HA the Dual and IT Nexus agaoSynchrony.
: Frontend Con
LEX SAN Contices Technic
pre-checks d
connected t as required bd Quad VPLEainst the limi
nnectivity
1nectivity cal Notes
dual or
o four by the EX it as
11
12
Chapter 2: Fro
EMC VPLEX SImplementat
ontend Connec
SAN Connecttion Planning
Fig
TheVPLagaGeopatdire
Noterecoengbala
ctivity
ivity g and Best Pr
gure 3 Host
e previous illuLEX engines (ainst the totaoSynchrony. h to all availa
ector for the a
e: Most host commendationsines and subsancing the I/O
ractices Tech
connectivit
ustration sho (eight directo
l limit as def Hosts using able director
active path w
connectivity fos for a dual ensequent hosts
O across all eng
nical Notes
ty for HA q
ows host conors). This confined in the R active/pass
rs and manuawith different
or hosts runninngine cluster. should alterngines.
quad engine
nectivity witnfiguration c
Release Noteive path failoal load balant hosts.
ng load balanc The hosts shonate between p
h dual HBAs ounts as eigs for that verover softwarence by select
cing software ould be config pairs of engin
connected tht IT Nexusesrsion of e should coning a differen
should followgured across twes effectively
o four s
nect a nt
w the wo load
Host Cro
oss-Cluste
Fig
er Conne
gure 4 Host
PowerPathenvironme
Host crossconfigurat
Host initia
Host multactive patmultipath
Separate merging o
Connectiv
Supported(Please re
Host crossthan 1ms
Host cross
VPLEX Wit
Host crossonly
Host crossenvironme
At least onconnectioeach VPLE
All Consisrequired t
Imp
ect
Cluster con
h VE provideent
s-cluster contions as liste
ators are zon
ipathing softth going to thing driver to
HBA ports shof the local an
vity at both si
d stretch clusefer to VPLEX
s-cluster con latency
s-cluster con
tness works w
s-cluster con
s-cluster conent only
ne backend son to the VPLEX cluster
stency Group to have the a
plementation
nnected acr
s an auto sta
nect appliesed in the VPLE
ed to both V
tware can behe local VPLE prefer all loc
hould be usend remote fa
ites follow sa
sters can be ESSM)
nect is limite
nect require
with Consiste
nect must be
nect is supp
storage arrayEX cluster at
s used in a huto-resume a
n Planning an
ross sites t
andby feature
s to specific hEX ESSM onl
PLEX clusters
e configured fEX cluster. Wcal cluster pa
d for the rembrics
ame rules as
configured u
ed to a VPLEX
s the use of V
ency Groups
e configured
ported in a VP
y is required that site. Ar
host cross-clu attribute set
Chapter 2:
EMC VPLnd Best Pract
to both VPL
e created spe
host OS and y.
s in a Metro.
for active patWhen feasibleaths over rem
mote cluster c
single host c
using host cro
X cluster sep
VPLEX Witne
only
using VPLEX
PLEX Metro sy
at each site wrrays are not
uster connec to true
: Frontend Con
LEX SAN Contices Technic
LEX Clusters
ecifically for
multipathing
th/passive pe, configure tmote cluster p
connection to
connectivity
oss-cluster c
paration of no
ss
X Distributed
ynchronous
with redunda cross conne
ct configurati
nnectivity
1nectivity cal Notes
s
this
g
path with he
paths.
o avoid
onnect
o more
Devices
ant cted to
on are
13
14
V
Chapter 2: Fro
EMC VPLEX SImplementat
VBLOCK
ontend Connec
SAN Connecttion Planning
Theacccanusehos
If usaut
#po
Powrem
PP glate
and VPL
Notedepsug
1
2
All Epatt
Engin
Direc
Clust
Engin
Direc
Clust
Engin
Direc
Clust
Dual
Quad
Singl
ctivity
ivity g and Best Pr
e unique soluess to both d
n be achievede of backbonest cross-clust
sing PowerPaostandby fea
owermt set a
werPath will tmote/non-pre
groups the pency is design
EX Front
e: All rules in pending on cusgested rule.
. Physical
a. Eacpor
i.
On
Co
ii.
Po
Po
Co
. ESX Clus
ESX clusters tterns:
ne #
ctor A
ter # 1,2,3,4,5,
ne #
ctor A
ter # 1,3,5,
ne #
ctor A
ter# 1,5
Engine
d Engine
e Engine
ractices Tech
ution provide datacenters. d using an exe fabrics mayter connect.
ath VE, the oature:
utostandby=
take care of seferred VPLEX
aths by VPLEnated as the
End Conn
BOLD cannot bstomer require
l FE connecti
ch VPLEX Direrts connect to
For single V
nly ports 0 an
onnect even V
For two VBL
orts 0 and 1 w
orts 2 and 3 u
onnect even V
ster Balancin
are evenly d
B
,6,7,8 1,2,3,4,5,6,7,8
B
,7 2,4,6,8
B
2,6
Engine 1
Engine 1
Engine 1
nical Notes
d by Host cro The latency xtended fabriy introduce a
The rtt must
nly thing tha
=on trigger=p
setting to autX cluster.
EX cluster and local/prefer
nectivity
be broken, hoement, but if t
vity
ector has 4 fro fabric A and
VBLOCKS con
nd 1 will be u
VPLEX front e
LOCKS conne
will be used
used for VBL
VPLEX front e
ng across VPL
distributed ac
8
A
2,4,6,8
A
3,7
Engine
Engine
oss-cluster c requirementic or fabrics tadditional lat be within 1m
at the custom
prox host=xxx
tostandby th
d the one witrred cluster.
y Rules
owever Rules i these are gene
ront end portd odd ports t
nnecting to s
used on each
end ports to
ecting to a si
for VBLOCK A
OCK B
end ports to
LEX Frontend
cross the VPL
B
1,3,5,7
B A
4,8 4
2
2
onnect requits for host cr
that span botency prevenms.
mer has to do
x
ose paths as
th the lowest
n Italics can beral requireme
ts. 0, 1, 2 and to fabric B.
ingle VPLEX
h director. 2
fabric A and
ngle VPLEX
A
fabric A and
d
LEX front end
A B
4,8 3,7
Engine 3
ires hosts haoss-cluster cth datacenteting a viable
is enable th
ssociated wit
t minimum p
be adjusted ents simply us
d 3. In all cas
and 3 are res
odd to fabric
odd to fabric
d in the follow
A
2,6
Engin
ave connect ers. The use of
e
th the
path
se the
ses even
served.
c B.
c B.
wing
B
1,3
ne 4
3
4
. Host / E
a. Eac
b. For diff
c. Min
d. Max
e. Anysambla
f. Eac
i.
A
A
ii.
Th
Th
. Pathing
a. Nonpoland
b. For preto adat
Imp
SX Cluster ru
ch ESX cluste
dual and quferent engine
nimum direct
aximum direct
y given ESX cme VPLEX fronde count.
ch ESX host s
2 across fa
VPLEX A Dire
VPLEX B Dire
2 across fa
he same VPLE
he same VPLE
policy
n cross conneicy in all case
d quad system
cross conneferred paths
alternate andastores are n
plementation
ules
er must conne
ad configs, Aes (see table
tors that an E
ctors that an E
luster connentend ports f
should see fo
bric A
ector port 0 (o
ector port 0 (o
bric B
EX A Director
EX B Director
ected configues. Round roms.
cted configu set per Data
d balance ove not all sendin
n Planning an
ect to a VPLE
A and B direc above for rec
ESX cluster co
ESX cluster c
ecting to a giv for all UCS bl
our paths to t
or 2 if second
or 2 if second
r port 1 (or 3
r port 1 (or 3
urations recobin should b
urations, fixedastore to the er the whole ng IO to a sin
Chapter 2:
EMC VPLnd Best Pract
EX A and a B d
ctors must becommendati
onnects to is
connects to is
ven VPLEX cllades regard
the same dat
d VBLOCK)
d VBLOCK)
if second VB
if second VB
ommend to ube avoided es
d pathing sh local VPLEX VPLEX front ngle VPLEX di
: Frontend Con
LEX SAN Contices Technic
director.
e picked fromons)
s 2 VPLEX dire
s 2 VPLEX dir
uster must uless of host /
tastore
BLOCK)
BLOCK)
use adaptive specially for
ould be used path only tak end (i.e. so tirector).
nnectivity
1nectivity cal Notes
m
ectors.
rectors.
se the / UCS
pathing dual
d and king care
that all
15
16
Chapter 3: ESX
EMC VPLEX SImplementat
Xi Path Loss H
SAN Connecttion Planning
This
Pat
vSp
PDL
Handling
ivity g and Best Pr
s chapter pre
h loss handl
phere Storag
L/APD Refere
ractices Tech
Cha
esents the fo
ling semantic
ge for vSphere
ences ...........
nical Notes
pter 3
llowing topic
cs (PDL and A
e 5.1 ............
....................
ESXi P
cs:
APD) ...........
...................
...................
Path Los
....................
....................
....................
ss Hand
...................
...................
...................
dling
......... 17
......... 20
......... 21
Path loss
Persistent deloss (PDL)
s handlin
vSpnewLoson t
A stbecoccdevMetpercodthe the on tmacmacvSp
Whcanoff omacdatregusett
Fig
evice
ng seman
phere can recwer server. Ths" (PDL). Eith
the failure co
torage devicecomes permaurs when a d
vice experientro Storage Cmanently un
des allow you device as PD device to be
the same datchine will nochine on a si
phere HA.
en a dataston power off vi only when ischine that is astore mightulate the powtings apply o
gure 5 Persi
Imp
ntics (PD
cognize two dhese are knowher of these condition.
e is consideranently unavadevice is uninces an unrec
Cluster WAN pavailable, it
ur host to recDL. The sensee considered tastore and at be killed. Vngle datasto
re enters a Prtual machinsuing I/O to running memt remain activwer off and reonly to a PDL
istent devic
plementation
L and AP
different typewn as "All Pa
conditions ca
red to be in thailable to yountentionally rcoverable ha partition. Wh sends SCSI sognize that te codes mus permanently
a PDL conditi VMware recoore, ensuring
Permanent Denes and resta the datastormory-intensivve in such sitestart operat condition an
ce loss pro
n Planning an
PD)
es of total paaths Down" (Aan be declare
he permanenur ESXi host.
removed, its rdware error,hen the stora sense codes the device ha
t be receivedy lost. If virtuion exists onmmends pla that PDL con
evice Loss (Part them laterre. Otherwiseve workloadstuations. VMtions for virtund not to an A
ocess flow
Chapter 3: ES
EMC VPLnd Best Pract
th failures toAPD) and "Peed by the ESX
nt device loss. Typically, th unique ID ch, or in the caage determin to the ESXi has failed andd on all pathsual machine f
one of the dacing all files nditions can
PDL) state, Hir. A virtual me, it remains as without iss
Mware offers aual machines APD conditio
SXi Path Loss
LEX SAN Contices Technic
o an ESXi 5.0ersistent DevXi server dep
s (PDL) state he PDL condihanges, whese of a vSphe
nes that the d host. The sen
register the s to the devic
files do not adatastores, th for a given v be mitigated
gh Availabilimachine is po
active. A virtuing I/O to th
advanced ops. The followon.
Handling
1nectivity cal Notes
u1 and vice pending
when it ition n the ere
device is nse state of ce for
all reside he virtual virtual d by
ty (HA) owered ual he
ptions to wing
17
18
vf
vf
Chapter 3: ESX
EMC VPLEX SImplementat
vSphere Stofor vSphere
vSphere Stofor vSphere
Xi Path Loss H
SAN Connecttion Planning
Advenahos
NoteHA wrule
disk
For Dis
das
HA whepowthe set be rvirt
disk
AdvparmacWh
VM
AdvparusindiskvSp
1
2
3
4
5
6
7
das
HA virtWhwer
rage 5.0 U1
rage 5.5
Handling
ivity g and Best Pr
vanced settinable vSphere sts and VMs i
e: PDL respon will not violatee should be se
k.terminateV
each host ink.terminateV
s.maskCleanS
Advanced Opereas in ESXiwers off and i virtual mach to false, the restarted, whual machine
k.terminateV
vanced Virtuaameter powechine's datasen set to DEF
Mkernel.Boot.
vanced Vmkeameter poweng enters thek.terminateV
phere web cli
. Browse
. Click the
. Under S
. In Advan
. Click the
. Click OK
. Reboot t
s.maskCleanS
Advanced Opual machinesen this optiore intentiona
ractices Tech
ngs have bee HA to respon
in the stretch
se works in coe the rule. If tht to “should”
VMonPDLDef
the cluster, VMOnPDLDef
ShutdownEn
ption. If thei 5.1 and late its home dathine should b responding hile if this op should be re
VMOnPDLDef
al Machine Oers off the virstore enters tFAULT, VMke
.terminateVM
ernel Option. ers off all virte PDL state. SVMOnPDLDef
ent:
to the host in
e Manage tab
ystem, click
nced Settings
e Edit button
K
the host
ShutdownEn
ption. This os that were p
on is set to trully powered o
nical Notes
en introducednd to a PDL c
hed cluster co
onjunction withe rule is set t to provide ava
fault set to tr
create and efault=TRUE, t
nabled set to
option is uner, a value of astore is not
be restarted. FDM master tion is set to
estarted.
fault set to d
Option. Defaurtual machine the PDL staternel.Boot.ter
MOnPDL set t
Default valutual machineSetting can bault paramet
n the vSpher
b and click S
Advanced Sy
s, select the
to edit the v
nabled set to
option is set tpowered off wue, HA restar off by a user.
d in VMware v condition. Thonsuming th
th DRS rules. to “should”, Vailability.
rue
edit /etc/vmw then reboot e
true:
nset in 5.0U1f true is assumt accessible, So, it must m will assume
o true, the res
efault:
ult value is FAe if any device. HA will notrminateVMOn
to true:
ue is FALSE. es on the systbe overriddenter. Can be s
re Web Client
ettings.
ystem Setting
appropriate
value.
default:
to TRUE by dwhile the PDLrts all virtual .
vSphere 5.0 he following e virtual stor
If the rule is sVMware HA wil
ware/settings each host.
, a value of fmed. When a HA cannot d
make a decis the virtual msponding FDM
ALSE. Whence that backst restart this nPDL is used
When set to tem when ston for each virt set only to TR
t navigator.
gs.
item.
efault. It alloL condition w machines, in
Update 1 an settings are rage.
et to “must”, Vll violate it. Th
s with
false is assum a virtual mac
etermine whsion. If this o
machine shouM will assum
TRUE, this s up the virtu virtual machd.
TRUE, this orage that thtual machine
RUE or FALSE.
ows HA to reswas in progrencluding thos
nd 5.5 to for the
VMware he DRS
med, hine
hether ption is
uld not me the
ual hine.
hey are e by . With
start ss. se that
A
APD Handlin
disk
AdvPDLwhipertheyrescvMSwanfunPleaconsett
1
2
Or w
1
2
3
4
5
6
Per
All
A stbecreas
In cas t
Thecon
g
k.Autoremov
vanced VmkeL’s are likely ich case a faimanently why will not autcan in order tSC environmnt to have to ctionality shoase note that
nfigurations. Ating changed
. ConnectUsing Te
. Run thisset -o "/
with vSphere
. Browse
. Click the
. Under S
. In Advan
. Click the
. Click OK
rmanent Devi
Remov
WAN pwan po
Paths Down
Remov
Remov
Disabl
torage devicecomes unavasons for an A
contrast with transient and
e host indefinnnectivity wit
Imp
veOnPDL set
ernel option. temporary beilover has oc
hen the failurtomatically b to bring the dent is that it do manual rould be disat this is recom Any vMSC cod. To disable
t to the ESXi ech Support M
s command to/Disk/Autore
e web client:
to the host in
e Manage tab
ystem, click
nced Settings
e Edit button
K
ice Loss
ve device from
partition, disaorts using vp
ve volume fro
ve FC ports fr
le FC ports on
e is considerilable to you
APD state can
the permaned expects the
nitely continuh the device
plementation
to 0:
Default is 1.ecause one scurred. If the
re or configurbe visible to t devices back handles thesescans all thbled on all hmmended foonfiguration te this feature
host using th Mode in ESX
o disable AutmoveOnPDL"
n the vSpher
b and click S
Advanced Sy
s, select the
to edit the v
m VPLEX, rem
able wan porplexcli
om storage v
rom ESXi hos
n switch.
red to be in thr ESXi host fo
n be, for exam
ent device loe device to be
ues to retry is. If the host's
n Planning an
. In the case site has becoe devices in aration error o
the hosts agak into servicese types of th
he time. For thosts that are
or Uniform or that could ca:
he console oXi 4.1 and ESX
toRemove: e" -i 0
re Web Client
ettings.
ystem Setting
appropriate
value.
move or offlin
rts from switc
iew.
st, can cause
he all paths dor an unspecmple, a failed
ss (PDL) state available a
ssued commas commands
Chapter 3: ES
EMC VPLnd Best Pract
e of a vMSC eome orphane a PDL state aof the vMSC eain. This will e. The whole hings automhis reason the part of a vM Non-Uniformause a PDL sh
r SSH. For moXi 5.x (KB art
esxcli system
t navigator.
gs.
item.
ne a LUN from
ch or log in v
other errors
down (APD) scified period d switch.
te, the host tgain.
ands in an at fail the retrie
SXi Path Loss
LEX SAN Contices Technic
environment ed from the ore removed
environment require a ma reason for haatically. So y
he PDL AutoRMSC configuram vMSC hould have t
ore informatiticle 101791
settings adv
m backend.
vplex and dis
.
state when it of time. The
reats the APD
ttempt to reees for a prolo
Handling
1nectivity cal Notes
the other, in
is fixed anual aving a you don’t Remove ation.
he
ion, see 0).
vanced
able
t e
D state
establish onged
19
20
v
A
A
Chapter 3: ESX
EMC VPLEX SImplementat
vSphere
Disable StorAPD Handlin
Change TimeLimits for StoAPD
Xi Path Loss H
SAN Connecttion Planning
perper
Witentcon
Storage
TheWhthe its ron y
If yocomsamthe hos
Pro
1
2
3
4
5
If yostatwith
Noteprobapp
Themacnee
Theexpnon
Thevalulong
rage g
eout orage
Handling
ivity g and Best Pr
iod of time, tformance pro
h vSphere 5.ers the APD s
ntinue retryin
for vSph
e storage all pen it is enab APD state fo
retry attempt your host.
ou disable thmmands in anme behavior a host to exce
st might beco
cedure
. Browse
. Click the
. Under S
. Under Aparamet
. Change
ou disabled tte. The internh the current
e: The host cablems as APD
propriate SCSI
e timeout parchine I/O com
eded, you can
e timer starts pires, the hosn-virtual mac
e default timeue of the timger than 140
ractices Tech
the host and oblems and p
.1, a default A state, the sysg non-virtual
here 5.1
paths down (led, the host
or a limited tits and termin
he APD handln attempt to as in ESXi veeed their inteome disconne
to the host in
e Manage tab
ystem, click
dvanced Syster and click
the value to
the APD handnal APD handt timeout valu
annot detect P when a storag sense codes.
rameter contrmmands to an change the
immediatelyst marks the Ahine I/O. Virt
eout parameteout if, for ex
0 seconds to
nical Notes
its virtual ma potentially b
APD handlingstem immedil machine co
(APD) handlint continues tome period. W
nates any I/O
ling, the host reconnect torsion 5.0. Thrnal I/O timeected from v
n the vSpher
b, and click S
Advanced Sy
stem Settings the Edit icon
0.
dling, you cadling feature ue for each d
DL conditionsge device perm
rols how mana storage deve default time
y after the de APD device atual machine
ter on your hxample, stora recover from
achines mighecoming unr
g feature waiately turns o
ommands for
ng on your ESo retry I/O co
When the timO. You can dis
t will indefino the APD dev
is behavior meout and becCenter Serve
re Web Client
Settings.
ystem Setting
s, select the n.
an re-enable i turns on imm
device in APD
s and continuemanently fails
ny seconds tvice in an all eout value.
evice enters t as unreachabe I/O will con
ost is 140 seage devices c
m a connectio
ht be at risk responsive.
s introducedon a timer anr a limited tim
SXi host is enommands to e period expsable the AP
itely continuvice. Continu
might cause ome unrespo
er.
t navigator.
gs.
Misc.APDHa
it when a devmediately anD.
es to treat the s in a way that
he ESXi host paths down
the APD stateble and fails ntinue to be r
econds. You connected toon loss.
of having
d. When a devd allows you
me period.
nabled by de a storage depires, the hosD handling fe
e to retry issuing to retry i virtual machonsive or fail
ndlingEnable
vice enters thd the timer s
device connec does not retu
t will retry no (APD) state.
e. When the t any pending
retried.
can increaseo your ESXi h
vice ur host to
efault. vice in
st stops eature
ued s the
ines on l. The
e
he APD starts
ctivity urn
n-virtual If
timeout g or new
e the ost take
PDL/APD
Pro
1
2
3
4
5
D Referen
httpft-h
http102
httpplay
httpft-h
http102
httpvce
httpvce
httpst%
httpvsp
httpplay
httpplay
cedure
. Browse
. Click the
. Under S
. Under Aclick the
. Change
nces
p://www.emca.pdf for ESX
p://www.vmw2-HI-RES.pdf
p://kb.vmwayKC&externa
p://www.emca.pdf
p://www.vmw2-HI-RES.pdf
p://pubs.vmwnter-server-5
p://pubs.vmwnter-server-5
ps://support%20practices
p://www.bocphere-ha-and
p://kb.vmwayKC&externa
p://kb.vmwayKC&externa
Imp
to the host in
e Manage tab
ystem, click
dvanced Syse Edit icon.
the default v
c.com/collatXi 5.0 U1 test
ware.com/fil for ESXi 5.0
are.com/selfsalId=200754
c.com/collat
ware.com/fil
ware.com/vs50-storage-gu
ware.com/vs55-storage-gu
t.emc.com/se%20for%20e
che.net/blog-pdl
are.com/selfsalId=203325
are.com/selfsalId=205962
plementation
n the vSpher
b, and click S
Advanced Sy
stem Settings
value.
teral/softwart scenarios
es/pdf/tech U1 vmware v
service/micro5
teral/softwar
es/pdf/tech
sphere-50/touide.pdf
sphere-55/touide.pdf
earch/?text=emc%20vple
/index.php/
service/micro0
service/micro2
n Planning an
re Web Client
Settings.
ystem Setting
s, select the
re/white-pap
paper/vSPHR vsphere metr
osites/searc
re/white-pap
paper/vSPHR
opic/com.vm
opic/com.vm
=implementaex
2014/07/14
osites/searc
osites/searc
Chapter 3: ES
EMC VPLnd Best Pract
t navigator.
gs.
Misc.APDTim
pers/h11065
R-CS-MTRO-Sro storage clu
h.do?langua
pers/h11065
R-CS-MTRO-S
mware.ICbase
mware.ICbase
ation%20and
4/yet-anothe
h.do?langua
h.do?langua
SXi Path Loss
LEX SAN Contices Technic
meout param
-vplex-with-v
STOR-CLSTR-uster case st
ge=en_US&c
-vplex-with-v
STOR-CLSTR-
e/PDF/vsphe
e/PDF/vsphe
d%20plannin
r-blog-post-a
ge=en_US&c
ge=en_US&c
Handling
2nectivity cal Notes
eter and
vmware-
USLET-udy
cmd=dis
vmware-
USLET-
re-esxi-
re-esxi-
ng%20be
about-
cmd=dis
cmd=dis
21
22
Chapter 4: VP
EMC VPLEX SImplementat
LEX Backend A
SAN Connecttion Planning
This
Bac
Array Connect
ivity g and Best Pr
Chapte
s chapter pre
ck-end/Stora
tivity
ractices Tech
er 4 V
esents the fo
age Array Con
nical Notes
VPLEX B
llowing topic
nnectivity ....
Backend
cs:
...................
Array C
....................
Connecti
...................
ivity
......... 23
AA
Back-end
Active/ActiveArrays
d/Storag
TheVPLrecocon
Direonlyswicon
Notesup
VPL(Opthe
Note
Higdivipreas nfolloas Iporhavpredireleavdiff
Appfor tstorgro
e
ge Array C
e best practicLEX is also caognizable as
nnect should
ect connect isy 1 array. Thtch ports. Si
nnectivity as t
e: Direct connported.
Each direclocal backcluster (re
Each direcacross dupaths to e
Each storafabrics, wthe back-e
LEX recommeptimal). This four active p
e: Exceeding t
h quantities ided up into sented from
not to exceedowing the ru
TLs), a four erts dedicatedve the ability sented throu
ector. This caving the otheferent array p
propriateness the group of rage arrays oup, or maski
Imp
Connectiv
ce for array coapable of bac being extrem be followed
s intended fois allows for ites with mul that provides
nect applies o
ctor in a VPLEk-end storageequired). Thi
ctor will haveal fabrics (re
every back-en
age array shoith each VPLend storage a
ends a maxim is considerepaths to the s
the maximum
of storage vo appropriatel the array to Vd the four actle of four act
engine VPLEX to that direc to connect to
ugh all 32 poan be accomer two ports fports.
s would be ju LUNs and lim
often have limng view they
plementation
vity
onnectivity isckend direct mely limited. with regards
or Proof of Co backend conltiple arrays os the optima
nly to backend
EX cluster mue array and tos is referred
e redundant pequired) Eachnd storage ar
ould have redEX Director harrays throug
mum of 4 actid optimal be
storage volum
of 4 active pa
olumes or eny sized group
VPLEX via grotive paths petive paths peX cluster coulctor. In othero 32 ports onrts and still mplished usin
for access to
udged basedmitations of tmits around ty can have.
Chapter 4
n Planning an
s to use A/B connect. Th. The followis to the direc
oncept, test/nnectivity wh
or large impll solution ov
d connectivity
ust have a mo every stora to as an ITL o
physical conh director is rray across b
dundant conhaving a minigh the dual fa
ive paths perecause each me.
aths per direct
ntire arrays pps (i.e. maskoups of four er VPLEX direr storage volld have eachr words, a qun a single arr
meet the cong only two po another set
d on things lik the physical the number o
4: VPLEX Back
EMC VPLnd Best Pract
fabrics for reis practice isng best practt connect wh
/dev and spehile reducingementationserall.
y. Frontend di
inimum of twage volume p or Initiator/T
nections to t required to h
both fabrics.
trollers connimum of two abrics (requi
r director to a director will
or per LUN is n
rovisioned toking views or array ports pctor limitatiolume per direh director conuad engine Vray for accessnnectivity ruleorts per back of volumes o
ke the plann storage arrayof LUNs per s
kend Array Con
LEX SAN Contices Technic
edundancy hos immediatelytices for fabr
here applicab
cific sites thag the overall cs should utiliz
rect connect is
wo I/O paths presented to tTarget/LUN.
the back-endhave redunda
nected to dua ports conneired).
a given LUN load balance
not supported
o VPLEX shour storage grouper VPLEX diron. As an exaector (also rennected to foPLEX cluster s to a single es of 4 ITLs pkend I/O modover the same
ed total IO wy. For examp
storage port,
nnectivity
2nectivity cal Notes
owever y ric ble.
at have cost of ze SAN
s not
to every that
d storage ant
al cted to
e across
d.
uld be ups) and ector so
ample, ferred to
our array would device per dule e or
workload ple, storage
23
24
Chapter 4: VP
EMC VPLEX SImplementat
LEX Backend A
SAN Connecttion Planning
Maxmaxis nPro
Loapat
1
2
3
4
5
6
Array Connect
ivity g and Best Pr
ximum perfoximum numbot based on per balancin
ad balancing hs on each d
. Zone VP
. Zone VP
. Repeat f
. Create agroups.
. Spread e
. Mask degroups.
tivity
ractices Tech
rmance, envber of ports o a single hosg of all availa
via Host Muldirector balan
PLEX director
PLEX director
for additiona
a separate po
each group o
evices to allo
nical Notes
ironment widon an array wt but the oveable resourc
ltipath betwences the load
A ports to on
B ports to a
al VPLEX engi
ort group with
of four ports a
ow access to
de, is achievewhile staying werall impact oes provides t
een VPLEX did equally bet
ne group of fo
different gro
nes.
hin the array
across array
the appropri
ed by load b within the IT of all resourc the best ove
rectors and ttween the arr
our array por
up of four ar
for each of t
engines for r
iate VPLEX in
alancing acro limits. Perfoes being utilrall performa
then from theray ports.
rts.
rray ports.
these logical
redundancy.
nitiators for b
oss ormance ized.
ance.
e four
path
oth port
7Fig
Thisconpra
TheandwithVPL
Theenggro
. gure 6 Activ
s illustration nsiderations sctices for all
e devices shod provisionedhin this devicLEX Director I
e VPLEX initiagines to increups such tha
Imp
ve/Active Ar
shows the p should apply arrays includ
ould be provid for access tce grouping s
TL group.
ators (backenease HA and at each VPLEX
plementation
rray Connec
physical conny to other actding third pa
sioned in su through spec should restri
nd ports) on a redundancy.X director act
Chapter 4
n Planning an
ctivity
nectivity to a tive/active arrty arrays.
ch a way as tcific FA ports ct access to
a single direc The array sts as a single
4: VPLEX Back
EMC VPLnd Best Pract
VMAX array.rrays. Follow
to create “dig and VPLEX p four specific
ctor should sshould be coe host per fou
kend Array Con
LEX SAN Contices Technic
Similar w the array be
gestible” chuports. The dec FA ports for
spread acrosnfigured intour paths.
nnectivity
2nectivity cal Notes
est
unks evices each
s o initiator
25
26
Chapter 4: VP
EMC VPLEX SImplementat
LEX Backend A
SAN Connecttion Planning
Thisdepmu
For
We foutwoInitthisturnGroandallo
VPLdireinitVPLthroB d
Thisnumdireove
Thisbes
Array Connect
ivity g and Best Pr
s could meanpending on pltiple fabrics
the example
Initiator G
Port Grou
Storage G
Masking
have dividedr masking vie
o FA’s each oniator Group a
s group of WWn become a mups. This wa
d 4. This is oowed as long
LEX virtual voector A and iniators conne
LEX would acough two VPLirector each
s is an optimmber of arrayectors in a duerall environm
s would allowst possible en
tivity
ractices Tech
n four physicort availabili in excess of
e above follow
Groups (HBAVMaxonly IG's (
ups (FA portsport;
Groups (LUNSymDon a V
View = Initia
d the backenews on the Vn the array. T
and will be pWN’s will becmember of a as repeated fnly one exam as the rules
olumes shoulnitiators fromcted to dual cess the bacLEX directors see four diffe
al configuraty ports. Additual-engine orment I/O wor
w for the greanvironment p
nical Notes
al paths or foity and whethf two.
wing basic lim
As); max of 32x; set port flabelong to 1 IG(rather than W
s): max of 512 ports may be
Ns/SymDevs);Dev may beloVMax
ator Group +
nd ports of thVMAX. Ports F The WWN’s oart of Maskincome the me second Mask for ports FC0mple of attac are followed
ld be added tm a director B fabrics and hckend array’ss (an A directoerent FA’s ac
tion that spretional hosts r quad-enginekload over m
atest possiblperformance.
our logical paher or not VP
mits on the V
2 WWN's per gs on the IG; G. Cascaded IWWN's) as me
2 PG's; ACLX elong to more
; max of 4096ong to more th
Port Group +
he VPLEX into FC00 and FC0 of these portsng View 1. Thmber of a seking View. T2 and FC03 phing to the V
d.
to Storage ViB. This transl having four ps storage voluor and a B diross at least
eads a single will attach toe VPLEX clus
more switches
le balancing .
aths per VPLPLEX is attach
VMAX:
IG; max of 81 an individuaInitiator Grouembers.
flag must be e than 1 PG
6 SymDevs pehan 1 SG; ma
+ Storage Gro
o two groups 01 for both ds are the mehe Initiator G
econd InitiatoThis is called placing themVMAX and oth
iews containlates to a sinpaths into twumes via eigirector). The two VMAX e
e host’s I/O oo different paster. This wils, VPLEX and
of all resourc
EX director hed to dual fa
192 IG's on a al WWN can ups have othe
enabled on th
er SG; a x of 8192 SG's
oup
allowing us tdirectors are mbers of the
Group createdor Group whic Cascading In
m in Masking her possibilit
ing initiatorsngle host witho VPLEX direht FA’s on th VPLEX A direngines if ava
over the maxairs of VPLEX ll help sprea
d array resou
ces resulting
abrics or
er
he
s
to create zoned to e first d with ch will in nitiator Views 3 ties are
s from a h two ectors. e array
ector and ailable.
imum d the rces.
g in the
AA
Active/PassiALUA Arrays
Fig
ThisClusper panwhi
Thearrapas
Somspedirethe to sCLAnecVPL
ive and
gure 7 Show
s illustration ster is a sing Storage Volu
nel displays tich FA they a
e proper outpay would havssive or non-p
me arrays havecial consideector needs t active and p
storage volumARiiON® havcessary to haLEX. Adhere t
Imp
ITLs per St
shows the ITgle engine anume per Dire
the ports on tre connected
put in the Shove double thepreferred SP
ve architecturation. Wheno have logica
passive or nomes if an actie limitationsve multiple g
to logical and
plementation
torage Volu
TLs per Storad is connectctor giving u
the VPLEX Did to.
ow ITLs panee count as it w on the array
ure and implen using an acal (zoning ann-preferred cive controller on the size o
groups to accd physical co
Chapter 4
n Planning an
ume
age Volume. ed to an actis a total of e
irector from w
l for an activ would also c.
ementation rctive-passive
nd masking) acontrollers. Tr should fail. of initiator orcommodate ponnectivity g
4: VPLEX Back
EMC VPLnd Best Pract
In this examve/active arr
eight logical p which the pa
ve/passive orontain the lo
requirementse or ALUA sup and physical That way you Additionallyr storage gro
provisioning uidelines dis
kend Array Con
LEX SAN Contices Technic
ple the VPLEray with four
paths. The Sths originate
r ALUA suppoogical paths f
s that necesspported arrayl connectivityu will not losey, arrays likeups. It may storage to thscussed earli
nnectivity
2nectivity cal Notes
EX paths how ITLs
e and
orted for the
sitate y, each y to both e access the
be he ier.
27
28
Chapter 4: VP
EMC VPLEX SImplementat
LEX Backend A
SAN Connecttion Planning
Fig
PoinfabexaSPb
ALUrecodirewill
WhVPLdev
ProwaycontwoITL pasa mActi
Noteplanon cnum
The“LoconDire
Array Connect
ivity g and Best Pr
gure 8 VS2 c
nts to note wric through w
amples showsb3(odd ports)
UA support alognize the noector with pro only report f
en provisionLEX initiator rvices are alre
per connectiys. You havenfiguration who or four pass or Initiator/Tssive/non-pre
maximum supive/Passive a
e: VNX2 will sn to upgrade t
configuring wimber of paths
e next set of dgical” config
nfiguration. Bector per LUN
tivity
ractices Tech
connectivity
would be thatwhich each Ss Fabric-A wi) for dual fab
llows for conon-preferred oper maximu four active pa
ing storage t registration p
ady presente
vity for activee the option ohich amountsive or non-pTarget/LUN Neferred) path
pported confi and ALUA.
support Activehe VNX2 to a cth two (2) acti
later.
diagrams depguration and Both are suppN configuratio
nical Notes
y to Active
t for each CLAP has connecth SPa0 & SP
bric redundan
nectivity sim path and ref
um path connaths.
to VPLEX, ensprior to deviced.
e/passive an of configuringt to two or foureferred pathexus. A minhs (Logical oguration is 8
e/Active conne code level thaive paths initia
picts both a f a four “activported configons.
e/Passive an
ARiiON, eachctions to all VPb0 (even poncy.
milar to Activefrain from usnectivity will
sure that mode presentatio
nd ALUA arrayg to a minimur active pathhs per Directoimum configr Physical) p
8 (eight) path
ectivity. If youat supports Acally. This will
four “active”e” path per Dgurations as
nd ALUA Arr
h SP has conn VPLEX directorts) and Fab
e/Passive arring it under n show eight IT
de 4 (ALUA) oon. Don’t try
ys can be haum configurahs per Director per LUN. T
guration is 4 (er Director fo
hs per Directo
u are configurintive/Active th eliminate the
path per Dir Director per L well as two a
rays
nection to eaors. The abovric-B with SP
rays. VPLEX w normal condTLs per direc
or mode 1 sey to change it
andled in a coation or maxitor per LUN a This is known (two active aor any given Lor per LUN fo
ng ALUA now en you should need to reduc
rector per LU LUN “Physica active paths
ach ve
Pa3 &
will itions. A
ctor but
et during t after
ouple of imum s well as n as an
and two LUN and r
but d plan ce the
N al” per
Thethe
As a
Fig
Runis a
e commands following dia
VPlexcli:/
VPlexcli:/volumes/<
an example:
gure 9 Backe
nning the lonssociated wi
Imp
used in the Vagram are:
ll **/hardwa
clusters/clus<storage volu
end port WWN
g listing on tith which VPL
plementation
VPlexcli to de
re/ports
ster-<cluster nume name>>ll
N identific
the hardwareLEX backend
Chapter 4
n Planning an
etermine the
number>/stol --full
cation
e/ports allowd port.
4: VPLEX Back
EMC VPLnd Best Pract
port WWNs
orage-elemen
ws you to dete
kend Array Con
LEX SAN Contices Technic
and the ITLs
nts/storage-
ermine which
nnectivity
2nectivity cal Notes
used in
h WWN
29
30
Chapter 4: VP
EMC VPLEX SImplementat
LEX Backend A
SAN Connecttion Planning
Fig
Fromcon
In tclusfor tmodOnl
Noteredidetr
Array Connect
ivity g and Best Pr
gure 10 ITL
m the storagntext. Runnin
his example ster connecte this volume ode 1 or modely in the case
e: All paths, wiscover. Over rimental effect
tivity
ractices Tech
association
e-volumes cong the ll --full
we have sixted to a VNX a only four pate 4 (ALUA), the of a trespas
whether active allocating thets on performa
nical Notes
n
ontext you cal command w
teen entries f array. Even tths go to the he paths goinss will they be
e or not, will pee number of paance and/or b
an select a sawill show the
for this volumhough this g array SP thatng to the othecome active
erform deviceaths beyond thbackend LUN p
ample volum ITLs.
me. This is agives us eightt owns the voer SP will noe.
discovery durhe supported
provisioning.
me and cd to t
single engint paths per dolume. In eitt be used for
ring an array limits will hav
that
ne VPLEX irector ther r I/O.
ve
Fig
This
gure 11 1:1
s drawing wa
Imp
Physical pa
as developed
plementation
ath configu
d from the ou
Chapter 4
n Planning an
uration
utput from the
4: VPLEX Back
EMC VPLnd Best Pract
e two comma
kend Array Con
LEX SAN Contices Technic
ands shown
nnectivity
3nectivity cal Notes
above.
31
32
S
Chapter 4: VP
EMC VPLEX SImplementat
Significant B
LEX Backend A
SAN Connecttion Planning
Fig
A slusinnumpor
Theconbit
Theof slow
Tha
Bit
Array Connect
ivity g and Best Pr
gure 12 Logi
light modificng only two Vmber of ITLs arts.
e above two insiderations f positions as
e CLARiiON CXspecifying the
w end ports.
at is, SPA9 wo
tivity
ractices Tech
ical path co
ation from thVPLEX backen and meets th
llustrations s for identifyin they apply to
X4 Series supe Ports would
ould have the
nical Notes
onfiguratio
he previous dnd ports per
he maximum
show the signg all possiblo the various
pports many d cause an ov
e significant
on
drawing help director. Thi supported li
nificant bit be ports. The
s modules on
y more SP porverlap betwe
byte pair as
s illustrate this gives us thimit as sprea
but there are following wi
n a CLARiiON
rts. As such teen SP A high
69, which is
he same conhe exact samading across
other bit ill help expla / VNX.
the original mh end ports a
SPB1.
cept but e
all four
ain the
method and SP B
A
Additional AConsideratio
The
SPA
Port
00
01
02
03
04
05
06
07
For
0 0-
4 8-
8 16
C 24
AndA an
Port
08
09
10
11
Arrapatpurnevis aman
StoAnyvolucon
Array ons
e new method
A0-7 and SPB
t SP
50:06:01:6
50:06:01:6
50:06:01:6
50:06:01:6
50:06:01:6
50:06:01:6
50:06:01:6
50:06:01:6
the higher p
-7
-15
6-23
4-31
d the 8th bytend SP B we h
t SP
50:06:01:6
50:06:01:6
50:06:01:6
50:06:01:6
ays, such as h from some
rposes of in-bver simultane high probabnagement an
rage volumey volumes whumes to be c
ntain data an
Imp
d is as follow
B0-7 are the s
P A
60:BB:20:02:
61:BB:20:02:
62:BB:20:02:
63:BB:20:02:
64:BB:20:02:
65:BB:20:02:
66:BB:20:02:
67:BB:20:02:
ort numbers
e cycles backhave:
P A
60:BB:24:02
61:BB:24:02
62:BB:24:02
63:BB:24:02
the Symmetr hosts to the
band manageeously be mability of data nd keep them
s provided bhich are not alaimed. For d are not a m
plementation
ws :
same as the
S
:07 50:06:0
:07 50:06:0
:07 50:06:0
:07 50:06:0
:07 50:06:0
:07 50:06:0
:07 50:06:0
:07 50:06:0
byte 12 is c
k to 0-7 for S
S
2:07 50:06
2:07 50:06
2:07 50:06
2:07 50:06
rix®, that do ie array. Suchement. Stora
asked directly corruption.
m outside of t
y arrays mus a multiple of the use case
multiple of 4K
Chapter 4
n Planning an
old method.
SP B
01:68:BB:20
01:69:BB:20
01:6A:BB:20
01:6B:BB:20
01:6C:BB:20
01:6D:BB:20
01:6E:BB:20
01:6F:BB:20
changed to re
P A and 8-F f
SP B
6:01:68:BB:2
6:01:69:BB:2
6:01:6A:BB:2
6:01:6B:BB:2
in-band manh a direct patage volumes y from the ar
It may be bes the VPLEX en
st have a cap 4KB will not
e of presentinK then those
4: VPLEX Back
EMC VPLnd Best Pract
0:02:07
0:02:07
0:02:07
0:02:07
0:02:07
0:02:07
0:02:07
:02:07
epresent the
for SP B. So f
24:02:07
24:02:07
24:02:07
24:02:07
agement mah should be provisioned rray to the host to dedicat
nvironment.
pacity that is show up in tng storage vo devices will
kend Array Con
LEX SAN Contices Technic
higher ports
for ports 8-11
ay require a d solely for the to the VPLEX
ost; otherwisete hosts for in
a multiple of the list of avaolumes to VP have to be m
nnectivity
3nectivity cal Notes
thus:
1 on SP
direct e X should e there n-band
f 4 KB. ailable
PLEX that migrated
33
34
Chapter 4: VP
EMC VPLEX SImplementat
LEX Backend A
SAN Connecttion Planning
to aalteunu
Remdocarra
Array Connect
ivity g and Best Pr
a volume thaternative woulused VPLEX d
member to recumentation ay best pract
tivity
ractices Tech
t is a multiplld be to use a
device.
eference the E for specific aices for confi
nical Notes
e of 4K first t a host based
EMC Simple S array configuiguring devic
then that devd copy utility
Support Matration requir
ces to VPLEX.
vice presente to move the
trix, Release rements. Rem.
ed to VPLEX. data to a new
Notes, and omember to fo
The w and
online ollow
This
Xtre
VPL
s chapter pre
emIO Overvie
LEX Connectiv
Imp
C
esents the fo
ew ...............
vity to Xtrem
plementation
Chapter
llowing topic
....................
mIO ...............
n Planning an
5 Xtr
cs:
...................
...................
Chapter 5
EMC VPLnd Best Pract
remIO C
....................
....................
5: XtremIO Con
LEX SAN Contices Technic
Connecti
...................
...................
nnectivity
3nectivity cal Notes
ivity
......... 36
......... 37
35
36
X
Chapter 5: Xtr
EMC VPLEX SImplementat
XtremIO
remIO Connect
SAN Connecttion Planning
Overview
TheThe
ThecallXMSope
Xtrepothosperpre
Theovetrueutil
Due
Add2.4
tivity
ivity g and Best Pr
w
e XtremIO Stoe system uses
e system opeled the XtremS host, which
erating if it is
emIO's array ential of flas
st ports in a bformance levdictable app
e XtremIO Stoer time, systee random I/Oization level,
e to its conte
• Even disand min
• Even dis
• No data
• Easy set
• Advance(deduplprotectio
ditional infor White Paper
ractices Tech
orage Array iss building blo
ration is conmIO Managemh can be eith disconnecte
architecture h, while line
balanced mavel, while malication beha
orage Array pm conditions
O. The cluster, number of v
nt-aware sto
stribution of dnimal flash w
stribution of
or metadata
tup and no tu
ed storage fuication and don (XDP), sna
mation abour, located at E
nical Notes
s an all-flash ocks, called
trolled via a ment Server (her a physicaed from the X
is specificalarly scaling anner. This al
aintaining conavior.
rovides a vers and accessr's performan
volumes, or a
rage architec
data blocks, ear
metadata
a hotspots
uning
unctionality, i data compresapshots, and
ut XtremIO caEMC.com.
system, bas X-Bricks, wh
stand-alone (XMS). Each X
l or a virtual XMS, but can
ly designed t all resourceslows the arransistency of
ry high level s patterns. It nce level is n aging effects
cture, XtremI
inherently le
including Inlssion), thin pd more
an be found i
sed on a scalich can be cl
dedicated Li XtremIO clus server. The anot be config
to deliver the such as CPU
ay to achieve performance
of performan is designed ot affected b.
O provides:
eading to ma
ine Data Red provisioning,
n the EMC VP
e-out architelustered toge
inux-based ster requires
array continugured or mon
e full performU, RAM, SSDse any desirede that is critic
nce that is co for high granby its capacit
aximum perfo
duction , advanced d
PLEX WITH XT
ecture. ether.
server, its own ues nitored.
mance s, and
cal to
onsistent nularity ty
ormance
ata
TREMIO
V
VPLEX Co
onnectivi
XtreConThisWhsimperStofolloconpos
Witthe septhastorconthe initgroaccvoluto tvolu
Thelimifromcon
ity to Xtr
emIO is a truentrollers has s is made poen deciding h
mply follow VPspective, yourage Controlowing diagra
ncept. This cossibilities tha
hin the Xtrem cluster. It is
parate initiaton most otherrage view or
ntain the obje administratoiators. Xtremuping of initiess to any giume to the inhat volume fume.
e following illited to thesem which you nfiguration ne
Imp
remIO
e active/activ access to allssible due to
how to propePLEX active/au want to sprlers balancin
ams illustrateoncept shou
at are not sho
mIO provisions not necessaor groups. Thr arrays. Mos storage grouects such as or wishes to
mIO simply aliators from a iven volume.nitiator group from the initi
ustrations de specific con should be abeeds.
plementation
ve array. Eac storage prov
o the Infinibaerly connect Vactive array cread the pathng the workloe multiple cold be easily a
own.
ning, a singleary to try andhe XtremIO pst arrays use
up as a few ex LUNs or devi use on the allows you to single host o When provi
p. You do noators in that
emonstrate bfigurations oble to extrap
n Planning an
ch Fibre Chanvisioned on aand connectiv VPLEX backe
connectivity bhs from eachoad evenly acmbinations a
adapted for a
e initiator grod configure a provisioning u a masking cxamples, andices, initiatorrray for acce
create initiat or clustered isioning volut select whic initiator grou
best practiceonly. We are olate a desig
Chapter 5
EMC VPLnd Best Pract
nnel and iSC all X-Bricks wvity between
end ports to t best practice VPLEX Direccross the ove and should c additional co
oup can acce single VPLEX
uses a much container calld these contars and the poss to those dtor groups wh host which yumes you simch port you wup. Any port
e recommend providing segn from for yo
5: XtremIO Con
LEX SAN Contices Technic
SI port on alwithin the clun Storage Con the XtremIO ces. From a hictor across Xterall cluster.
clearly show onfiguration
ess all storagX cluster into simpler appled a maskinainers typicaorts on the ardevices from hich are a simyou want to h
mply connect want to allow
can access a
dations but aeveral combinour customer
nnectivity
3nectivity cal Notes
l Storage uster. ntrollers. cluster igh level tremIO The the
e within o two proach ng view, ally rray that those mple have the access any
re not nations r’s
37
38
SS
Chapter 5: Xtr
EMC VPLEX SImplementat
Single EnginSingle X-Bric
remIO Connect
SAN Connecttion Planning
Fig
ThisBricrecoeacandStoava
Thisbas1:1 of tinit
Tota
ne / ck
tivity
ivity g and Best Pr
gure 13 VPLE
s illustration ck cluster. Thommending f
ch XtremIO Std blue colorinrage Control
ailability.
s illustration sis. Based o basis therebhe XtremIO cial deployme
al Bandwidth
64 Gb/s
ractices Tech
EX single en
is a VPLEX shis configurafour paths petorage Controngs. Both VPlers on both
is mapping tn total availa
by leaving twcluster. This ent and still g
h VPLEX T
nical Notes
ngine - Xtr
ingle engine ation meets ther Director peoller are conn
PLEX Director fabrics. This
the VPLEX baable bandwid
wo ports avail would reduc
give the same
Total Bandwi
32 Gb/
remIO single
cluster connhe VPLEX baer Storage Vonected to bors are cross cs connectivity
ackend portsdth, you maylable on eachce the port coe maximum t
idth XtremIO
/s
e X-Brick
nected to an Xckend conneolume. Eachth fabrics asonnected to y allows for t
s to the Xtremy consider mah VPLEX Direount on the fathroughput.
(as depicted
XtremIO singectivity rule VPLEX Direc depicted by both XtremIO
the highest le
mIO ports on apping the pctor for futurabrics at time
d)
gle X-
ctor and y the red O evel of
a 2:1 orts on a
re scale e of
S
Single EnginDual X-Brick
Fig
ThisBric
Thisrecoallo
Tota
ne / k
gure 14 VPLE
s illustration ck XtremIO cl
s example foommended f
owing for max
al Bandwidth
64 Gb/s
Imp
EX single en
demonstrateluster.
llows the fou for performanximum throu
h VPLEX T
plementation
ngine - Xtr
es a single en
ur paths per Vnce. The portghput using
Total Bandwi
64 Gb/
n Planning an
remIO dual X
ngine VPLEX
VPLEX Directrts are mappe all available
idth XtremIO
/s (as depict
Chapter 5
EMC VPLnd Best Pract
X-Brick
cluster conn
tor per Storaged 1:1 betwe
e ports on bot
ted)
5: XtremIO Con
LEX SAN Contices Technic
nected to a d
ge Volume rueen the two cth clusters.
nnectivity
3nectivity cal Notes
ual X-
ule as clusters
39
40
S
Chapter 5: Xtr
EMC VPLEX SImplementat
Single EnginQuad X-Brick
remIO Connect
SAN Connecttion Planning
Fig
Thisclus
Thisper bes
Tota
ne / k
tivity
ivity g and Best Pr
gure 15 VPLE
s illustration ster to a quad
s configuratio Storage Volu
st practice of
al Bandwidth
64 Gb/s
ractices Tech
EX single en
demonstrated X-Brick Xtre
on meets theume. This co spreading ac
h VPLEX T
nical Notes
ngine - Xtr
es one possiemIO cluster
e VPLEX best onfiguration across all ava
Total Bandwi
64 Gb/
remIO quad X
bility of connr.
practices of also demonsilable resour
idth XtremIO
/s (128 tota
X-Brick
necting a sin
four paths pstrates a comrces evenly.
(as depicted
l available)
gle engine V
per VPLEX Dirmmon perform
d)
PLEX
rector mance
S
Dual Engine Single X-Bric
Fig
Thisxbri
DueChafounduFibrclushasif yo
Thesavbe e
Tota
/ ck
gure 16 VPLE
s illustration ick XtremIO c
e to the limitaannel ports wr paths per D
u requiremenre Channel psters. The tos not been peou would exp
e advantage tving eight por easier.
al Bandwidth
128 Gb/s
Imp
EX dual engi
demonstrate cluster.
ation of eachwe have to doDirector per Snt of two pathorts on each tal available
erformed comperience any
to going to thrts altogethe
h VPLEX T
plementation
ine - Xtrem
es connectiv
h Storage Conouble up on tStorage Volumhs per Directo VPLEX Direc bandwidth w
mparing these difference o
he minimum r on the fabri
Total Bandwi
32 Gb
n Planning an
mIO single X
ity of a dual
ntroller for th the paths frome. You couor per Storag
ctor and creat would remaine two possibr not in overa
configurationics. Future s
idth XtremIO
b/s
Chapter 5
EMC VPLnd Best Pract
X-Brick
engine VPLEX
he X-Brick havm VPLEX in old drop to th
ge Volume thte a 1:1 port n the same.
ble configuratall performan
n would be inscale of the X
(as depicted
5: XtremIO Con
LEX SAN Contices Technic
X cluster to a
ving only two order to achie
e minimum Vereby only u mapping be Performancetions so we cnce.
n that you woXtremIO woul
d)
nnectivity
4nectivity cal Notes
a single
o Fibre eve the
VPLEX sing two tween e testing
can’t say
ould be d also
41
42
X
Chapter 5: Xtr
EMC VPLEX SImplementat
Dual Engine X-Brick
remIO Connect
SAN Connecttion Planning
Fig
Thisa du
Thisbas1:1 of tinit
Tota
/ Dual
tivity
ivity g and Best Pr
gure 17 VPLE
s illustration ual X-Brick Xt
s illustration sis. Based o basis therebhe XtremIO cial deployme
al Bandwidth
128 Gb/s
ractices Tech
EX dual engi
demonstratetremIO cluste
is mapping tn total availa
by leaving twcluster. This ent and still g
h VPLEX T
nical Notes
ine - Xtrem
es connectiver.
the VPLEX baable bandwid
wo ports avail would reduc
give the same
Total Bandwi
64 Gb
mIO dual X-B
ity between a
ackend portsdth, you maylable on eachce the port coe maximum t
idth XtremIO
b/s
Brick
a dual engin
s to the Xtremy consider mah VPLEX Direount on the fa throughput.
(as depicte
e VPLEX clus
mIO ports on apping the pctor for futurabrics at time
d)
ster and
a 2:1 orts on a
re scale e of
Dual Engine Quad X-Brick
Fig
Thisqua
Thisrecoallo
Tota
/ k
gure 18 VPLE
s illustration ad X-Brick clu
s example foommended f
owing for max
al Bandwidth
128 Gb/s
Imp
EX dual engi
demonstrateuster.
llows the fou for performanximum throu
h VPLEX T
s
plementation
ine - Xtrem
es a VPLEX d
ur paths per Vnce. The portghput using
Total Bandwi
128 Gb
n Planning an
mIO quad X-B
ual engine c
VPLEX Directrts are mappe all available
idth XtremIO
b/s
Chapter 5
EMC VPLnd Best Pract
Brick
luster conne
tor per Storaged 1:1 betwe
e ports on bot
(as depicted
5: XtremIO Con
LEX SAN Contices Technic
ctivity to an X
ge Volume rueen the two cth clusters.
d)
nnectivity
4nectivity cal Notes
XtremIO
ule as clusters
43
44
S
Chapter 5: Xtr
EMC VPLEX SImplementat
Quad EngineSingle X-Bric
remIO Connect
SAN Connecttion Planning
Fig
ThisXtre
Thispatdesof twporthebanclus
Tota
e / ck
tivity
ivity g and Best Pr
gure 19 VPLE
s illustration emIO single X
s configuratiohs per Direct
sirable from awo paths per
rt count requise ports avai
ndwidth is dister in this co
al Bandwidth
256 Gb/s
ractices Tech
EX quad engi
demonstrate X-Brick cluste
on meets thetor per Storag
a cost point or Director perired on the failable for futuctated by theonfiguration.
h VPLEX T
nical Notes
ine - Xtrem
es connectiver.
e VPLEX backge Volume. T
of view. Redur Storage Volabrics by a toure scale on te throughput
Total Bandwi
32 Gb/
mIO single X
ity of a VPLEX
kend connectThe ports areucing to a milume would rotal of 16 por the XtremIO t of the availa
idth XtremIO
/s
X-Brick
X quad engin
tivity rule rece mapped 8:1inimum VPLE
reduce the corts overall. T cluster. Theable four por
(as depicted
ne cluster to
commending1 and may noEX ndu requirost by reduciThis would al total availab
rts on the Xtre
d)
an
g four ot be rement ing the so leave
ble emIO
Quad EngineDual X-Brick
Fig
Thisdua
Thispatdesof twporthe
Tota
e / k
gure 20 VPLE
s illustration al X-Brick Xtre
s configuratiohs per Direct
sirable from awo paths per
rt count requise ports avai
al Bandwidth
256 Gb/s
Imp
EX quad engi
demonstrateemIO cluster
on meets thetor per Storag
a cost point or Director perired on the failable for futu
h VPLEX T
plementation
ine - Xtrem
es the conne.
e VPLEX backge Volume. T
of view. Redur Storage Volabrics by a toure scale on t
Total Bandwi
64 Gb/
n Planning an
mIO dual X-B
ectivity of a q
kend connect The ports areucing to a milume would rotal of 16 por the XtremIO
idth XtremIO
/s
Chapter 5
EMC VPLnd Best Pract
Brick
uad engine V
tivity rule rece mapped 4:1inimum VPLE
reduce the corts overall. T cluster.
(as depicted
5: XtremIO Con
LEX SAN Contices Technic
VPLEX cluste
commending 1 and may noEX ndu requirost by reduciThis would al
d)
nnectivity
4nectivity cal Notes
r to a
g four ot be rement ing the so leave
45
46
Chapter 5: Xtr
EMC VPLEX SImplementat
Quad EngineQuad X-Brick
remIO Connect
SAN Connecttion Planning
Figbot
e / k
tivity
ivity g and Best Pr
gure 21 VPLEttom)
ractices Tech
EX quad engi
nical Notes
ine - Xtrem
mIO quad X-BBrick (Fabrric A top Fa
abric B
Thequabrocon
ThisStominalsoon t
Tota
e previous twad engine VPken up into t
nnectivity on
s configuratiorage Volume
nimal configuo allows add the fabrics.
al Bandwidth
256 Gb/s
Imp
o illustrationLEX cluster a
two parts sho the bottom.
on meets thee. The ports auration requiring future ca
h VPLEX T
plementation
ns combined and a quad Xowing Fabric
e VPLEX best are mapped orements on V
apacity on the
Total Bandwi
128 Gb
n Planning an
demonstrateX-Brick Xtrem connectivity
practices for on a 2:1 bas VPLEX as reqe XtremIO as
idth XtremIO
b/s
Chapter 5
EMC VPLnd Best Pract
e the connecIO cluster. T
y on the top a
r four paths pis which allouired by the well as redu
(as depicted
5: XtremIO Con
LEX SAN Contices Technic
ctivity betweeThe illustratioand Fabric B
per Director pows dropping ndu pre-cheucing the por
d)
nnectivity
4nectivity cal Notes
en a ons are
per g to the ck. This
rt count
47