Difference between revisions of "Flying zones sol11"
From pressy's brainbackup
(Created page with "== Flying Zones in sol11 running OSC4.1 == <pre> root@clnode01:~# pkg list -a ha-cluster/data-service/ha-zones NAME (PUBLISHER) VERSION ...") |
|||
Line 1: | Line 1: | ||
== Flying Zones in sol11 running OSC4.1 == | == Flying Zones in sol11 running OSC4.1 == | ||
+ | === configure the environment === | ||
+ | <pre> | ||
+ | root@clnode01:~# cldevice status | ||
+ | === Cluster DID Devices === | ||
+ | |||
+ | Device Instance Node Status | ||
+ | --------------- ---- ------ | ||
+ | /dev/did/rdsk/d1 clnode01 Ok | ||
+ | clnode02 Ok | ||
+ | |||
+ | /dev/did/rdsk/d2 clnode01 Unmonitored | ||
+ | |||
+ | /dev/did/rdsk/d4 clnode02 Unmonitored | ||
+ | |||
+ | root@clnode01:~# | ||
+ | |||
+ | root@clnode01:~# clrg create ha-group01 | ||
+ | root@clnode01:~# clrt register SUNW.HAStoragePlus | ||
+ | root@clnode02:~# zpool import | ||
+ | pool: ha-data | ||
+ | id: 3768387658467332306 | ||
+ | state: ONLINE | ||
+ | action: The pool can be imported using its name or numeric identifier. | ||
+ | config: | ||
+ | |||
+ | ha-data ONLINE | ||
+ | c0t600144F0DE794B00000053A403AA0001d0 ONLINE | ||
+ | |||
+ | root@clnode01:~# clrs create -g ha-group01 -t SUNW.HAStoragePlus -p Zpools=ha-data ha-data-rs | ||
+ | root@clnode01:~# clrg online -M -n clnode01 ha-group01 | ||
+ | |||
+ | root@clnode01:~# cluster status -t rg,rs | ||
+ | |||
+ | === Cluster Resource Groups === | ||
+ | |||
+ | Group Name Node Name Suspended State | ||
+ | ---------- --------- --------- ----- | ||
+ | ha-group01 clnode01 No Online | ||
+ | clnode02 No Offline | ||
+ | |||
+ | |||
+ | === Cluster Resources === | ||
+ | |||
+ | Resource Name Node Name State Status Message | ||
+ | ------------- --------- ----- -------------- | ||
+ | ha-data-rs clnode01 Online Online | ||
+ | clnode02 Offline Offline | ||
+ | root@clnode01:~# clrg switch -n clnode02 ha-group01 | ||
+ | root@clnode01:~# cluster status -t rg,rs | ||
+ | |||
+ | === Cluster Resource Groups === | ||
+ | |||
+ | Group Name Node Name Suspended State | ||
+ | ---------- --------- --------- ----- | ||
+ | ha-group01 clnode01 No Offline | ||
+ | clnode02 No Online | ||
+ | |||
+ | |||
+ | === Cluster Resources === | ||
+ | |||
+ | Resource Name Node Name State Status Message | ||
+ | ------------- --------- ----- -------------- | ||
+ | ha-data-rs clnode01 Offline Offline | ||
+ | clnode02 Online Online | ||
+ | |||
+ | root@clnode01:~# | ||
+ | root@clnode01:~# clrg remaster + | ||
+ | root@clnode01:~# df -h | tail -1 | ||
+ | ha-data 20G 31K 20G 1% /ha-data | ||
+ | root@clnode01:~# | ||
+ | </pre> | ||
+ | |||
+ | === the zone === | ||
+ | <pre> | ||
+ | root@clnode01:~# more /etc/hosts | ||
+ | # | ||
+ | # Copyright 2009 Sun Microsystems, Inc. All rights reserved. | ||
+ | # Use is subject to license terms. | ||
+ | # | ||
+ | # Internet host table | ||
+ | # | ||
+ | ::1 localhost | ||
+ | 127.0.0.1 localhost loghost | ||
+ | 192.168.56.200 clnode01 clnode01.local | ||
+ | 192.168.56.201 clnode02 clnode02.local | ||
+ | #### | ||
+ | 192.168.56.210 clzone01 clzone01.local | ||
+ | #### | ||
+ | 192.168.56.100 vsol01 vsol01.local | ||
+ | root@clnode01:~# | ||
+ | root@clnode01:~# clreslogicalhostname create -g ha-group01 -h clzone01 clzone01-lh-rs | ||
+ | root@clnode01:~# | ||
+ | root@clnode01:~# clrs status | ||
+ | |||
+ | === Cluster Resources === | ||
+ | |||
+ | Resource Name Node Name State Status Message | ||
+ | ------------- --------- ----- -------------- | ||
+ | clzone01-lh-rs clnode01 Online Online - LogicalHostname online. | ||
+ | clnode02 Offline Offline | ||
+ | |||
+ | ha-data-rs clnode01 Online Online | ||
+ | clnode02 Offline Offline | ||
+ | |||
+ | root@clnode01:~# ping clzone01 | ||
+ | clzone01 is alive | ||
+ | root@clnode01:~# | ||
+ | root@clnode01:~# zonecfg -z clzone01 | ||
+ | Use 'create' to begin configuring a new zone. | ||
+ | zonecfg:clzone01> create -b | ||
+ | zonecfg:clzone01> set zonepath=/ha-zones/clzone01 | ||
+ | zonecfg:clzone01> set autoboot=false | ||
+ | zonecfg:clzone01> set ip-type=shared | ||
+ | zonecfg:clnode01> add attr | ||
+ | zonecfg:clnode01:attr> set name=osc-ha-zone | ||
+ | zonecfg:clnode01:attr> set type=boolean | ||
+ | zonecfg:clnode01:attr> set value=true | ||
+ | zonecfg:clnode01:attr> end | ||
+ | zonecfg:clzone01> verify | ||
+ | zonecfg:clzone01> commit | ||
+ | zonecfg:clzone01> exit | ||
+ | root@clnode01:~# | ||
+ | root@clnode01:~# | ||
+ | root@clnode01:~# | ||
+ | root@clnode01:~# zoneadm list -cv | ||
+ | ID NAME STATUS PATH BRAND IP | ||
+ | 0 global running / solaris shared | ||
+ | - clzone01 configured /ha-zones/clzone01 solaris shared | ||
+ | root@clnode01:~# zoneadm -z clzone01 install | ||
+ | The following ZFS file system(s) have been created: | ||
+ | ha-data/ha-zones/clzone01 | ||
+ | Progress being logged to /var/log/zones/zoneadm.20140623T122141Z.clzone01.install | ||
+ | Image: Preparing at /ha-zones/clzone01/root. | ||
+ | |||
+ | AI Manifest: /tmp/manifest.xml.Qqaqbg | ||
+ | SC Profile: /usr/share/auto_install/sc_profiles/enable_sci.xml | ||
+ | Zonename: clzone01 | ||
+ | Installation: Starting ... | ||
+ | |||
+ | Creating IPS image | ||
+ | Startup linked: 1/1 done | ||
+ | Installing packages from: | ||
+ | solaris | ||
+ | origin: http://192.168.56.101/ | ||
+ | ha-cluster | ||
+ | origin: http://192.168.56.101:8080/ | ||
+ | DOWNLOAD PKGS FILES XFER (MB) SPEED | ||
+ | Completed 187/187 34366/34366 231.6/231.6 485k/s | ||
+ | |||
+ | PHASE ITEMS | ||
+ | Installing new actions 48351/48351 | ||
+ | Updating package state database Done | ||
+ | Updating image state Done | ||
+ | Creating fast lookup database Done | ||
+ | Installation: Succeeded | ||
+ | |||
+ | Note: Man pages can be obtained by installing pkg:/system/manual | ||
+ | |||
+ | done. | ||
+ | |||
+ | Done: Installation completed in 783.353 seconds. | ||
+ | |||
+ | |||
+ | Next Steps: Boot the zone, then log into the zone console (zlogin -C) | ||
+ | |||
+ | to complete the configuration process. | ||
+ | |||
+ | Log saved in non-global zone as /ha-zones/clzone01/root/var/log/zones/zoneadm.20140623T122141Z.clzone01.install | ||
+ | root@clnode01:~# | ||
+ | root@clnode01:~# zoneadm list -cv | ||
+ | ID NAME STATUS PATH BRAND IP | ||
+ | 0 global running / solaris shared | ||
+ | - clzone01 installed /ha-zones/clzone01 solaris shared | ||
+ | root@clnode01:~# zoneadm -z clzone01 boot | ||
+ | root@clnode01:~# zlogin -C clzone01 | ||
+ | [snip] | ||
+ | root@clnode01:~# zoneadm list -cv | ||
+ | ID NAME STATUS PATH BRAND IP | ||
+ | 0 global running / solaris shared | ||
+ | 1 clzone01 running /ha-zones/clzone01 solaris shared | ||
+ | root@clnode01:~# zoneadm -z clzone01 shutdown | ||
+ | root@clnode01:~# | ||
+ | root@clnode01:~# zoneadm -z clzone01 detach -F | ||
+ | root@clnode01:~# | ||
+ | root@clnode01:~# zoneadm list -cv | ||
+ | ID NAME STATUS PATH BRAND IP | ||
+ | 0 global running / solaris shared | ||
+ | - clzone01 configured /ha-zones/clzone01 solaris shared | ||
+ | root@clnode01:~# | ||
+ | root@clnode01:~# beadm list -H | ||
+ | non-sc;0f5f7660-63ad-6a84-ccd0-8bbea3d9ec7c;;;66463744;static;1400556054 | ||
+ | non-sc-backup-1;51da94ac-2d1b-c9d1-c202-e3a125480b3f;;;68608;static;1403282602 | ||
+ | s11.1.19.0.6.0;5f5a858a-e2e7-69ef-897c-8bb96173bd55;NR;/;7785575936;static;1403532512 | ||
+ | solaris;65176f0c-2c86-642d-ec65-932ecb4f755a;;;3269120;static;1399981144 | ||
+ | root@clnode01:~# | ||
+ | root@clnode01:~# df -b / | ||
+ | Filesystem avail | ||
+ | rpool/ROOT/s11.1.19.0.6.0 12539479 | ||
+ | root@clnode01:~# zfs get org.opensolaris.libbe:uuid rpool/ROOT/s11.1.19.0.6.0 | ||
+ | NAME PROPERTY VALUE SOURCE | ||
+ | rpool/ROOT/s11.1.19.0.6.0 org.opensolaris.libbe:uuid 5f5a858a-e2e7-69ef-897c-8bb96173bd55 local | ||
+ | root@clnode02:~# zfs get org.opensolaris.libbe:uuid rpool/ROOT/s11.1.19.0.6.0 | ||
+ | NAME PROPERTY VALUE SOURCE | ||
+ | rpool/ROOT/s11.1.19.0.6.0 org.opensolaris.libbe:uuid 003c22b8-be3a-eb9f-be05-8385af25e5b6 local | ||
+ | root@clnode02:~# | ||
+ | root@clnode02:~# zfs set org.opensolaris.libbe:uuid=5f5a858a-e2e7-69ef-897c-8bb96173bd55 rpool/ROOT/s11.1.19.0.6.0 | ||
+ | root@clnode02:~# zfs get org.opensolaris.libbe:uuid rpool/ROOT/s11.1.19.0.6.0 | ||
+ | NAME PROPERTY VALUE SOURCE | ||
+ | rpool/ROOT/s11.1.19.0.6.0 org.opensolaris.libbe:uuid 5f5a858a-e2e7-69ef-897c-8bb96173bd55 local | ||
+ | root@clnode02:~# beadm list -H | ||
+ | non-sc;13d2053a-c6b1-ce44-85cb-daf36b2dbfa0;;;154624;static;1400556067 | ||
+ | s11.1.19.0.6.0;5f5a858a-e2e7-69ef-897c-8bb96173bd55;NR;/;6764256256;static;1400550024 | ||
+ | s11.1.19.0.6.0-backup-1;d640f3d4-a023-6bb3-fecf-81029189443c;;;68608;static;1403282566 | ||
+ | solaris;65176f0c-2c86-642d-ec65-932ecb4f755a;;;3840512;static;1399981144 | ||
+ | root@clnode02:~# | ||
+ | root@clnode01:~# clrg switch -n clnode02 ha-group01 | ||
+ | root@clnode01:~# clrg status | ||
+ | |||
+ | === Cluster Resource Groups === | ||
+ | |||
+ | Group Name Node Name Suspended Status | ||
+ | ---------- --------- --------- ------ | ||
+ | ha-group01 clnode01 No Offline | ||
+ | clnode02 No Online | ||
+ | root@clnode02:~# zoneadm list -cv | ||
+ | ID NAME STATUS PATH BRAND IP | ||
+ | 0 global running / solaris shared | ||
+ | - clnode01 configured /ha-zones/clzone01 solaris shared | ||
+ | root@clnode02:~# zoneadm -z clnode01 attach -F | ||
+ | root@clnode02:~# zoneadm list -cv | ||
+ | ID NAME STATUS PATH BRAND IP | ||
+ | 0 global running / solaris shared | ||
+ | - clnode01 installed /ha-zones/clzone01 solaris shared | ||
+ | root@clnode02:~# | ||
+ | root@clnode02:~# zoneadm -z clnode01 boot | ||
+ | root@clnode02:~# zoneadm list -cv | ||
+ | ID NAME STATUS PATH BRAND IP | ||
+ | 0 global running / solaris shared | ||
+ | 1 clnode01 running /ha-zones/clzone01 solaris shared | ||
+ | root@clnode02:~# | ||
+ | </pre> | ||
+ | |||
+ | === Zone under Cluster Control === | ||
<pre> | <pre> | ||
root@clnode01:~# pkg list -a ha-cluster/data-service/ha-zones | root@clnode01:~# pkg list -a ha-cluster/data-service/ha-zones |
Revision as of 19:06, 6 August 2014
Contents
Flying Zones in sol11 running OSC4.1
configure the environment
root@clnode01:~# cldevice status === Cluster DID Devices === Device Instance Node Status --------------- ---- ------ /dev/did/rdsk/d1 clnode01 Ok clnode02 Ok /dev/did/rdsk/d2 clnode01 Unmonitored /dev/did/rdsk/d4 clnode02 Unmonitored root@clnode01:~# root@clnode01:~# clrg create ha-group01 root@clnode01:~# clrt register SUNW.HAStoragePlus root@clnode02:~# zpool import pool: ha-data id: 3768387658467332306 state: ONLINE action: The pool can be imported using its name or numeric identifier. config: ha-data ONLINE c0t600144F0DE794B00000053A403AA0001d0 ONLINE root@clnode01:~# clrs create -g ha-group01 -t SUNW.HAStoragePlus -p Zpools=ha-data ha-data-rs root@clnode01:~# clrg online -M -n clnode01 ha-group01 root@clnode01:~# cluster status -t rg,rs === Cluster Resource Groups === Group Name Node Name Suspended State ---------- --------- --------- ----- ha-group01 clnode01 No Online clnode02 No Offline === Cluster Resources === Resource Name Node Name State Status Message ------------- --------- ----- -------------- ha-data-rs clnode01 Online Online clnode02 Offline Offline root@clnode01:~# clrg switch -n clnode02 ha-group01 root@clnode01:~# cluster status -t rg,rs === Cluster Resource Groups === Group Name Node Name Suspended State ---------- --------- --------- ----- ha-group01 clnode01 No Offline clnode02 No Online === Cluster Resources === Resource Name Node Name State Status Message ------------- --------- ----- -------------- ha-data-rs clnode01 Offline Offline clnode02 Online Online root@clnode01:~# root@clnode01:~# clrg remaster + root@clnode01:~# df -h | tail -1 ha-data 20G 31K 20G 1% /ha-data root@clnode01:~#
the zone
root@clnode01:~# more /etc/hosts # # Copyright 2009 Sun Microsystems, Inc. All rights reserved. # Use is subject to license terms. # # Internet host table # ::1 localhost 127.0.0.1 localhost loghost 192.168.56.200 clnode01 clnode01.local 192.168.56.201 clnode02 clnode02.local #### 192.168.56.210 clzone01 clzone01.local #### 192.168.56.100 vsol01 vsol01.local root@clnode01:~# root@clnode01:~# clreslogicalhostname create -g ha-group01 -h clzone01 clzone01-lh-rs root@clnode01:~# root@clnode01:~# clrs status === Cluster Resources === Resource Name Node Name State Status Message ------------- --------- ----- -------------- clzone01-lh-rs clnode01 Online Online - LogicalHostname online. clnode02 Offline Offline ha-data-rs clnode01 Online Online clnode02 Offline Offline root@clnode01:~# ping clzone01 clzone01 is alive root@clnode01:~# root@clnode01:~# zonecfg -z clzone01 Use 'create' to begin configuring a new zone. zonecfg:clzone01> create -b zonecfg:clzone01> set zonepath=/ha-zones/clzone01 zonecfg:clzone01> set autoboot=false zonecfg:clzone01> set ip-type=shared zonecfg:clnode01> add attr zonecfg:clnode01:attr> set name=osc-ha-zone zonecfg:clnode01:attr> set type=boolean zonecfg:clnode01:attr> set value=true zonecfg:clnode01:attr> end zonecfg:clzone01> verify zonecfg:clzone01> commit zonecfg:clzone01> exit root@clnode01:~# root@clnode01:~# root@clnode01:~# root@clnode01:~# zoneadm list -cv ID NAME STATUS PATH BRAND IP 0 global running / solaris shared - clzone01 configured /ha-zones/clzone01 solaris shared root@clnode01:~# zoneadm -z clzone01 install The following ZFS file system(s) have been created: ha-data/ha-zones/clzone01 Progress being logged to /var/log/zones/zoneadm.20140623T122141Z.clzone01.install Image: Preparing at /ha-zones/clzone01/root. AI Manifest: /tmp/manifest.xml.Qqaqbg SC Profile: /usr/share/auto_install/sc_profiles/enable_sci.xml Zonename: clzone01 Installation: Starting ... Creating IPS image Startup linked: 1/1 done Installing packages from: solaris origin: http://192.168.56.101/ ha-cluster origin: http://192.168.56.101:8080/ DOWNLOAD PKGS FILES XFER (MB) SPEED Completed 187/187 34366/34366 231.6/231.6 485k/s PHASE ITEMS Installing new actions 48351/48351 Updating package state database Done Updating image state Done Creating fast lookup database Done Installation: Succeeded Note: Man pages can be obtained by installing pkg:/system/manual done. Done: Installation completed in 783.353 seconds. Next Steps: Boot the zone, then log into the zone console (zlogin -C) to complete the configuration process. Log saved in non-global zone as /ha-zones/clzone01/root/var/log/zones/zoneadm.20140623T122141Z.clzone01.install root@clnode01:~# root@clnode01:~# zoneadm list -cv ID NAME STATUS PATH BRAND IP 0 global running / solaris shared - clzone01 installed /ha-zones/clzone01 solaris shared root@clnode01:~# zoneadm -z clzone01 boot root@clnode01:~# zlogin -C clzone01 [snip] root@clnode01:~# zoneadm list -cv ID NAME STATUS PATH BRAND IP 0 global running / solaris shared 1 clzone01 running /ha-zones/clzone01 solaris shared root@clnode01:~# zoneadm -z clzone01 shutdown root@clnode01:~# root@clnode01:~# zoneadm -z clzone01 detach -F root@clnode01:~# root@clnode01:~# zoneadm list -cv ID NAME STATUS PATH BRAND IP 0 global running / solaris shared - clzone01 configured /ha-zones/clzone01 solaris shared root@clnode01:~# root@clnode01:~# beadm list -H non-sc;0f5f7660-63ad-6a84-ccd0-8bbea3d9ec7c;;;66463744;static;1400556054 non-sc-backup-1;51da94ac-2d1b-c9d1-c202-e3a125480b3f;;;68608;static;1403282602 s11.1.19.0.6.0;5f5a858a-e2e7-69ef-897c-8bb96173bd55;NR;/;7785575936;static;1403532512 solaris;65176f0c-2c86-642d-ec65-932ecb4f755a;;;3269120;static;1399981144 root@clnode01:~# root@clnode01:~# df -b / Filesystem avail rpool/ROOT/s11.1.19.0.6.0 12539479 root@clnode01:~# zfs get org.opensolaris.libbe:uuid rpool/ROOT/s11.1.19.0.6.0 NAME PROPERTY VALUE SOURCE rpool/ROOT/s11.1.19.0.6.0 org.opensolaris.libbe:uuid 5f5a858a-e2e7-69ef-897c-8bb96173bd55 local root@clnode02:~# zfs get org.opensolaris.libbe:uuid rpool/ROOT/s11.1.19.0.6.0 NAME PROPERTY VALUE SOURCE rpool/ROOT/s11.1.19.0.6.0 org.opensolaris.libbe:uuid 003c22b8-be3a-eb9f-be05-8385af25e5b6 local root@clnode02:~# root@clnode02:~# zfs set org.opensolaris.libbe:uuid=5f5a858a-e2e7-69ef-897c-8bb96173bd55 rpool/ROOT/s11.1.19.0.6.0 root@clnode02:~# zfs get org.opensolaris.libbe:uuid rpool/ROOT/s11.1.19.0.6.0 NAME PROPERTY VALUE SOURCE rpool/ROOT/s11.1.19.0.6.0 org.opensolaris.libbe:uuid 5f5a858a-e2e7-69ef-897c-8bb96173bd55 local root@clnode02:~# beadm list -H non-sc;13d2053a-c6b1-ce44-85cb-daf36b2dbfa0;;;154624;static;1400556067 s11.1.19.0.6.0;5f5a858a-e2e7-69ef-897c-8bb96173bd55;NR;/;6764256256;static;1400550024 s11.1.19.0.6.0-backup-1;d640f3d4-a023-6bb3-fecf-81029189443c;;;68608;static;1403282566 solaris;65176f0c-2c86-642d-ec65-932ecb4f755a;;;3840512;static;1399981144 root@clnode02:~# root@clnode01:~# clrg switch -n clnode02 ha-group01 root@clnode01:~# clrg status === Cluster Resource Groups === Group Name Node Name Suspended Status ---------- --------- --------- ------ ha-group01 clnode01 No Offline clnode02 No Online root@clnode02:~# zoneadm list -cv ID NAME STATUS PATH BRAND IP 0 global running / solaris shared - clnode01 configured /ha-zones/clzone01 solaris shared root@clnode02:~# zoneadm -z clnode01 attach -F root@clnode02:~# zoneadm list -cv ID NAME STATUS PATH BRAND IP 0 global running / solaris shared - clnode01 installed /ha-zones/clzone01 solaris shared root@clnode02:~# root@clnode02:~# zoneadm -z clnode01 boot root@clnode02:~# zoneadm list -cv ID NAME STATUS PATH BRAND IP 0 global running / solaris shared 1 clnode01 running /ha-zones/clzone01 solaris shared root@clnode02:~#
Zone under Cluster Control
root@clnode01:~# pkg list -a ha-cluster/data-service/ha-zones NAME (PUBLISHER) VERSION IFO ha-cluster/data-service/ha-zones (ha-cluster) 4.1-6.1 i-- root@clnode01:~# cd /opt/SUNWsczone/sczbt/util root@clnode01:/opt/SUNWsczone/sczbt/util# ls sczbt_config sczbt_register root@clnode01:/opt/SUNWsczone/sczbt/util# root@clnode01:/opt/SUNWsczone/sczbt/util# root@clnode01:/opt/SUNWsczone/sczbt/util# root@clnode01:/opt/SUNWsczone/sczbt/util# cp -p sczbt_config sczbt_config_clzone01-res root@clnode01:/opt/SUNWsczone/sczbt/util# grep -v "^#" sczbt_config > sczbt_config_clzone01-res root@clnode01:/opt/SUNWsczone/sczbt/util# vi sczbt_config_clzone01-res root@clnode01:/opt/SUNWsczone/sczbt/util# cat sczbt_config_clzone01-res RS=clnode01-zone-rs RG=ha-group01 PARAMETERDIR=/ha-zones/clzone01/cl-params SC_NETWORK=true SC_LH=clzone01-lh-rs FAILOVER=true HAS_RS=ha-data-rs Zonename="clzone01" Zonebrand="solaris" Zonebootopt="" Milestone="svc:/milestone/multi-user-server" LXrunlevel="3" SLrunlevel="3" Mounts="" root@clnode01:/opt/SUNWsczone/sczbt/util# zoneadm -z clzone01 attach -F root@clnode01:/opt/SUNWsczone/sczbt/util# ./sczbt_register -f ./sczbt_config_clzone01-res sourcing ./sczbt_config_clzone01-res Registration of resource clnode01-zone-rs succeeded. Validation of resource clnode01-zone-rs succeeded. root@clnode01:/opt/SUNWsczone/sczbt/util# root@clnode01:/# clrs list -g ha-group01 clnode01-zone-rs clzone01-lh-rs ha-data-rs root@clnode01:/# root@clnode01:/# clrs enable clnode01-zone-rs root@clnode01:/# root@clnode01:/# root@clnode01:/# clrs status === Cluster Resources === Resource Name Node Name State Status Message ------------- --------- ----- -------------- clnode01-zone-rs clnode01 Online Online clnode02 Offline Offline clzone01-lh-rs clnode01 Online Online - LogicalHostname online. clnode02 Offline Offline - LogicalHostname offline. ha-data-rs clnode01 Online Online clnode02 Offline Offline root@clnode01:/# root@clnode01:/# root@clnode01:/# zoneadm list -cv ID NAME STATUS PATH BRAND IP 0 global running / solaris shared 2 clzone01 running /ha-zones/clzone01 solaris shared root@clnode01:/# root@clnode01:/# clrg switch -n clnode02 ha-group01