the new disks:
2. c0t2d0 <SUN36G cyl 24620 alt 2 hd 27 sec 107> rootdisk
/pci@1c,600000/scsi@2/sd@2,0
3. c0t3d0 <SUN36G cyl 24620 alt 2 hd 27 sec 107>
/pci@1c,600000/scsi@2/sd@3,0
first, copy volume table of content from original mirrored disk:
prtvtoc /dev/rdsk/c0t0d0s2 | fmthard -s - /dev/rdsk/c0t2d0s2
prtvtoc /dev/rdsk/c0t0d0s2 | fmthard -s - /dev/rdsk/c0t3d0s2
init partitions acting as metadevices:
metainit d30 1 1 c0t2d0s0
metainit d40 1 1 c0t3d0s0
metainit d31 1 1 c0t2d0s1
metainit d41 1 1 c0t3d0s1
metainit d35 1 1 c0t2d0s5
metainit d45 1 1 c0t3d0s5
metainit d36 1 1 c0t2d0s6
metainit d46 1 1 c0t3d0s6
metainit d37 1 1 c0t2d0s7
metainit d47 1 1 c0t3d0s7
attach devices to current mirror:
metattach d0 d30
metattach d1 d31
metattach d5 d35
metattach d6 d36
metattach d7 d37
metattach d0 d40
metattach d1 d41
metattach d5 d45
metattach d6 d46
metattach d7 d47
if metainit says "Metainit : no such file or directory", the limit defined for metainit number ( metainit d_number_ 1 1 c0t5d0s7) was higher than the limit _number_ allowed in the system configuration. The configuration file /kernel/drv/md.conf had to be changed regading the number you are willing to use ("nmd" parameter from 128 to your intended number). Changing kernel driver files requires a reboot.
[ add comment ] ( 2 views ) | [ 0 trackbacks ] | permalink
Path to scadm: /usr/platform/SUNW,Sun-Fire-V240/sbin
# ./scadm help
USAGE: scadm <command> [options]
For a list of commands, type "scadm help"
scadm - COMMANDS SUPPORTED
help, date, set, show, resetrsc, download, send_event, modem_setup,
useradd, userdel, usershow, userpassword, userperm, shownetwork,
loghistory, version
scadm - COMMAND DETAILS
scadm help => this message
scadm date [-s] | [[mmdd]HHMM | mmddHHMM[cc]yy][.SS] => print or set date
scadm set <variable> <value> => set variable to value
scadm show [variable] => show variable(s)
scadm resetrsc [-s] => reset SC (-s soft reset)
scadm download [boot] <file> => program firmware or [boot] monitor
scadm send_event [-c] "message" => send message as event (-c CRITICAL)
scadm modem_setup => connect to modem port
scadm useradd <username> => add SC user account
scadm userdel <username> => delete SC user account
scadm usershow [username] => show user details
scadm userpassword <username> => set user password
scadm userperm <username> [cuar] => set user permissions
scadm shownetwork => show network configuration
scadm loghistory => show SC event log
scadm version [-v] => show SC version (-v verbose)
[ add comment ] ( 2 views ) | [ 0 trackbacks ] | permalink
# rm /var/log/messages
# lsof -n -P -l | grep syslog
syslogd 6442 0 txt REG 3,6 35832 426222 /sbin/syslogd
# cd /proc/6442/fd
[root@vm02 fd]# ls -la
total 0
dr-x------ 2 root root 0 Apr 24 04:31 .
dr-xr-xr-x 5 root root 0 Nov 7 04:41 ..
lrwx------ 1 root root 64 Apr 24 04:31 0 -> socket:[59016]
l-wx------ 1 root root 64 Apr 24 04:31 1 -> /var/log/messages (deleted)
l-wx------ 1 root root 64 Apr 24 04:31 2 -> /var/log/secure
l-wx------ 1 root root 64 Apr 24 04:31 3 -> /var/log/maillog
l-wx------ 1 root root 64 Apr 24 04:31 4 -> /var/log/cron
l-wx------ 1 root root 64 Apr 24 04:31 5 -> /var/log/spooler
l-wx------ 1 root root 64 Apr 24 04:31 6 -> /var/log/boot.log
# cat 1
...
...
Dec 4 23:29:34 vm02 yum-updatesd-helper: error getting update info: rpmdb open failed
Dec 5 00:29:42 vm02 yum-updatesd-helper: error getting update info: rpmdb open failed
[ add comment ] ( 7 views ) | [ 0 trackbacks ] | permalink
[ add comment ] ( 5 views ) | [ 0 trackbacks ] | permalink
[ add comment ] ( 3 views ) | [ 0 trackbacks ] | permalink
Other nrpe than nrpe_nt is available as Windows service to connect Nagios monitoring with Windows servers. More info is at: http://www.itefix.no/i2/winrpe.
[ add comment ] ( 7 views ) | [ 0 trackbacks ] | permalink
To create 180GB disk for VM issue:
dd if=/dev/zero of=space.img bs=1024 count=0 seek=$[1024*1000*180]
or
dd if=/dev/zero of=space.tmp bs=1G count=180
[ add comment ] ( 6 views ) | [ 0 trackbacks ] | permalink
[root@cl1 ~]# crm_resource -L
resource_system2 (heartbeat::ocf:Xen)
resource_system3 (heartbeat::ocf:Xen)
resource_system4 (heartbeat::ocf:Xen)
[root@cl1 ~]# crm_mon
Defaulting to one-shot mode
You need to have curses available at compile time to enable console mode
============
Last updated: Tue Jul 7 11:33:27 2009
Current DC: cl1 (a9972101-4a50-4cb7-8dd6-91844a5d23e5)
2 Nodes configured.
3 Resources configured.
============
Node: cl2 (9f8ae28f-a1c4-43b3-8b13-374cc82d5be6): online
Node: cl1 (a9972101-4a50-4cb7-8dd6-91844a5d23e5): online
resource_system2 (heartbeat::ocf:Xen): Started cl2
resource_system3 (heartbeat::ocf:Xen): Started cl1
resource_system4 (heartbeat::ocf:Xen): Started cl2
[root@cl1 ~]# crm_resource --verbose --migrate --resource resource_system2
[root@cl1 ~]# crm_mon
Defaulting to one-shot mode
You need to have curses available at compile time to enable console mode
============
Last updated: Tue Jul 7 11:36:14 2009
Current DC: cl1 (a9972101-4a50-4cb7-8dd6-91844a5d23e5)
2 Nodes configured.
3 Resources configured.
============
Node: cl2 (9f8ae28f-a1c4-43b3-8b13-374cc82d5be6): online
Node: cl1 (a9972101-4a50-4cb7-8dd6-91844a5d23e5): online
resource_system2 (heartbeat::ocf:Xen): Started cl1
resource_system3 (heartbeat::ocf:Xen): Started cl2
resource_system4 (heartbeat::ocf:Xen): Started cl2
[root@cl1 ~]# crm_resource --un-migrate --resource resource_system2
[root@cl1 ~]# crm_mon
Defaulting to one-shot mode
You need to have curses available at compile time to enable console mode
============
Last updated: Tue Jul 7 11:39:20 2009
Current DC: cl1 (a9972101-4a50-4cb7-8dd6-91844a5d23e5)
2 Nodes configured.
3 Resources configured.
============
Node: cl2 (9f8ae28f-a1c4-43b3-8b13-374cc82d5be6): online
Node: cl1 (a9972101-4a50-4cb7-8dd6-91844a5d23e5): online
resource_system2 (heartbeat::ocf:Xen): Started cl2
resource_system3 (heartbeat::ocf:Xen): Started cl1
resource_system4 (heartbeat::ocf:Xen): Started cl2
[root@cl1 ~]# crm_resource --query-xml --resource resource_system2
resource_system2 (heartbeat::ocf:Xen): Started cl2
raw xml:
<primitive class="ocf" type="Xen" provider="heartbeat" id="resource_system2">
<meta_attributes id="resource_system2_meta_attrs">
<attributes>
<nvpair name="target_role" id="resource_system2_metaattr_target_role" value="started"/>
</attributes>
</meta_attributes>
<instance_attributes id="resource_system2_instance_attrs">
<attributes>
<nvpair id="2ad5bac8-69f0-4443-8ee3-2e13392f9c1d" name="xmfile" value="/etc/xen/system2"/>
<nvpair id="fddcde50-6717-4f10-8126-ecaf562b2a80" name="allow_migrate" value="1"/>
<nvpair id="1f2202e4-3df7-42e6-8d2f-a455908b2793" name="allow_mem_management" value="0"/>
<nvpair id="aea95d53-ae2b-44b2-85f6-e03239fce97d" name="reserved_Dom0_memory" value="512"/>
</attributes>
</instance_attributes>
</primitive>
[ add comment ] ( 5 views ) | [ 0 trackbacks ] | permalink
http://www.nabble.com/adding-a-reposito ... 20485.html
[root@vm01 AFS]# wget http://dl.openafs.org/dl/openafs/1.4.10 ... noarch.rpm
[ add comment ] ( 7 views ) | [ 0 trackbacks ] | permalink
------------------------------------------
bash-2.05# scstat -q
-- Quorum Summary --
Quorum votes possible: 3
Quorum votes needed: 2
Quorum votes present: 3
-- Quorum Votes by Node --
Node Name Present Possible Status
--------- ------- -------- ------
Node votes: node14 1 1 Online
Node votes: node18 1 1 Online
-- Quorum Votes by Device --
Device Name Present Possible Status
----------- ------- -------- ------
Device votes: /dev/did/rdsk/d3s2 1 1 Online
bash-2.05# scinstall -pv
Sun Cluster 3.1u3 for Solaris 9 sparc
SUNWscr: 3.1.0,REV=2003.03.25.13.14, 113801-12 117949-08 115364-10 115571- 03
SUNWscu: 3.1.0,REV=2003.03.25.13.14, 113801-12 117949-08 115364-10 115571- 03
SUNWscsck: 3.1.0,REV=2003.09.10.18.59, 115953-03
SUNWscnm: 3.1.0,REV=2004.10.02.17.53
SUNWscdev: 3.1.0,REV=2003.03.25.13.14, 113801-12 117949-08
SUNWscgds: 3.1.0,REV=2003.03.25.13.14, 115061-01
SUNWscman: 3.1.0,REV=2003.03.25.13.14, 113801-12 117949-08
SUNWscsal: 3.1.0,REV=2003.03.25.13.14, 113801-12 117949-08
SUNWscsam: 3.1.0,REV=2003.03.25.13.14, 113801-12
SUNWscvm: 3.1.0,REV=2003.03.25.13.14, 113801-12
SUNWmdm: 3.1.0,REV=2003.03.25.13.14, 115053-03
SUNWjdmk-runtime: 5.1,REV=34
SUNWjdmk-runtime-jmx: 5.1,REV=34
SUNWcacaocfg: 1.0,REV=25
SUNWcacao: 1.0,REV=25
SUNWscmasa: 3.1.0,REV=2004.10.02.17.53, 118627-01 115364-10
SUNWscva: 3.1.0,REV=2003.03.25.13.14, 115055-01
SUNWscspm: 3.1.0,REV=2004.10.02.17.53
SUNWscspmu: 3.1.0,REV=2004.10.02.17.53, 115364-10
SUNWscspmr: 3.1.0,REV=2004.10.02.17.53
SUNWexplo: 4.3.1,REV=2004.06.25.07.21
------------------------------------------
bash-2.05# lustatus
Boot Environment Is Active Active Can Copy
Name Complete Now On Reboot Delete Status
-------------------------- -------- ------ --------- ------ ----------
s9be yes yes yes no -
s10be yes no no yes -
bash-2.05# cat /etc/lutab
# DO NOT EDIT THIS FILE BY HAND. This file is not a public interface.
# The format and contents of this file are subject to change.
# Any user modification to this file may result in the incorrect
# operation of Live Upgrade.
1:s9be:C:0
1:/:/dev/dsk/c0t1d0s0:1
1:boot-device:/dev/dsk/c0t1d0s0:2
2:s10be:C:0
2:/:/dev/dsk/c0t2d0s0:1
2:boot-device:/dev/dsk/c0t2d0s0:2
-------------------------------------------------------------
bash-2.05# cat /etc/lutab
# DO NOT EDIT THIS FILE BY HAND. This file is not a public interface.
# The format and contents of this file are subject to change.
# Any user modification to this file may result in the incorrect
# operation of Live Upgrade.
1:s9be:C:0
1:/:/dev/dsk/c0t1d0s0:1
1:boot-device:/dev/dsk/c0t1d0s0:2
2:s10be:C:0
2:/:/dev/dsk/c0t2d0s0:1
2:boot-device:/dev/dsk/c0t2d0s0:2
bash-2.05# cd /etc/lu
bash-2.05# cat ICF.*
s9be:-:/dev/dsk/c0t1d0s1:swap:4194828
s9be:/:/dev/dsk/c0t1d0s0:ufs:20974140
s10be:-:/dev/dsk/c0t2d0s1:swap:4194828
s10be:/:/dev/dsk/c0t2d0s0:ufs:20974140
s10be:/global/.devices/node@1:/dev/dsk/c0t1d0s6:ufs:1048707
-------------------------------------------------------------
cd /net/sunray/export/seif_unzipped/opt/ses/software/Solaris10U3/
nohup luupgrade -u -n s10be -s /net/sunray/export/seif_unzipped/opt/ses/software/Solaris10U3/
----------------------------------------------------------------
cd /net/sunray/export/seif_unzipped/opt/ses/software/Solaris10U3/Solaris_10/Tools/Installers/
----------------------------------------------------------------
# scstat -W
-- Cluster Transport Paths --
Endpoint Endpoint Status
-------- -------- ------
Transport path: node14:qfe1 node18:qfe1 Path online
Transport path: node14:qfe2 node18:qfe2 faulted
----------------------------------------------------------------
bash-2.05# cat /etc/cluster/nodeid
1
----------------------------------------------------------------
bash-2.05# cd /usr/cluster/lib/rgm/
bash-2.05# ls
rt rtreg rtsupport
bash-2.05# find .
-------------------------------------------------------------------
bash-2.05# lumount -n s10be /mnt
/mnt
bash-2.05# ./installer -noconsole -nodisplay -altroot /mnt -state /var/tmp/stateFile
bash-2.05# cd ..
bash-2.05# cd Solaris_sparc/
bash-2.05# cd Product/sun_cluster/Solaris_10/Tools/
bash-2.05# ls
defaults dot.release lib locale scinstall
bash-2.05# ./scinstall -R /mnt/ -u update
.
./rt
./rt/haevent
./rt/haevent/event_monitor_check
./rt/haevent/event_monitor_start
./rt/haevent/event_monitor_stop
./rt/haevent/event_probe
./rt/haevent/event_svc_start
./rt/haevent/event_svc_stop
./rt/haevent/event_update
./rt/haevent/event_validate
./rt/hafoip
./rt/hafoip/hafoip_ipmp_callback
./rt/hafoip/hafoip_monitor_check
./rt/hafoip/hafoip_monitor_start
./rt/hafoip/hafoip_monitor_stop
./rt/hafoip/hafoip_prenet_start
./rt/hafoip/hafoip_retry
./rt/hafoip/hafoip_start
./rt/hafoip/hafoip_stop
./rt/hafoip/hafoip_update
./rt/hafoip/hafoip_validate
./rt/hascip
./rt/hascip/hascip_boot
./rt/hascip/hascip_fini
./rt/hascip/hascip_init
./rt/hascip/hascip_ipmp_callback
./rt/hascip/hascip_monitor_check
./rt/hascip/hascip_monitor_start
./rt/hascip/hascip_monitor_stop
./rt/hascip/hascip_prenet_start
./rt/hascip/hascip_retry
./rt/hascip/hascip_start
./rt/hascip/hascip_stop
./rt/hascip/hascip_update
./rt/hascip/hascip_validate
./rt/hastorage
./rt/hastorage/hastorage_monitor_check
./rt/hastorage/hastorage_monitor_start
./rt/hastorage/hastorage_monitor_stop
./rt/hastorage/hastorage_prenet_start
./rt/hastorage/hastorage_start
./rt/hastorage/hastorage_stop
./rt/hastorage/hastorage_validate
./rt/hastorageplus
./rt/hastorageplus/hastorageplus_monitor_check
./rt/hastorageplus/hastorageplus_monitor_start
./rt/hastorageplus/hastorageplus_monitor_stop
./rt/hastorageplus/hastorageplus_postnet_stop
./rt/hastorageplus/hastorageplus_prenet_start
./rt/hastorageplus/hastorageplus_prenet_start_private
./rt/hastorageplus/hastorageplus_start
./rt/hastorageplus/hastorageplus_stop
./rt/hastorageplus/hastorageplus_update
./rt/hastorageplus/hastorageplus_update_private
./rt/hastorageplus/hastorageplus_validate
./rt/hastorageplus/hastorageplus_validate_private
./rt/rgoffload
./rt/rgoffload/rgofl_monitor_start
./rt/rgoffload/rgofl_monitor_stop
./rt/rgoffload/rgofl_probe
./rt/rgoffload/rgofl_svc_start
./rt/rgoffload/rgofl_svc_stop
./rt/rgoffload/rgofl_update
./rt/rgoffload/rgofl_validate
./rt/hamasa
./rt/hamasa/cmas_service_ctrl_check
./rt/hamasa/cmas_service_ctrl_start
./rt/hamasa/cmas_service_ctrl_stop
./rt/hamasa/scmasa_monitor_check
./rt/hamasa/scmasa_monitor_start
./rt/hamasa/scmasa_monitor_stop
./rt/hamasa/scmasa_probe
./rt/hamasa/scmasa_svc_start
./rt/hamasa/scmasa_svc_stop
./rt/hamasa/scmasa_update
./rt/hamasa/scmasa_validate
./rtreg
./rtreg/SUNW.Event
./rtreg/SUNW.HAStorage
./rtreg/SUNW.HAStoragePlus
./rtreg/SUNW.LogicalHostname
./rtreg/SUNW.RGOffload
./rtreg/SUNW.SharedAddress
./rtreg/SUNW.gds
./rtreg/SUNW.scmasa
./rtsupport
-----------------------------------------------------------------------------
bash-2.05# pwd
/usr/cluster/lib/rgm/rt
bash-2.05# ls
haevent hamasa hastorage rgoffload
hafoip hascip hastorageplus
bash-2.05# cd /etc/cluster/ccr/
bash-2.05# ls
dcs_service_1 dcs_service_keys
dcs_service_1.bak dcs_service_keys.bak
dcs_service_10 did_instances
dcs_service_10.bak did_instances.bak
dcs_service_2 did_types
dcs_service_2.bak did_types.bak
dcs_service_3 directory
dcs_service_3.bak directory.bak
dcs_service_4 dpm_status_table
dcs_service_4.bak dpm_status_table.bak
dcs_service_5 epoch
dcs_service_5.bak epoch.bak
dcs_service_6 infrastructure
dcs_service_6.bak infrastructure.bak
dcs_service_7 rgm_rt_SUNW.LogicalHostname
dcs_service_7.bak rgm_rt_SUNW.LogicalHostname:2
dcs_service_8 rgm_rt_SUNW.LogicalHostname:2.bak
dcs_service_8.bak rgm_rt_SUNW.SharedAddress
dcs_service_9 rgm_rt_SUNW.SharedAddress:2
dcs_service_9.bak rgm_rt_SUNW.SharedAddress:2.bak
dcs_service_classes rgm_vp_version
dcs_service_classes.bak rgm_vp_version.bak
-----------------------------------------------------------------------------
scrgadm -a -t SUNW.gds
clrt register rgs
-----------------------------------------------------------------------------
scrgadm -r -t gds
-----------------------------------------------------------------------------
scdidadm -L
metadb -a -c 3 -f
root( c0t1d0 )
metadb -a -c 3 -f c0t1d0s7
root@blade-02 # scdidadm -L
1 u60-10:/dev/rdsk/c0t0d0 /dev/did/rdsk/d1
2 u60-10:/dev/rdsk/c0t1d0 /dev/did/rdsk/d2
3 u60-10:/dev/rdsk/c5t9d0 /dev/did/rdsk/d3
3 blade-02:/dev/rdsk/c4t9d0 /dev/did/rdsk/d3
4 u60-10:/dev/rdsk/c5t10d0 /dev/did/rdsk/d4
4 blade-02:/dev/rdsk/c4t10d0 /dev/did/rdsk/d4
5 u60-10:/dev/rdsk/c5t11d0 /dev/did/rdsk/d5
5 blade-02:/dev/rdsk/c4t11d0 /dev/did/rdsk/d5
6 blade-02:/dev/rdsk/c0t1d0 /dev/did/rdsk/d6
7 blade-02:/dev/rdsk/c0t2d0 /dev/did/rdsk/d7
root@blade-02 # scdidadm -L | grep c4t11d0
5 blade-02:/dev/rdsk/c4t11d0 /dev/did/rdsk/d5
root@blade-02 # cat /etc/hosts
#
# Internet host table
#
::1 localhost
127.0.0.1 localhost
192.168.217.108 blade-02 loghost blade-02.
192.168.217.120 u60-10
root@blade-02 # scdidadm -L | grep c4t11d0
root@blade-02 # metaset -s Z1-DS -a -h blade-02 u60-10
root@blade-02 # metaset -s Z1-DS -a -m blade-02 u60-10
root@blade-02 # metaset -s Z1-DS -a /dev/did/rdsk/d5
root@blade-02 # metaset -s Z1-DS
Set name = Z1-DS, Set number = 1
Host Owner
blade-02 Yes
u60-10
Mediator Host(s) Aliases
blade-02
u60-10
Driv Dbase
d5 Yes
root@blade-02 # metainit -s Z1-DS d1 1 1 /dev/did/rdsk/d5s0
Z1-DS/d1: Concat/Stripe is setup
root@blade-02 # newfs /dev/md/Z1-DS/dsk/d1
root@blade-02 # mount /dev/md/Z1-DS/dsk/d1 /share/
----------------------------------------------------------
1. Create diskset
# metaset -s Z1-DS -a -h NODE1 NODE2
2. Add the same nodes as diskset mediators to each diskset.
# metaset -s Z1-DS -a -m NODE1 NODE2
3. Add the disk(s) chosen previouslyto the
Z1-DS diskset.
# metaset -s Z1-DS -a /dev/did/rdsk/dx [/dev/did/rdsk/dy]
4. Verify the status of the new disksets.
# metaset -s Z1-DS
# medstat -s Z1-DS
#O
5. #check metaset disk for a slice 0 and put it into a metadevice
# metainit -s Z1-DS d1 1 1 /dev/did/rdsk/d#s0
6. add a FS to zhe metadevice
# newfs ...
7. on both nodes create mountpints and vfstab entries
/etc/vfstab
/dev/md/Z1-DS/dsk/d1 /dev/md/Z1-DS/rdsk/d100 /Mountpoint ufs 2 no -
----------------------------------------------------------
root@blade-02 # cluster show| grep -i name
----------------------------------------------------------
root@u60-02 # eeprom | grep local
local-mac-address?=true
root@u60-02 # cat /etc/hostname.hme0
u60-02 group sc_ipmp0 -failover
root@u60-02 # cat /etc/hosts
#
# Internet host table
#
::1 localhost
127.0.0.1 localhost
192.168.217.112 u60-02 loghost u60-02.
192.168.217.122 u60-12
------------------------------------------------------------
root@blade-02 # cat /etc/hostname.*
eri0-test deprecated -failover group sc_ipmp0 up
addif blade-02 up
qfe0-test deprecated -failover group sc_ipmp0 up
------------------------------------------------------------
root@u60-10 # cldg list
Z1-DS
root@u60-10 # cldg show
=== Device Groups ===
Device Group Name: Z1-DS
Type: SVM
failback: false
Node List: blade-02, u60-10
preferenced: true
numsecondaries: 1
diskset name: Z1-DS
root@u60-10 # cldg switch -n u60-10 Z1-DS
------------------------------------------------------------
root@blade-02 # clresourcetype register SUNW.HAStoragePlus
root@blade-02 # clresourcegroup create Z1-DS
root@blade-02 # clresource create -g Z1-DS -t SUNW.HAStoragePlus -p FilesystemMountPoints=/zone01 zone01
root@blade-02 # clrg show
=== Resource Groups and Resources ===
Resource Group: Z1-DS
RG_description: <NULL>
RG_mode: Failover
RG_state: Unmanaged
Failback: False
Nodelist: u60-10 blade-02
--- Resources for Group Z1-DS ---
Resource: zone01
Type: SUNW.HAStoragePlus:6
Type_version: 6
Group: Z1-DS
R_description:
Resource_project_name: default
Enabled{u60-10}: True
Enabled{blade-02}: True
Monitored{u60-10}: True
Monitored{blade-02}: True
root@blade-02 # clresourcegroup online -M Z1-DS
root@blade-02 # clrg show
=== Resource Groups and Resources ===
Resource Group: Z1-DS
RG_description: <NULL>
RG_mode: Failover
RG_state: Managed
Failback: False
Nodelist: u60-10 blade-02
--- Resources for Group Z1-DS ---
Resource: zone01
Type: SUNW.HAStoragePlus:6
Type_version: 6
Group: Z1-DS
R_description:
Resource_project_name: default
Enabled{u60-10}: True
Enabled{blade-02}: True
Monitored{u60-10}: True
Monitored{blade-02}: True
root@blade-02 # cldg switch -n blade-02 Z1-DS
root@blade-02 # clresource status
=== Cluster Resources ===
Resource Name Node Name State Status Message
------------- --------- ----- --------------
zone01 u60-10 Online Online
blade-02 Offline Offline
root@blade-02 # cldg show
=== Device Groups ===
Device Group Name: Z1-DS
Type: SVM
failback: false
Node List: blade-02, u60-10
preferenced: true
numsecondaries: 1
diskset name: Z1-DS
-----------------------------------------------------------
root@blade-02 # clrg switch -n blade-02 Z1-DS
------------------------------------------------------------
182 cd /mnt/Solaris_sparc/Product/sun_cluster_agents/Solaris_10/Packages/
183 ls -la | grep zone
184 pkgadd -d . SUNWsczone
---------------------------------------------------------------
root@u60-10 # clrt register SUNW.gds
root@u60-10 # clrt list
SUNW.LogicalHostname:3
SUNW.SharedAddress:2
SUNW.HAStoragePlus:6
SUNW.gds:6
--------------------------------------------------------------
root@u60-10 # ls /opt/SUNWsczone/sczbt/bin/
clear_zone functions probe_sczbt start_sczbt stop_sczbt validate_sczbt
--------------------------------------------------------------
root@u60-10 # cd /opt/SUNWsczone/sczbt/util/
root@u60-10 # cp sczbt_config sczbt_config.bak
--------------------------------------------------------------
root@u60-10 # ./sczbt_register -f /zone01/paramdir/sczbt_config
sourcing /zone01/paramdir/sczbt_config
Registration of resource zone01-rs succeeded.
Validation of resource zone01-rs succeeded.
---------------------------------------------------------------
root@u60-10 # cat /zone01/paramdir/sczbt_config | grep -v "^#"
RS=zone01-rs
RG=Z1-DS
PARAMETERDIR=/zone01/paramdir
SC_NETWORK=false
SC_LH=
FAILOVER=true
HAS_RS=zone01
Zonename="zone01"
Zonebrand="native"
Zonebootopt=""
Milestone="multi-user-server"
LXrunlevel="3"
SLrunlevel="3"
Mounts=""
-----------------------------------------------------------------
root@u60-10 # scstat -g
-- Resource Groups and Resources --
Group Name Resources
---------- ---------
Resources: Z1-DS zone01 zone01-rs
-- Resource Groups --
Group Name Node Name State Suspended
---------- --------- ----- ---------
Group: Z1-DS u60-10 Online No
Group: Z1-DS blade-02 Offline No
-- Resources --
Resource Name Node Name State Status Message
------------- --------- ----- --------------
Resource: zone01 u60-10 Online Online
Resource: zone01 blade-02 Offline Offline
Resource: zone01-rs u60-10 Offline Offline
Resource: zone01-rs blade-02 Offline Offline
-------------------------------------------------------------------
root@u60-10 # clrs enable zone01-rs
root@u60-10 # scstat -g
-- Resource Groups and Resources --
Group Name Resources
---------- ---------
Resources: Z1-DS zone01 zone01-rs
-- Resource Groups --
Group Name Node Name State Suspended
---------- --------- ----- ---------
Group: Z1-DS u60-10 Online No
Group: Z1-DS blade-02 Offline No
-- Resources --
Resource Name Node Name State Status Message
------------- --------- ----- --------------
Resource: zone01 u60-10 Online Online
Resource: zone01 blade-02 Offline Offline
Resource: zone01-rs u60-10 Online Online
Resource: zone01-rs blade-02 Offline Offline
----------------------------------------------------------------------------
root@u60-10 # bootadm update-archive
updating /platform/sun4u/boot_archive
-------------------------------------------------------------------
[ add comment ] ( 4 views ) | [ 0 trackbacks ] | permalink