CCR Initialization Failure
807567Feb 14 2010 — edited Feb 23 2010I'm trying to install Sun Cluster on LDOM 1.3. I tried 2 nodes which isn't working, so now I'm trying to get just 1 node up.
Config:
2 Public Vnets
1 Private Vnet
1 Shared Quorum Disk (EMC CX-500 1Gb) whole disk given (using EMC Powerpath /dev/dsk/emcpower15c) given to the LDOM.
VFSTAB:
#/dev/dsk/c0d1s0 /dev/rdsk/c0d1s0 /globaldevices ufs 1 no -
/dev/dsk/c0d1s0 /dev/rdsk/c0d1s0 /global/.devices/node@1 ufs 2 no global
Ifconfig -a
[root@fsdev2w]# ifconfig -a
lo0: flags=2001000849<UP,LOOPBACK,RUNNING,MULTICAST,IPv4,VIRTUAL> mtu 8232 index 1
inet 127.0.0.1 netmask ff000000
vnet0: flags=9000843<UP,BROADCAST,RUNNING,MULTICAST,IPv4,NOFAILOVER> mtu 1500 index 2
inet 10.25.23.92 netmask ff000000 broadcast 10.255.255.255
groupname sc_ipmp0
ether 0:14:4f:fb:74:7f
vnet0:1: flags=1000843<UP,BROADCAST,RUNNING,MULTICAST,IPv4> mtu 1500 index 2
inet 0.0.0.0 netmask ff000000 broadcast 0.255.255.255
vnet1: flags=69000842<BROADCAST,RUNNING,MULTICAST,IPv4,NOFAILOVER,STANDBY,INACTIVE> mtu 0 index 4
inet 0.0.0.0 netmask 0
groupname sc_ipmp0
ether 0:14:4f:f9:c:13
clprivnet0: flags=1009843<UP,BROADCAST,RUNNING,MULTICAST,MULTI_BCAST,PRIVATE,IPv4> mtu 1500 index 3
inet 192.168.204.33 netmask fffffff0 broadcast 192.168.204.47
ether 0:0:0:0:0:1
(I brought up vnet1, the install didnt, and didn't work. bringing it up made no difference either)
Install goes fine with /globaldevices mounted as the quorum disk (This disk is shared between the 2 nodes on a seperate physical server LDOM which I've shut down)
The LOFI method doesn't work either.
On reboot I get the following:
Sun Blade T6340 Server Module, No Keyboard
Copyright 2009 Sun Microsystems, Inc. All rights reserved.
OpenBoot 4.30.6, 8192 MB memory available, Serial #83403189.
Ethernet address 0:14:4f:f8:a1:b5, Host ID: 84f8a1b5.
Boot device: ch1bl1gldm3wbdsk File and args:
SunOS Release 5.10 Version Generic_141444-09 64-bit
Copyright 1983-2009 Sun Microsystems, Inc. All rights reserved.
Use is subject to license terms.
Hostname: fsdev2w
NOTICE: mddb: unable to get devid for 'vdc', 0x7
NOTICE: mddb: unable to get devid for 'vdc', 0x7
NOTICE: mddb: unable to get devid for 'vdc', 0x7
Configuring devices.
Reading ZFS config: done.
fsdev2w console login: root
Password:
Last login: Sun Feb 14 23:00:57 on console
Sun Microsystems Inc. SunOS 5.10 Generic January 2005
[root@fsdev2w]# Booting in cluster mode
NOTICE: CMM: Node fsdev2w (nodeid = 1) with votecount = 1 added.
NOTICE: CMM: Node fsdev2w: attempting to join cluster.
NOTICE: CMM: Cluster has reached quorum.
NOTICE: CMM: Node fsdev2w (nodeid = 1) is up; new incarnation number = 1266206912.
NOTICE: CMM: Cluster members: fsdev2w.
NOTICE: CMM: node reconfiguration #1 completed.
Feb 14 23:08:35 fsdev2w cl_runtime: NOTICE: CMM: Node fsdev2w: joined cluster.
Feb 14 23:08:35 fsdev2w ip: ip: joining multicasts failed (18) on clprivnet0 - will use link layer broadcasts for multicast
Configuring DID devices
obtaining access to all attached disks
Configuring the /dev/global directory (global devices)
Feb 14 23:08:52 fsdev2w Cluster.CCR: /usr/cluster/bin/scgdevs: Cannot register devices as HA.
Feb 14 23:08:57 fsdev2w : ccr_initialize failure
Feb 14 23:09:02 fsdev2w last message repeated 8 times
Feb 14 23:09:03 fsdev2w svc.startd[8]: system/cluster/scdpm:default failed repeatedly: transitioned to maintenance (see 'svcs -xv' for details)
fsdev2w console login:
I understand the CCR & CMM are crucial to a cluster functioning (Found out the hard way when I rebooted the second node in a 2 node config and the first node panicked :( )
Can someone PLEASE PLEASE HELP!
Please help! Driving me nuts. Should be pretty straightforward I thoguht.
Edited by: nvmurali on Feb 14, 2010 8:15 PM