With SG Linux using a bonded device, can a second bond device be used as a standby interface?
ServiceGuard (Linux) Standby network interfaces
rhel 6.3 and 11.20.10 , cmcheckconf, network probe fails with messages couldn't talk to itself
Hello,
I would like to ask if somebody did experience during building of sglx 11.20.10 cluster on Rhel 6.3 during cmcheckconf error messages like :
[root@node1 ~]# cmcheckconf -v -C cluster.conf
Begin cluster verification...
Checking cluster file: cluster.conf.
MAX_CONFIGURED_PACKAGES configured to 10.
Checking nodes ... Done
Checking existing configuration ... Done
MAX_CONFIGURED_PACKAGES configured to 10.
Gathering storage information
Found 1 devices on node node1
Found 1 devices on node node2
Analysis of 2 devices should take approximately 1 seconds
0%----10%----20%----30%----40%----50%----60%----70%----80%----90%----100%
Gathering network information
Beginning network probing (this may take a while)
Completed network probing
Network interface bond2 on node node1 couldn't talk to itself.
Network interface bond3 on node node1 couldn't talk to itself.
Network interface bond2 on node node2 couldn't talk to itself.
Network interface bond3 on node node2 couldn't talk to itself.
cmcheckconf: Unable to reconcile configuration file cluster.conf
with discovered configuration information.
and what is the solution to fixing/workarounding the issue.
The interfaces have full connectivity and name resolution is setup according to guide.
thanks
ServiceGuard for Linux A11.20.10 cmclonfd
Hello,
I've installed SG for Linux on a two Suse Linux servers. In /var/log/messages I get a lot of the following messages:
Mar 18 14:45:57 node1 cmclconfd[47486]: Child Process terminated due to an exit(2).
Mar 18 14:45:57 node1 cmclconfd[47493]: Child Process terminated due to an exit(2).
Mar 18 14:45:58 node1 cmclconfd[47500]: Child Process terminated due to an exit(1).
Mar 18 14:45:58 node1cmclconfd[47542]: Child Process terminated due to an exit(2).
Mar 18 14:45:58 node1 cmclconfd[47528]: Child Process terminated due to an exit(127).
Mar 18 14:45:58 node1cmclconfd[47552]: Child Process terminated due to an exit(0).
What might be the cause of these messages?
Thanks.
P.S. this thread has been moved from HP-UX > Serviceguard to Linux > Serviceguard - HP Forums Moderator
cluster issue
hi ...
First of all tell me can we istall service gaurd cluster for Redhat linux.
if yes then i am not able to FAILOVER my server on another node .
P.S.This thread has been moved from Servers >Planning to Linux > Serviceguard-HP Forums Moderator
Error while switching the cluster
I am explaining you the senario.
We have two nodes and on both node we have insatalled redhat linux and for the cluster we have install HP service gaurd.
The problem what we are facing is , when our NODE1 goes down the confingured NODE2 for failover does not take place automatically . I dont know about if it shifts mannually because I was assigned this case now and no one is there for details.
Also when I am using cmviewcl -v i got the output on node1
CLUSTER STATUS
JISP_DATABASE_CLUSTER up
NODE STATUS STATE
hathdb1 up running
Cluster_Lock_LUN:
DEVICE STATUS
/dev/cciss/c0d0p1 up
Network_Parameters:
INTERFACE STATUS NAME
PRIMARY up eth0
PRIMARY up eth1
PACKAGE STATUS STATE AUTO_RUN NODE
oracle up running disabled hathdb1
Policy_Parameters:
POLICY_NAME CONFIGURED_VALUE
Failover configured_node
Failback manual
Script_Parameters:
ITEM STATUS MAX_RESTARTS RESTARTS NAME
Service up 0 0 oracle_db_mon
Service up 5 0 oracle_lsnr_mon
Subnet up 202.88.149.0
Subnet up 192.168.0.0
Node_Switching_Parameters:
NODE_TYPE STATUS SWITCHING NAME
Primary up enabled hathdb1 (current)
Alternate up enabled hathdb2
NODE STATUS STATE
hathdb2 up running
Cluster_Lock_LUN:
DEVICE STATUS
/dev/cciss/c0d0p1 up
Network_Parameters:
INTERFACE STATUS NAME
PRIMARY up eth0
PRIMARY up eth1
and when i go to /oracle and more to clusterciew i got the output as below
[root@hathdb1 oracle]#
[root@hathdb1 oracle]# ll
total 2352
-rw-r--r-- 1 root root 8106 Feb 14 2011 1
drwxr-xr-x 2 root root 4096 Apr 24 2010 backup
-rw-r--r-- 1 root root 1603 Apr 9 2010 clusterciew
-rwx------ 1 root root 8105 Feb 15 2011 oracle.conf
-rwx------ 1 root root 8106 Feb 14 2011 oracle.conf-FEB02
-rwx------ 1 root root 8105 Aug 18 2010 oracle.conf.old
-rwx------ 1 root root 39407 Aug 4 2011 oracle.ctrl
-rwx------ 1 root root 39407 Aug 4 2011 oracle.ctrl_04-08-2011
-rwx------ 1 root root 39407 Feb 7 2007 oracle.ctrl.back
-rwx------ 1 root root 39407 Aug 18 2010 oracle.ctrl.back.old
-rwx------ 1 root root 39457 Feb 14 2011 oracle.ctrl-FEB02
-rwx------ 1 root root 39407 Feb 14 2011 oracle.ctrl-FEB-13-11
-rw-r--r-- 1 root root 610796 Mar 11 15:25 oracle.ctrl.log
-rw-r--r-- 1 root root 1454460 Apr 9 2010 oracle.ctrl.log_primary
-rwx------ 1 root root 39407 Aug 18 2010 oracle.ctrl.old
[root@hathdb1 oracle]# more
usage: more [-dflpcsu] [+linenum | +/pattern] name1 name2 ...
[root@hathdb1 oracle]#
[root@hathdb1 oracle]#
[root@hathdb1 oracle]#
[root@hathdb1 oracle]#
[root@hathdb1 oracle]# more clusterciew
CLUSTER STATUS
JISP_DATABASE_CLUSTER down
NODE STATUS STATE
hathdb1 down unknown
Cluster_Lock_LUN:
DEVICE STATUS
/dev/cciss/c0d0p1 unknown
Network_Parameters:
INTERFACE STATUS NAME
PRIMARY unknown eth0
PRIMARY unknown eth1
NODE STATUS STATE
hathdb2 down unknown
Cluster_Lock_LUN:
DEVICE STATUS
/dev/cciss/c0d0p1 unknown
Network_Parameters:
INTERFACE STATUS NAME
PRIMARY unknown eth0
PRIMARY unknown eth1
UNOWNED_PACKAGES
PACKAGE STATUS STATE AUTO_RUN NODE
oracle down unowned
Policy_Parameters:
POLICY_NAME CONFIGURED_VALUE
Failover unknown
Failback unknown
Script_Parameters:
ITEM STATUS NODE_NAME NAME
Subnet unknown hathdb1 202.88.149.0
Subnet unknown hathdb1 192.168.0.0
Subnet unknown hathdb2 202.88.149.0
Subnet unknown hathdb2 192.168.0.0
Node_Switching_Parameters:
NODE_TYPE STATUS SWITCHING NAME
Primary down hathdb1
Alternate down hathdb2
[root@hathdb1 oracle]#
[root@hathdb1 oracle]#
i didnot understand this one can you explain why it is so.
Also I am new to cluster so give me the solution i will be thankfull to you.
For details please find the attachment.
Thanks and regaurd.
Ashish
SG Cluster for Linux with SG Ext for RAC on KVM - is it supported?
Hi Team,
We want to configure SG Cluster for Linux with SG Ext for RAC on 2 KVM based VM's...
Is this a supported configuration?
As per http://bizsupport1.austin.hp.com/bc/docs/support/SupportManual/c03398707/c03398707.pdf KVM is not listed as a supported hypervisor, please refer page no:2
Also kindly point me to some documentation for SG Cluster configuration for Linux with SG Ext for RAC.
Any input would be greatly appreciated
Regards,
Rudra
HP has a Serviceguard for Linux course
www.hp.com/education. look in the linux catalog
How to failover the package from one node to another node with out package down
Hi All,
How to failover the package from one node to another node with out package downtime / with out halting the node.
Please let me know the steps to be performed with syntax.
How to remove the service from one node of three nodes cluster completely from Linux
Hi All,
There are three Redhat Linux nodes are configured with service guard cluster. I would like to remove service guard completely from one node of the cluster but rest of nodes should be running with packages
I would like to know the steps with commands starting from deleting cluster configuration to removing service guard cluster RPMs.
Cannot create cluster , 'java.util.NoSuchElementException' is shown
--- Error Report Begin --- This error report was automatically generated by HP software. Operating System Version: 2.6.32-220.el6.x86_64 Operating System Architecture: amd64 Java Runtime Environment Version: 1.7.0_45 --- Product Information --- Product Name: Serviceguard Manager Version: B.03.31 Build Id: Tue Jun 05 04:58:32 PDT 2012 Operating System: Linux Error Id: Z63B3CC6E --- Error Information --- java.util.NoSuchElementException at java.util.StringTokenizer.nextToken(StringTokenizer.java:349) at java.util.StringTokenizer.nextToken(StringTokenizer.java:377) at com.hp.sgmgr.utils.gadgets.GenericErrorPageG$createClusterAction.perform(GenericErrorPageG.java:203) at com.hp.vse.ui.glib.GlibServlet.invokeAction(GlibServlet.java:399) at com.hp.vse.ui.glib.GlibServlet.invokeAction(GlibServlet.java:408) at com.hp.vse.ui.glib.GlibServlet.invokeAction(GlibServlet.java:408) at com.hp.vse.ui.glib.GlibServlet.invokeAction(GlibServlet.java:408) at com.hp.vse.ui.glib.GlibServlet.invokeAction(GlibServlet.java:385) at com.hp.vse.ui.glib.GlibServlet.doProcessRequest(GlibServlet.java:328) at com.hp.vse.ui.glib.GlibServlet.processRequest(GlibServlet.java:70) at com.hp.vse.ui.glib.GlibServlet.doPost(GlibServlet.java:40) at javax.servlet.http.HttpServlet.service(HttpServlet.java:637) at javax.servlet.http.HttpServlet.service(HttpServlet.java:717) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:290) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206) at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:233) at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:191) at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:127) at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:102) at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:109) at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:298) at org.apache.coyote.http11.Http11Processor.process(Http11Processor.java:857) at org.apache.coyote.http11.Http11Protocol$Http11ConnectionHandler.process(Http11Protocol.java:588) at org.apache.tomcat.util.net.JIoEndpoint$Worker.run(JIoEndpoint.java:489) at java.lang.Thread.run(Thread.java:744) --- Log Messages --- GET /sgmgr/main/3; POST /sgmgr/main/0; GET /sgmgr/main/0; GET /sgmgr/main/1; POST /sgmgr/main/1;HTTP_SESSION_ID_FOR_PAGEG=[F8050C1D6F086DA3F86CD27FE37300D5];actionField=[createClusterButton];confirmation_field=[];formCommand=[];page=[genericErrorPage];selectionField=[]; --- Error Report End ---
P.S. This thread has been moved from HP-UX > Serviceguard to Linux > Serviceguard. - Hp Forum moderator
How to configure volume group to use in Serviceguard
Host based storage migration on Linux SG clsuter nodes
Hi All,
Have created the document for host based storage migration on Linux SG cluster nodes. There are couple of clusters which have RHEL 4 / RHEL 5 with A.11.18.06 / A.11.18.01.
Have created the migration steps for Lock LUN and migration of SG cluster package's file system
Please review the below mentioned steps completely and let me know if you need any things to be added / changed.
My Questions:
1)Please let me know if we can do online migration of Lock LUN on HP SG version A.11.18.06 / A.11.18.01.
2)I understand that migration of SG cluster package's file system can be done online.
Steps created for migration of Lock LUN and SG cluster package migration:
- 1. Lock LUN Migration – Offline :
1.1. Take server configuration backup:
#/opt/linuxdepots/tcs/scripts/linux_cfg_bkup.sh
#cmviewcl > /root/backup/cmviewcl
#cmviewcl –v > /root/backup/cmviewcl-v
#cmviewconf > /root/backup/cmviewconf
1.2. Storage team needs to assign the 1GB LUN for lock LUN for all the nodes of the cluster:
1.3. Unix team need to scan 1GB LUN (Lock LUN) on all three servers:
#/opt/hp/hp_fibreutils/hp_rescan –a
#powermt check
#powermt config
#powermt save
Note: Now you can identify the new device. For say, we can assume new device name is /dev/emcpowerXY
1.4. On Node 1, Please create one partition (id – 83):
The partition name will be /dev/emcpowerXY1
1.5. On all three nodes, run below command and confirm if newly created partition is available on all nodes:
#partprobe
1.6. Halt all packages, nodes and entire cluster:
1.6.1. First we have to halt all the packages after getting confirmation from application team:
#cmhaltpkg –v <package_name>
1.6.2. Second we have to halt all the nodes and stop entire cluster:
#cmhaltcl –v
1.6.3. Verify if the cluster is stopped:
#cmviewcl
1.7. On Node1, run following steps to edit cluster configuration and add new device as lock LUN:
#mkdir /root/storage_migration
#cd /root/storage_migration
#cmviewconf > cmviewconf_original
#cmgetconf -c <clsutername> <clustername.ascii>
Note: Now edit the file“<clustername.ascii>” and update new device for lock LUN for all nodes (for all nodes, lock LUN device name will be same).
Example:
Old:
cluster lock lun name: /dev/emcpowerl1
New:
cluster lock lun name: /dev/emcpowerXY1
1.8. Verify if we have edited properly:
#cmcheckconf -v -C <clustername.ascii> ------------à (C is capital)
1.9. Apply / Distribute new configuration changes to all nodes of the cluster:
#cmapplyconf -v -C anyname.ascii (C is capital)
1.10. Now start the cluster and make all nodes join the cluster:
#cmruncl –v
Note: Now the cluster will be started and all nodes will join the cluster.
Run below command to join particular node to the cluster if that node is joined:
#cmrunnode –v <hostname>
Verify if cluster started and all the nodes joined the cluster:
#cmviewcl
1.11. Now start all the packages which were running earlier in respective nodes:
#cmrunpkg <package name> -n <node name>
Note: Packages cannot be started if the packages is not enabled on particular node. Run below command to enable the package.
#cmmodpkg –n <hostname> -e <package name>
1.12. Verify if all the packages are started and running on respective nodes as same as before migration:
#cmviewcl OR cmviewcl -v
- 2. Migrate Service Guard cluster packages file systems / volumes:
2.1. Ensure to have full backup of the file system in place before you start SG cluster file system migration.
2.2. Storage team needs to assign the required LUNs for SG cluster file system migration.
2.3. Scan for new disks On Node - 1
# powermt display dev=all > /tmp/power.before.out
#opt/hp/hp_fibreutils/hp_rescan –a
# powermt config
#powermt save
# powermt check
# powermt display dev=all > /tmp/power.after.out
# diff /tmp/power.before.out /tmp/power.after.out > /tmp/diff.out
# grep -i "Logical Device" /tmp/diff.out --------------à This command will show new LUN IDs.
NEW LUN id:
> Logical device ID=14B6 -- emcpowerd -- 112
> Logical device ID=14B7 -- emcpowerf -- 112
> Logical device ID=14B8 -- emcpowerc -- 112
> Logical device ID=14B9 -- emcpowerg -- 112
2.4. Scan for new disks on Node 2 & 3.....
# powermt display dev=all > /tmp/power.before.out
#opt/hp/hp_fibreutils/hp_rescan –a
# powermt config
#powermt save
# powermt check
# powermt display dev=all > /tmp/power.after.out
# diff /tmp/power.before.out /tmp/power.after.out > /tmp/diff.out
# grep -i "Logical Device" /tmp/diff.out --------------à This command will show new LUN IDs.
NEW LUN id:
> Logical device ID=14B6 -- emcpowerd -- 112
> Logical device ID=14B7 -- emcpowerf -- 112
> Logical device ID=14B8 -- emcpowerc -- 112
> Logical device ID=14B9 -- emcpowerg -- 112
2.5. Create partitions on all new disks On Node - 1
# fdisk /dev/emcpowerXX
2.6. Scan for partition changed On node 2 & 3
# partprobe
#fdisk –l </dev/emcpowerXX
2.7. Do pvcreate in all new disks on Node 1
# pvcreate /dev/emcpowerXX1
2.8. Extend VG with new disks on Node 1
#vgextend VolGroup01 /dev/emcpowerXX1 /dev/emcpowerXY1
Volume group "VolGroup01" successfully extended
2.9. Confirm the new LUNs are part of VG on Node 1:
# vgdisplay –v <VG name> | more
2.10. Now mirror the existing disk to new disk on Node1:
# lvconvert -m1 --corelog /dev/<VG name>/<LV name> /dev/emcpowerXX1 /dev/emcpowerXY1
2.11. Verify if mirroring is done 100% on Node1:
#lvs --noheadings /dev/<VG name>/<LV name>
2.12. Split the mirror from old disks once the mirror completed on Node1:
# lvconvert -m0 /dev/<VG name >/<LV name> /dev/emcpowerYX1 /dev/emcpoweYY1
(emcpowerYX1 and emcpoweYY1 are old disks)
2.13. Verify the new LUNs are part of LV now on Node1
# lvdisplay -m /dev/<VG name> /<LV name>
2.14. Run partprobe on all nodes to make above changes effective on other nodes:
#partprobe
2.15. Verify if the cluster packages file systems are working fine on other nodes
Note: We need to halt the packages on running node and start the packages on other nodes one by one and verify if the package is working fine and packages file systems are getting mounted when the packages are started on other nodes. Once the package is test verified on all nodes. We have start the package back on original node.
# cmhaltpkg –v <package_name>
#cmrunpkg –v –n <nodename> <packagename>
#vgdisplay
#lvdisplay -----------à Need to check if lvdisplay is showing new device.
2.16. Remove the old LUNs from the VG on node 1
Note : This should be done once you get confirmation from Application team and the LV which has been mirrored is running fine.
# vgreduce <VG name> /dev/emcpowerYX
2.17. Remove the LVM header from old disks
# pvremove /dev/emcpowerYX1
2.18. Remove the partitions from the old disks by fdisk command
2.19. Ask SAN team to remove the disk after removed the disk from LVM and fdisk
2.20. Once SAN confirmed that disk has been removed then remove the old path on all nodes by below commands
Linux ServiceGuard with DRBD
Hello.
I have to mount a Linux cluster with two storage arrays that are not connected via SAN.
Is this supported Serviceguard Linux with DRBD?
Thanks.
SG 11.20.22 RHEL 6.5 cmcheckconf error + Ora Tkit
hi i try to update our oracle env.
actual env:
Rhel 6.3
SG 11.20.10
Ora Tkit A05.01.01
after update:
RHEL 6.5
SG 11.20.22
Ora Tkit A05.01.12
update proccess or new installation works fine without any errors or warnings.
i can create/failover/start packages.
but if i do
/usr/local/cmcluster/bin/cmcheckconf -v -C /usr/local/cmcluster/conf/cmcluster.conf -p /usr/local/cmcluster/conf/pkg_ref_conf;
i´ve got many errors:
Begin cluster verification... Checking cluster file: /usr/local/cmcluster/conf/cmcluster.conf. MAX_CONFIGURED_PACKAGES configured to 300. Checking nodes ... Done Checking existing configuration ... Done MAX_CONFIGURED_PACKAGES configured to 300. Gathering storage information Gathering network information Beginning network probing (this may take a while) Completed network probing Cluster clu-rbg01 is an existing cluster Begin checking consistency of files mentioned in cmclfiles2check /usr/local/cmcluster/conf/cmclfiles2check is the same across nodes lnx-rbg09 lnx-rbg10 /etc/nsswitch.conf is the same across nodes lnx-rbg09 lnx-rbg10 /etc/services is the same across nodes lnx-rbg09 lnx-rbg10 /usr/local/cmcluster/conf/cmknowncmds is the same across nodes lnx-rbg09 lnx-rbg10 /usr/local/cmcluster/conf/cmcluster.rc is the same across nodes lnx-rbg09 lnx-rbg10 All files were the same on all nodes "-p" option specified. Opening package reference file: /usr/local/cmcluster/conf/pkg_ref_conf. Validating package poct1rbg via /usr/local/cmcluster/conf/scripts/mscripts/master_control_script.sh ... Waiting for up to 300 seconds for the validation. On node lnx-rbg09, validation of package poct1rbg succeeded with: /var/tmp/master_control_script_env.poct1rbg.708031: line 84: SG_LNX-RBG09_PID[0]=: command not found ..... ..... /var/tmp/master_control_script_env.poct1rbg.708031: line 99: SG_LNX-RBG10_PID[2]=: command not found PR_mode:DISABLED /var/tmp/master_control_script_env.poct1rbg.708031: line 84: SG_LNX-RBG09_PID[0]=: command not found ..... ..... /var/tmp/master_control_script_env.poct1rbg.708031: line 99: SG_LNX-RBG10_PID[2]=: command not found Verifying if the IP addresses could be added to corresponding subnets /var/tmp/master_control_script_env.poct1rbg.708031: line 84: SG_LNX-RBG09_PID[0]=: command not found /var/tmp/master_control_script_env.poct1rbg.708031: line 85: SG_LNX-RBG10_PID[0]=: command not found /var/tmp/master_control_script_env.poct1rbg.708031: line 91: SG_LNX-RBG09_PID[1]=: command not found /var/tmp/master_control_script_env.poct1rbg.708031: line 92: SG_LNX-RBG10_PID[1]=: command not found /var/tmp/master_control_script_env.poct1rbg.708031: line 98: SG_LNX-RBG09_PID[2]=: command not found /var/tmp/master_control_script_env.poct1rbg.708031: line 99: SG_LNX-RBG10_PID[2]=: command not found Install directory operation mode selected. Oracle Toolkit validation The toolkit configuration file in the toolkit configuration directory will be backed up and a new file will be created when the package configuration is applied. /var/tmp/master_control_script_env.poct1rbg.708031: line 84: SG_LNX-RBG09_PID[0]=: command not found ..... ..... /var/tmp/master_control_script_env.poct1rbg.708031: line 99: SG_LNX-RBG10_PID[2]=: command not found PR_mode:DISABLED /var/tmp/master_control_script_env.poct1rbg.708031: line 84: SG_LNX-RBG09_PID[0]=: command not found ...... ...... /var/tmp/master_control_script_env.poct1rbg.708031: line 99: SG_LNX-RBG10_PID[2]=: command not found Verifying if the IP addresses could be added to corresponding subnets /var/tmp/master_control_script_env.poct1rbg.708031: line 84: SG_LNX-RBG09_PID[0]=: command not found /var/tmp/master_control_script_env.poct1rbg.708031: line 85: SG_LNX-RBG10_PID[0]=: command not found /var/tmp/master_control_script_env.poct1rbg.708031: line 91: SG_LNX-RBG09_PID[1]=: command not found /var/tmp/master_control_script_env.poct1rbg.708031: line 92: SG_LNX-RBG10_PID[1]=: command not found /var/tmp/master_control_script_env.poct1rbg.708031: line 98: SG_LNX-RBG09_PID[2]=: command not found /var/tmp/master_control_script_env.poct1rbg.708031: line 99: SG_LNX-RBG10_PID[2]=: command not found Install directory operation mode selected. Oracle Toolkit validation The toolkit configuration file in the toolkit configuration directory will be backed up and a new file will be created when the package configuration is applied. /var/tmp/master_control_script_env.poct1rbg.708031: line 84: SG_LNX-RBG09_PID[0]=: command not found ..... ..... /var/tmp/master_control_script_env.poct1rbg.708031: line 99: SG_LNX-RBG10_PID[2]=: command not found Validation for package poct1rbg succeeded via /usr/local/cmcluster/conf/scripts/mscripts/master_control_script.sh. Validation for package poct5rbg succeeded via /usr/local/cmcluster/conf/scripts/mscripts/master_control_script.sh. Maximum configured packages parameter is 300. Modifying configuration on node lnx-rbg09 Modifying configuration on node lnx-rbg10 Modifying the cluster configuration for cluster clu-rbg01 Modifying the package configuration for package poct1rbg. Modifying the package configuration for package poct5rbg. Use the cmapplyconf command to apply the configuration.
Can anybody help me or have the same situation?
cmquerycl not work
Hi I try to config SG cluster with 2 nodes at RHEL 6 but when y try to execute the command cmquerycl show me the following error:
[root@wamas1 conf]# cmquerycl -v -C cluster.conf -n wamas1.rslapita.local -n wamas2.rslapita.local
Warning: Unable to determine local domain name for wamas1
Looking for other clusters ... Done
ERROR: Node wamas1.rslapita.local is refusing Serviceguard communication.
Please make sure that the proper security access is configured on node
wamas1.rslapita.local through either file-based authorization(pre-A.11.16 version) or
access control policies (version A.11.16 or higher)
and/or that the host name lookup on node wamas1.rslapita.local
resolves the IP address correctly.
Please make sure that cmclnodelist on node wamas1.rslapita.local has entries
for all the nodes in the cluster.
Failed to gather configuration information.
My cmclnodelist is configured:
[root@wamas1 conf]# cat cmclnodelist
wamas1.rslapita.local root
wamas2.rslapita.local root
My etc/hosts file:
192.168.105.35 wamas1 wamas1.rslapita.local
192.168.105.36 wamas2 wamas2.rslapita.local
What can I do?
Any body Help Me?
Regards
Victor
Cannot get disk partition for device /dev/sde Invalid data for cluster lock LUN configuration
I guys!
I have some trouble with my config.
I try to configure my sg cluster with 2 nodes, when try to check the config showme the following error.
cmcheckconf -v -C cluster.conf
Begin cluster verification...
Checking cluster file: cluster.conf.
MAX_CONFIGURED_PACKAGES configured to 300.
Checking nodes ... Done
Checking existing configuration ... Done
MAX_CONFIGURED_PACKAGES configured to 300.
Gathering storage information
Found 1 devices on node wamas1
Found 1 devices on node wamas2
Analysis of 2 devices should take approximately 1 seconds
0%----10%----20%----30%----40%----50%----60%----70%----80%----90%----100%
Gathering network information
Beginning network probing (this may take a while)
Completed network probing
Cannot get disk partition for device /dev/sdeInvalid data for cluster lock LUN configuration
Maximum configured packages parameter is 300.
cmcheckconf: Unable to verify cluster file: cluster.conf.
My cluster.conf
NODE_NAME wamas1
NETWORK_INTERFACE eth1
HEARTBEAT_IP xx.xx.xx.xx
NETWORK_INTERFACE eth5
HEARTBEAT_IP yy.yy.yy.yy
NETWORK_INTERFACE bond0
STATIONARY_IP zz.zz.zz.zz
CLUSTER_LOCK_LUN /dev/sde
NODE_NAME wamas2
NETWORK_INTERFACE eth1
HEARTBEAT_IP xx.xx.xx.xx
NETWORK_INTERFACE eth5
HEARTBEAT_IP yy.yy.yy.yy
NETWORK_INTERFACE bond0
STATIONARY_IP zz.zz.zz.zz
CLUSTER_LOCK_LUN /dev/sde
[root@wamas1 conf]# fdisk -l /dev/sde
Disk /dev/sde: 499 MB, 499974144 bytes
16 heads, 60 sectors/track, 1017 cylinders
Units = cylinders of 960 * 512 = 491520 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 1048576 bytes
Disk identifier: 0xa71b4a14
Device Boot Start End Blocks Id System
/dev/sde1 1 1017 488130 83 Linux
What Can I Do?
Regards,
VB
Serviceguard on non-HP servers
Hi,
We intend to migrate the HP-UX environment to RHEL. Unfortunately, the Linux environemnt is not running on HP servers which raises the question, if it is supported at all to run SGLX on non-HP x86 servers?
The support matrix does only list HP gear for the server part...
Thanks in advance!
Service command /etc/opt/omni/server/sg/csfailover.ksh is not executable
Created SG 12.0 on RHEL 6.6 x64
Created a failover package for virtual ip address
Created a failover package for file system (/fs1). It will be mounted only one node at a time.
Installed DP on shared file system on both nodes.
Then tried to create a failover package for DP service.
During cmcheckconf getting the below error.
Because sg is expecting the file "/etc/opt/omni/server/sg/csfailover.ksh" on both nodes. But it will be availble on mounted file system node only, here file system is mounted in host1. Please helm to solve this issue.
Note: But SG 11.20 it worked fine.
On node <host2>, validation of package ob2cl failed with:
PR_mode:DISABLED
ERROR: Function sg_validate_services
Service command /etc/opt/omni/server/sg/csfailover.ksh is not executable
ERROR: Failed to validate /usr/local/cmcluster/conf/scripts/sg/service.sh
On node <host2>, validation of package ob2cl succeeded with:
PR_mode:DISABLED
cmcheckconf: Error found in package file: /usr/local/cmcluster/conf/pkg/ob2cl/ob2cl.conf.
service guard manager error,unable to create package pkg_name,need help
while creating a general package with out any extra tool kit service guard manager shows unable to create package error.It's in redhat linux 6.5
package not getting up in node
I am using two node ans hp servigurd is running on it
yeaterday i restarted my server after My package is not getting up please help
CLUSTER STATUS
dbcluster up
NODE STATUS STATE
dbnode01 up running
dbnode02 up running
UNOWNED_PACKAGES
PACKAGE STATUS STATE AUTO_RUN NODE
dbpkg down halted disabled unowned
[root@dbnode02 ~]#
Earliear i removed the mounted partion from my node, yeaterday i restarted my server after package is not getting UP pls help.
P.S. This thread has been moved from HP-UX > Serviceguard to Linux > Serviceguard. - Hp Forum Moderator