Quantcast
Channel: Serviceguard topics
Viewing all 361 articles
Browse latest View live

ServiceGuard (Linux) Standby network interfaces

$
0
0

With SG Linux using a bonded device, can a second bond device be used as a standby interface?


rhel 6.3 and 11.20.10 , cmcheckconf, network probe fails with messages couldn't talk to itself

$
0
0

Hello,

 

I would like to ask if somebody did experience during building of sglx 11.20.10 cluster on Rhel 6.3 during cmcheckconf error messages like :

 

[root@node1 ~]# cmcheckconf -v -C cluster.conf

Begin cluster verification...

Checking cluster file: cluster.conf.

MAX_CONFIGURED_PACKAGES configured to 10.

Checking nodes ... Done

Checking existing configuration ... Done

MAX_CONFIGURED_PACKAGES configured to 10.

Gathering storage information

Found 1 devices on node node1

Found 1 devices on node node2

Analysis of 2 devices should take approximately 1 seconds

0%----10%----20%----30%----40%----50%----60%----70%----80%----90%----100%

Gathering network information

Beginning network probing (this may take a while)

Completed network probing

Network interface bond2 on node node1 couldn't talk to itself.

Network interface bond3 on node node1 couldn't talk to itself.

Network interface bond2 on node node2 couldn't talk to itself.

Network interface bond3 on node node2 couldn't talk to itself.

cmcheckconf: Unable to reconcile configuration file cluster.conf

 with discovered configuration information.

 

and what is the solution to fixing/workarounding the issue.

 

The interfaces have full connectivity and name resolution is setup according to guide.

 

thanks

ServiceGuard for Linux A11.20.10 cmclonfd

$
0
0

Hello,

 

I've installed SG for Linux on a two Suse Linux servers. In /var/log/messages I get a lot of the following messages:

Mar 18 14:45:57 node1 cmclconfd[47486]: Child Process terminated due to an exit(2).
Mar 18 14:45:57 node1 cmclconfd[47493]: Child Process terminated due to an exit(2).
Mar 18 14:45:58 node1 cmclconfd[47500]: Child Process terminated due to an exit(1).
Mar 18 14:45:58 node1cmclconfd[47542]: Child Process terminated due to an exit(2).
Mar 18 14:45:58 node1 cmclconfd[47528]: Child Process terminated due to an exit(127).
Mar 18 14:45:58 node1cmclconfd[47552]: Child Process terminated due to an exit(0).


What might be the cause of these messages?

 

Thanks.

 

 

P.S. this thread has been moved  from HP-UX > Serviceguard to Linux > Serviceguard - HP Forums Moderator

cluster issue

$
0
0

hi ...

 

First of all tell me can we istall service gaurd cluster for Redhat linux.

 

 

if yes then i am not able to FAILOVER my server on another node .

 

 

 

P.S.This thread has been moved from Servers >Planning to Linux > Serviceguard-HP Forums Moderator

Error while switching the cluster

$
0
0

I am explaining you the senario.

 

We have two nodes and on both node we have insatalled redhat linux and for the cluster we have install HP service gaurd.

 

The problem what we are facing is , when our NODE1 goes down the confingured NODE2 for failover does not take place automatically . I dont know about if it shifts mannually because I was assigned this case now and no one is there for details.

 

Also when I am using cmviewcl  -v i got the output on node1

 

CLUSTER STATUS

JISP_DATABASE_CLUSTER up

NODE STATUS STATE

hathdb1 up running

Cluster_Lock_LUN:

DEVICE STATUS

/dev/cciss/c0d0p1 up

Network_Parameters:

INTERFACE STATUS NAME

PRIMARY up eth0

PRIMARY up eth1

PACKAGE STATUS STATE AUTO_RUN NODE

oracle up running disabled hathdb1

Policy_Parameters:

POLICY_NAME CONFIGURED_VALUE

Failover configured_node

Failback manual

Script_Parameters:

ITEM STATUS MAX_RESTARTS RESTARTS NAME

Service up 0 0 oracle_db_mon

Service up 5 0 oracle_lsnr_mon

Subnet up 202.88.149.0

Subnet up 192.168.0.0

Node_Switching_Parameters:

NODE_TYPE STATUS SWITCHING NAME

Primary up enabled hathdb1 (current)

Alternate up enabled hathdb2

NODE STATUS STATE

hathdb2 up running

Cluster_Lock_LUN:

DEVICE STATUS

/dev/cciss/c0d0p1 up

Network_Parameters:

INTERFACE STATUS NAME

PRIMARY up eth0

PRIMARY up eth1

 

and when i go to /oracle and more to clusterciew i got the output as below

 

[root@hathdb1 oracle]#

[root@hathdb1 oracle]# ll

total 2352

-rw-r--r-- 1 root root 8106 Feb 14 2011 1

drwxr-xr-x 2 root root 4096 Apr 24 2010 backup

-rw-r--r-- 1 root root 1603 Apr 9 2010 clusterciew

-rwx------ 1 root root 8105 Feb 15 2011 oracle.conf

-rwx------ 1 root root 8106 Feb 14 2011 oracle.conf-FEB02

-rwx------ 1 root root 8105 Aug 18 2010 oracle.conf.old

-rwx------ 1 root root 39407 Aug 4 2011 oracle.ctrl

-rwx------ 1 root root 39407 Aug 4 2011 oracle.ctrl_04-08-2011

-rwx------ 1 root root 39407 Feb 7 2007 oracle.ctrl.back

-rwx------ 1 root root 39407 Aug 18 2010 oracle.ctrl.back.old

-rwx------ 1 root root 39457 Feb 14 2011 oracle.ctrl-FEB02

-rwx------ 1 root root 39407 Feb 14 2011 oracle.ctrl-FEB-13-11

-rw-r--r-- 1 root root 610796 Mar 11 15:25 oracle.ctrl.log

-rw-r--r-- 1 root root 1454460 Apr 9 2010 oracle.ctrl.log_primary

-rwx------ 1 root root 39407 Aug 18 2010 oracle.ctrl.old

[root@hathdb1 oracle]# more

usage: more [-dflpcsu] [+linenum | +/pattern] name1 name2 ...

[root@hathdb1 oracle]#

[root@hathdb1 oracle]#

[root@hathdb1 oracle]#

[root@hathdb1 oracle]#

[root@hathdb1 oracle]# more clusterciew

CLUSTER STATUS

JISP_DATABASE_CLUSTER down

NODE STATUS STATE

hathdb1 down unknown

Cluster_Lock_LUN:

DEVICE STATUS

/dev/cciss/c0d0p1 unknown

Network_Parameters:

INTERFACE STATUS NAME

PRIMARY unknown eth0

PRIMARY unknown eth1

NODE STATUS STATE

hathdb2 down unknown

Cluster_Lock_LUN:

DEVICE STATUS

/dev/cciss/c0d0p1 unknown

Network_Parameters:

INTERFACE STATUS NAME

PRIMARY unknown eth0

PRIMARY unknown eth1

UNOWNED_PACKAGES

PACKAGE STATUS STATE AUTO_RUN NODE

oracle down unowned

Policy_Parameters:

POLICY_NAME CONFIGURED_VALUE

Failover unknown

Failback unknown

Script_Parameters:

ITEM STATUS NODE_NAME NAME

Subnet unknown hathdb1 202.88.149.0

Subnet unknown hathdb1 192.168.0.0

Subnet unknown hathdb2 202.88.149.0

Subnet unknown hathdb2 192.168.0.0

Node_Switching_Parameters:

NODE_TYPE STATUS SWITCHING NAME

Primary down hathdb1

Alternate down hathdb2

[root@hathdb1 oracle]#

[root@hathdb1 oracle]#

 

i didnot understand this one can you explain why it is so.

 

Also I am new to cluster so give me the solution i will be thankfull to you.

 

For details please find the attachment.

 

Thanks and regaurd.

Ashish

SG Cluster for Linux with SG Ext for RAC on KVM - is it supported?

HP has a Serviceguard for Linux course

$
0
0
i am not sure if people are aware that hp has a serviceguard on Linux course that gives each student their own cluster and goes through installation and cluster creation to setting up a oracle database package.

www.hp.com/education. look in the linux catalog

How to failover the package from one node to another node with out package down

$
0
0

Hi All,

 

 

How to failover the package from one node to another node with out package downtime / with out halting the node.

 

 

Please let me know the steps to be performed with syntax.


How to remove the service from one node of three nodes cluster completely from Linux

$
0
0

Hi All,

 

There are three Redhat Linux nodes are configured with service guard cluster.  I would like to remove service guard completely from one node of the cluster but rest of nodes should be running with packages 

 

I would like to know the steps with commands starting from deleting cluster configuration to removing service guard cluster RPMs.

 

 

Cannot create cluster , 'java.util.NoSuchElementException' is shown

$
0
0
--- Error Report Begin ---
This error report was automatically generated by HP software.
Operating System Version: 2.6.32-220.el6.x86_64
Operating System Architecture: amd64
Java Runtime Environment Version: 1.7.0_45
--- Product Information ---
Product Name: Serviceguard Manager
Version: B.03.31
Build Id: Tue Jun 05 04:58:32 PDT 2012
Operating System: Linux
Error Id: Z63B3CC6E
--- Error Information ---
java.util.NoSuchElementException
	at java.util.StringTokenizer.nextToken(StringTokenizer.java:349)
	at java.util.StringTokenizer.nextToken(StringTokenizer.java:377)
	at com.hp.sgmgr.utils.gadgets.GenericErrorPageG$createClusterAction.perform(GenericErrorPageG.java:203)
	at com.hp.vse.ui.glib.GlibServlet.invokeAction(GlibServlet.java:399)
	at com.hp.vse.ui.glib.GlibServlet.invokeAction(GlibServlet.java:408)
	at com.hp.vse.ui.glib.GlibServlet.invokeAction(GlibServlet.java:408)
	at com.hp.vse.ui.glib.GlibServlet.invokeAction(GlibServlet.java:408)
	at com.hp.vse.ui.glib.GlibServlet.invokeAction(GlibServlet.java:385)
	at com.hp.vse.ui.glib.GlibServlet.doProcessRequest(GlibServlet.java:328)
	at com.hp.vse.ui.glib.GlibServlet.processRequest(GlibServlet.java:70)
	at com.hp.vse.ui.glib.GlibServlet.doPost(GlibServlet.java:40)
	at javax.servlet.http.HttpServlet.service(HttpServlet.java:637)
	at javax.servlet.http.HttpServlet.service(HttpServlet.java:717)
	at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:290)
	at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206)
	at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:233)
	at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:191)
	at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:127)
	at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:102)
	at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:109)
	at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:298)
	at org.apache.coyote.http11.Http11Processor.process(Http11Processor.java:857)
	at org.apache.coyote.http11.Http11Protocol$Http11ConnectionHandler.process(Http11Protocol.java:588)
	at org.apache.tomcat.util.net.JIoEndpoint$Worker.run(JIoEndpoint.java:489)
	at java.lang.Thread.run(Thread.java:744)
--- Log Messages ---
GET /sgmgr/main/3;
POST /sgmgr/main/0;
GET /sgmgr/main/0;
GET /sgmgr/main/1;
POST /sgmgr/main/1;HTTP_SESSION_ID_FOR_PAGEG=[F8050C1D6F086DA3F86CD27FE37300D5];actionField=[createClusterButton];confirmation_field=[];formCommand=[];page=[genericErrorPage];selectionField=[];
--- Error Report End ---

 

P.S. This thread has been moved from HP-UX > Serviceguard to Linux > Serviceguard. - Hp Forum moderator

How to configure volume group to use in Serviceguard

Host based storage migration on Linux SG clsuter nodes

$
0
0

Hi All,

 

Have created the document for host based storage migration on Linux SG cluster nodes. There are couple of clusters which have RHEL 4 / RHEL 5 with A.11.18.06 / A.11.18.01.

 

Have created the migration steps for Lock LUN and migration of SG cluster package's file system

 

 

Please review the below mentioned steps completely and let me know if you need any things to be added / changed.

 

 

My Questions:

 

 

1)Please let me know if we can do online migration of Lock LUN on HP SG version A.11.18.06 / A.11.18.01.

2)I understand that migration of SG cluster package's file system can be done online. 

 

 

Steps created for migration of Lock LUN and SG cluster package migration:

 

 

  1. 1.      Lock LUN Migration – Offline :

 

1.1.   Take server configuration backup:

#/opt/linuxdepots/tcs/scripts/linux_cfg_bkup.sh

 

      #cmviewcl > /root/backup/cmviewcl

      #cmviewcl –v > /root/backup/cmviewcl-v

      #cmviewconf > /root/backup/cmviewconf

 

1.2.   Storage team needs to assign the 1GB LUN for lock LUN for all the nodes of the cluster:

 

1.3.   Unix team need to scan 1GB LUN (Lock LUN) on all three servers:

 

#/opt/hp/hp_fibreutils/hp_rescan –a

#powermt check

#powermt config

#powermt save

 

Note: Now you can identify the new device. For say, we can assume new device name is /dev/emcpowerXY

 

1.4.   On Node 1, Please create one partition (id – 83):

The partition name will be /dev/emcpowerXY1

 

1.5.   On all three nodes, run below command and confirm if newly created partition is available on all nodes:

#partprobe

 

1.6.   Halt all packages, nodes and entire cluster:

 

1.6.1.      First we have to halt all the packages after getting confirmation from application team:

#cmhaltpkg –v <package_name>

 

1.6.2.      Second we have to halt all the nodes and stop entire cluster:

#cmhaltcl –v

 

1.6.3.      Verify if the cluster is stopped:

#cmviewcl

 

1.7.   On Node1, run following steps to edit cluster configuration and add new device as lock LUN:

#mkdir /root/storage_migration

#cd /root/storage_migration

#cmviewconf > cmviewconf_original

#cmgetconf -c <clsutername> <clustername.ascii>

      Note: Now edit the file<clustername.ascii>” and update new device for lock LUN for all nodes (for all nodes, lock LUN device name will be same).

      Example:

            Old:

            cluster lock lun name:            /dev/emcpowerl1

            New:

            cluster lock lun name:            /dev/emcpowerXY1

 

1.8.   Verify if we have edited properly:

#cmcheckconf -v -C <clustername.ascii>  ------------à (C is capital)

 

1.9.   Apply / Distribute new configuration changes to all nodes of the cluster:

#cmapplyconf -v -C anyname.ascii (C is capital)

 

1.10.              Now start the cluster and make all nodes join the cluster:

#cmruncl –v

Note: Now the cluster will be started and all nodes will join the cluster.

            Run below command to join particular node to the cluster if that node is joined:

#cmrunnode –v <hostname>

            Verify if cluster started and all the nodes joined the cluster:

            #cmviewcl

 

1.11.              Now start all the packages which were running earlier in respective nodes:

#cmrunpkg <package name> -n <node name>

 

Note: Packages cannot be started if the packages is not enabled on particular node. Run below command to enable the package.

#cmmodpkg –n <hostname> -e <package name>

 

1.12.              Verify if all the packages are started and running on respective nodes as same as before migration:

#cmviewcl      OR      cmviewcl -v

 

  1. 2.            Migrate Service Guard cluster packages file systems / volumes:

 

2.1.   Ensure to have full backup of the file system in place before you start SG cluster file system migration.

 

2.2.   Storage team needs to assign the required LUNs for SG cluster file system migration.

 

2.3.   Scan for new disks On Node - 1

# powermt display dev=all > /tmp/power.before.out

 

#opt/hp/hp_fibreutils/hp_rescan –a

 

# powermt config

 

#powermt save

 

# powermt check

 

# powermt display dev=all > /tmp/power.after.out

 

# diff /tmp/power.before.out /tmp/power.after.out > /tmp/diff.out

 

# grep -i "Logical Device" /tmp/diff.out   --------------à This command will show new LUN IDs.

 

NEW LUN id:

 

> Logical device ID=14B6 -- emcpowerd  -- 112

> Logical device ID=14B7 -- emcpowerf  -- 112

> Logical device ID=14B8 -- emcpowerc  -- 112

> Logical device ID=14B9 -- emcpowerg  -- 112

 

 

 

2.4.   Scan for new disks on Node 2 & 3.....

 

# powermt display dev=all > /tmp/power.before.out

 

#opt/hp/hp_fibreutils/hp_rescan –a

 

# powermt config

 

#powermt save

 

# powermt check

 

# powermt display dev=all > /tmp/power.after.out

 

# diff /tmp/power.before.out /tmp/power.after.out > /tmp/diff.out

 

# grep -i "Logical Device" /tmp/diff.out   --------------à This command will show new LUN IDs.

 

NEW LUN id:

 

> Logical device ID=14B6 -- emcpowerd  -- 112

> Logical device ID=14B7 -- emcpowerf  -- 112

> Logical device ID=14B8 -- emcpowerc  -- 112

> Logical device ID=14B9 -- emcpowerg  -- 112

 

 

2.5.   Create partitions on all new disks On Node - 1

# fdisk /dev/emcpowerXX

 

2.6.   Scan for partition changed On node 2 & 3

 

# partprobe

 

#fdisk –l </dev/emcpowerXX

 

 

2.7.   Do pvcreate in all new disks on Node 1

 

# pvcreate /dev/emcpowerXX1

 

 

 

2.8.   Extend VG with new disks on Node 1

 

#vgextend VolGroup01 /dev/emcpowerXX1 /dev/emcpowerXY1

 Volume group "VolGroup01" successfully extended

 

 

2.9.   Confirm the new LUNs are part of VG on Node 1:

 

# vgdisplay –v <VG name> | more

2.10.              Now mirror the existing disk to new disk on Node1:

# lvconvert -m1 --corelog /dev/<VG name>/<LV name> /dev/emcpowerXX1 /dev/emcpowerXY1

 

2.11.              Verify if mirroring is done 100% on Node1:

#lvs --noheadings /dev/<VG name>/<LV name>

 

2.12.              Split the mirror from old disks once the mirror completed on Node1:

 

# lvconvert -m0 /dev/<VG name >/<LV name> /dev/emcpowerYX1 /dev/emcpoweYY1

 (emcpowerYX1 and emcpoweYY1 are old disks)

 

2.13.              Verify the new LUNs are part of LV now on Node1

# lvdisplay -m /dev/<VG name> /<LV name>

 

2.14.              Run partprobe on all nodes to make above changes effective on other nodes:

#partprobe

 

2.15.              Verify if the cluster packages file systems are working fine on other nodes

Note: We need to halt the packages on running node and start the packages on other nodes one by one and verify if the package is working fine and packages file systems are getting mounted when the packages are started on other nodes. Once the package is test verified on all nodes. We have start the package back on original node.

# cmhaltpkg –v <package_name>

       #cmrunpkg –v –n <nodename> <packagename>

       #vgdisplay

        #lvdisplay   -----------à Need to check if lvdisplay is showing new device.

 

 

2.16.              Remove the old LUNs from the VG on node 1

Note : This should be done once you get confirmation from Application team and the LV which has been mirrored is running fine.

# vgreduce <VG name> /dev/emcpowerYX

 

2.17.              Remove the LVM header from old disks

# pvremove /dev/emcpowerYX1

 

2.18.              Remove the partitions from the old disks by fdisk command

 

2.19.              Ask SAN team to remove the disk after removed the disk from LVM and fdisk

 

2.20.              Once SAN confirmed that disk has been removed then remove the old path on all nodes by below commands

 

Linux ServiceGuard with DRBD

$
0
0

Hello.

 

I have to mount a Linux cluster with two storage arrays that are not connected via SAN.
Is this supported Serviceguard Linux with DRBD?

 

Thanks.

 

 

SG 11.20.22 RHEL 6.5 cmcheckconf error + Ora Tkit

$
0
0

hi i try to update our oracle env.

 

actual env:

Rhel 6.3

SG 11.20.10

Ora Tkit A05.01.01

 

after update:

RHEL 6.5

SG 11.20.22

Ora Tkit A05.01.12

 

update proccess or new installation works fine without any errors or warnings.

i can create/failover/start packages.

 

but if i do

/usr/local/cmcluster/bin/cmcheckconf -v -C /usr/local/cmcluster/conf/cmcluster.conf -p /usr/local/cmcluster/conf/pkg_ref_conf;

 

i´ve got many errors:

 

Begin cluster verification...
Checking cluster file: /usr/local/cmcluster/conf/cmcluster.conf.
MAX_CONFIGURED_PACKAGES configured to 300.
Checking nodes ... Done
Checking existing configuration ... Done
MAX_CONFIGURED_PACKAGES configured to 300.
Gathering storage information
Gathering network information
Beginning network probing (this may take a while)
Completed network probing
Cluster clu-rbg01 is an existing cluster
Begin checking consistency of files mentioned in cmclfiles2check
/usr/local/cmcluster/conf/cmclfiles2check is the same across nodes lnx-rbg09 lnx-rbg10
/etc/nsswitch.conf is the same across nodes lnx-rbg09 lnx-rbg10
/etc/services is the same across nodes lnx-rbg09 lnx-rbg10
/usr/local/cmcluster/conf/cmknowncmds is the same across nodes lnx-rbg09 lnx-rbg10
/usr/local/cmcluster/conf/cmcluster.rc is the same across nodes lnx-rbg09 lnx-rbg10
All files were the same on all nodes
"-p" option specified. Opening package reference file: /usr/local/cmcluster/conf/pkg_ref_conf.
Validating package poct1rbg via /usr/local/cmcluster/conf/scripts/mscripts/master_control_script.sh ...
Waiting for up to 300 seconds for the validation.
On node lnx-rbg09, validation of package poct1rbg succeeded with:
/var/tmp/master_control_script_env.poct1rbg.708031: line 84: SG_LNX-RBG09_PID[0]=: command not found
.....
.....
/var/tmp/master_control_script_env.poct1rbg.708031: line 99: SG_LNX-RBG10_PID[2]=: command not found
PR_mode:DISABLED
/var/tmp/master_control_script_env.poct1rbg.708031: line 84: SG_LNX-RBG09_PID[0]=: command not found
.....
.....
/var/tmp/master_control_script_env.poct1rbg.708031: line 99: SG_LNX-RBG10_PID[2]=: command not found
Verifying if the IP addresses could be added to corresponding subnets
/var/tmp/master_control_script_env.poct1rbg.708031: line 84: SG_LNX-RBG09_PID[0]=: command not found
/var/tmp/master_control_script_env.poct1rbg.708031: line 85: SG_LNX-RBG10_PID[0]=: command not found
/var/tmp/master_control_script_env.poct1rbg.708031: line 91: SG_LNX-RBG09_PID[1]=: command not found
/var/tmp/master_control_script_env.poct1rbg.708031: line 92: SG_LNX-RBG10_PID[1]=: command not found
/var/tmp/master_control_script_env.poct1rbg.708031: line 98: SG_LNX-RBG09_PID[2]=: command not found
/var/tmp/master_control_script_env.poct1rbg.708031: line 99: SG_LNX-RBG10_PID[2]=: command not found
Install directory operation mode selected.
Oracle Toolkit validation
The toolkit configuration file in the toolkit configuration directory will be backed up and a new file will be created when the package configuration is applied.
/var/tmp/master_control_script_env.poct1rbg.708031: line 84: SG_LNX-RBG09_PID[0]=: command not found
.....
.....
/var/tmp/master_control_script_env.poct1rbg.708031: line 99: SG_LNX-RBG10_PID[2]=: command not found
PR_mode:DISABLED
/var/tmp/master_control_script_env.poct1rbg.708031: line 84: SG_LNX-RBG09_PID[0]=: command not found
......
......
/var/tmp/master_control_script_env.poct1rbg.708031: line 99: SG_LNX-RBG10_PID[2]=: command not found
Verifying if the IP addresses could be added to corresponding subnets
/var/tmp/master_control_script_env.poct1rbg.708031: line 84: SG_LNX-RBG09_PID[0]=: command not found
/var/tmp/master_control_script_env.poct1rbg.708031: line 85: SG_LNX-RBG10_PID[0]=: command not found
/var/tmp/master_control_script_env.poct1rbg.708031: line 91: SG_LNX-RBG09_PID[1]=: command not found
/var/tmp/master_control_script_env.poct1rbg.708031: line 92: SG_LNX-RBG10_PID[1]=: command not found
/var/tmp/master_control_script_env.poct1rbg.708031: line 98: SG_LNX-RBG09_PID[2]=: command not found
/var/tmp/master_control_script_env.poct1rbg.708031: line 99: SG_LNX-RBG10_PID[2]=: command not found
Install directory operation mode selected.
Oracle Toolkit validation
The toolkit configuration file in the toolkit configuration directory will be backed up and a new file will be created when the package configuration is applied.
/var/tmp/master_control_script_env.poct1rbg.708031: line 84: SG_LNX-RBG09_PID[0]=: command not found
.....
.....
/var/tmp/master_control_script_env.poct1rbg.708031: line 99: SG_LNX-RBG10_PID[2]=: command not found
Validation for package poct1rbg succeeded via /usr/local/cmcluster/conf/scripts/mscripts/master_control_script.sh.
Validation for package poct5rbg succeeded via /usr/local/cmcluster/conf/scripts/mscripts/master_control_script.sh.
Maximum configured packages parameter is 300.
Modifying configuration on node lnx-rbg09
Modifying configuration on node lnx-rbg10
Modifying the cluster configuration for cluster clu-rbg01
Modifying the package configuration for package poct1rbg.
Modifying the package configuration for package poct5rbg.
Use the cmapplyconf command to apply the configuration.

 

Can anybody help me or have the same situation?

 

 

cmquerycl not work

$
0
0

Hi I try to config SG cluster with 2 nodes at RHEL 6 but when y try to execute the command cmquerycl show me the following error:

 

[root@wamas1 conf]# cmquerycl -v -C cluster.conf -n wamas1.rslapita.local -n wamas2.rslapita.local
Warning: Unable to determine local domain name for wamas1
Looking for other clusters ... Done
ERROR: Node wamas1.rslapita.local is refusing Serviceguard communication.
Please make sure that the proper security access is configured on node
wamas1.rslapita.local through either file-based authorization(pre-A.11.16 version) or
access control policies (version A.11.16 or higher)
and/or that the host name lookup on node wamas1.rslapita.local
resolves the IP address correctly.
Please make sure that cmclnodelist on node wamas1.rslapita.local has entries
for all the nodes in the cluster.
Failed to gather configuration information.

 

My cmclnodelist is configured:

 

[root@wamas1 conf]# cat cmclnodelist
wamas1.rslapita.local root
wamas2.rslapita.local root

 

My etc/hosts file:

 

192.168.105.35  wamas1 wamas1.rslapita.local
192.168.105.36  wamas2 wamas2.rslapita.local

 

What can I do?

 

Any body Help Me?

 

Regards

 

Victor


Cannot get disk partition for device /dev/sde Invalid data for cluster lock LUN configuration

$
0
0

I guys!

 

I have some trouble with my config.

 

I try to configure my sg cluster with 2 nodes, when try to check the config showme the following error.

 

 cmcheckconf -v -C cluster.conf
Begin cluster verification...
Checking cluster file: cluster.conf.
MAX_CONFIGURED_PACKAGES configured to 300.
Checking nodes ... Done
Checking existing configuration ... Done
MAX_CONFIGURED_PACKAGES configured to 300.
Gathering storage information
Found 1 devices on node wamas1
Found 1 devices on node wamas2
Analysis of 2 devices should take approximately 1 seconds
0%----10%----20%----30%----40%----50%----60%----70%----80%----90%----100%
Gathering network information
Beginning network probing (this may take a while)
Completed network probing
Cannot get disk partition for device /dev/sdeInvalid data for cluster lock LUN configuration
Maximum configured packages parameter is 300.
cmcheckconf: Unable to verify cluster file: cluster.conf.

 

My cluster.conf

NODE_NAME               wamas1
  NETWORK_INTERFACE     eth1
    HEARTBEAT_IP        xx.xx.xx.xx
  NETWORK_INTERFACE     eth5
    HEARTBEAT_IP       yy.yy.yy.yy
  NETWORK_INTERFACE     bond0
    STATIONARY_IP       zz.zz.zz.zz
  CLUSTER_LOCK_LUN      /dev/sde


NODE_NAME               wamas2
    NETWORK_INTERFACE     eth1
    HEARTBEAT_IP        xx.xx.xx.xx
  NETWORK_INTERFACE     eth5
    HEARTBEAT_IP       yy.yy.yy.yy
  NETWORK_INTERFACE     bond0
    STATIONARY_IP       zz.zz.zz.zz
  CLUSTER_LOCK_LUN      /dev/sde

 

[root@wamas1 conf]# fdisk -l /dev/sde

Disk /dev/sde: 499 MB, 499974144 bytes
16 heads, 60 sectors/track, 1017 cylinders
Units = cylinders of 960 * 512 = 491520 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 1048576 bytes
Disk identifier: 0xa71b4a14

   Device Boot      Start         End      Blocks   Id  System
/dev/sde1               1        1017      488130   83  Linux

 

What Can I Do?

 

Regards,

 

VB

Serviceguard on non-HP servers

$
0
0

Hi,

 

We intend to migrate the HP-UX environment to RHEL. Unfortunately, the Linux environemnt is not running on HP servers which raises the question, if it is supported at all to run SGLX on non-HP x86 servers?

 

The support matrix does only list HP gear for the server part...

 

Thanks in advance!

Service command /etc/opt/omni/server/sg/csfailover.ksh is not executable

$
0
0

Created SG 12.0 on RHEL 6.6 x64

 

Created a failover package for virtual ip address

 

Created a failover package for file system (/fs1). It will be mounted only one node at a time.

 

Installed DP on shared file system on both nodes.

 

Then tried to create a failover package for DP service.

 

During cmcheckconf getting the below error.

 

Because sg is expecting the file "/etc/opt/omni/server/sg/csfailover.ksh" on both nodes. But it will be availble on mounted file system node only, here file system is mounted in host1. Please helm to solve this issue.

 

Note: But SG 11.20 it worked fine.

 

 

On node <host2>, validation of package ob2cl failed with:
PR_mode:DISABLED
ERROR: Function sg_validate_services
Service command /etc/opt/omni/server/sg/csfailover.ksh is not executable
ERROR: Failed to validate /usr/local/cmcluster/conf/scripts/sg/service.sh
On node <host2>, validation of package ob2cl succeeded with:
PR_mode:DISABLED
cmcheckconf: Error found in package file: /usr/local/cmcluster/conf/pkg/ob2cl/ob2cl.conf.

service guard manager error,unable to create package pkg_name,need help

$
0
0

 while creating a general package with out any extra tool kit service guard manager shows unable to create package error.It's in redhat linux 6.5

 

package not getting up in node

$
0
0

I am using two node ans hp servigurd is running on it

 

 yeaterday i restarted my server after My package is not getting up please help

 

CLUSTER        STATUS
dbcluster      up

  NODE           STATUS       STATE
  dbnode01       up           running
  dbnode02       up           running

UNOWNED_PACKAGES

    PACKAGE        STATUS        STATE         AUTO_RUN     NODE
    dbpkg          down          halted        disabled     unowned
[root@dbnode02 ~]#

 

 

Earliear i removed the mounted partion from my node, yeaterday i restarted my server after package is not getting UP pls help.

 

 

P.S. This thread has been moved from HP-UX > Serviceguard to Linux > Serviceguard. - Hp Forum Moderator

Viewing all 361 articles
Browse latest View live


<script src="https://jsc.adskeeper.com/r/s/rssing.com.1596347.js" async> </script>