Home » Server Options » RAC & Failsafe » Wrong step in Clusterware Removal (11.2.0.1 SLES 11 SP3)
Wrong step in Clusterware Removal [message #614329] Tue, 20 May 2014 22:07 Go to next message
juniordbanewbie
Messages: 250
Registered: April 2014
Senior Member
Hi,

I'm currently removing a clusterware on a two node RAC installation.

from http://docs.oracle.com/cd/E11882_01/rac.112/e16794/adddelclusterware.htm#CWADD90992

Quote:


Run the following command as either root or the user that installed Oracle Clusterware to determine whether the node you want to delete is active and whether it is pinned:

$ olsnodes -s -t
If the node is pinned, then run the crsctl unpin css command. Otherwise, proceed to the next step.


here's the output I have


grid@ORCL01:olsnodes -s -t
orcl01 Active Unpinned ->this is probably the cause of every thing
orcl02 Active Unpinned


After completing grid@ORCL02:$ORACLE_HOME/deinstall/runInstaller -detachHome ORACLE_HOME=$GRID_HOMR -silent -local, I've the following summary

CLEAN OPERATION SUMMARY
Following RAC listeners(s) were de-configured successfully: LISTENER_SCAN3, LISTENER_SCAN2, LISTENER_SCAN1
Oracle Clusterware was already stopped and de-configured on node "ORCL02"
Oracle Clusterware is already stopped and de-configured on node "ORCL01"
Oracle Clusterware is stopped and de-configured successfully
Sucessfully detached Oracle home '/u01/app/11.2.0/grid' from the central inventory on that local node.
Sucessfully deleted directory '/u01/app/11.2.0/grid' on the local node.
Sucessfully deleted directory '/u01/app/grid' on the local node.
Oracle install successfully cleaned up the temporary directories.
grid@ORCL02


Btw by the I run step 8 I encounter the following error

grid@ORCL01:~> cluvfy stage -help

ERROR:
Oracle Grid Infrastructure not configured.
You cannot run '/u01/app/11.2.0/grid/bin/cluvfy' without the Oracle Grid Infrastructure.

log from $GRID_HOME/orcl01/alertorcl01.log

[crsd(10950)]CRS-5504:Node down event reported for node 'orcl02'.
2014-05-14 19:23:57.994
[crsd(10950)]CRS-2773:Server 'orcl02' has been removed from pool 'Generic'.
2014-05-14 19:23:57.994
[crsd(10950)]CRS-2773:Server 'orcl02' has been removed from pool 'ora.RECOV'.
2014-05-14 19:23:57.994
[crsd(10950)]CRS-2787:Server pool 'ora.RECOV' has fallen below its minimum size. Details at (:CRSPE00140:) in /u01/app/11.2.0/grid/log/orcl01/crsd/crsd.log.
[client(18183)]CRS-10001:ADVM/ACFS is not supported on SUSE
[client(18185)]CRS-10001:ACFS-9201: Not Supported
2014-05-19 18:38:41.732
[/u01/app/11.2.0/grid/bin/orarootagent.bin(11204)]CRS-5822:Agent '/u01/app/11.2.0/grid/bin/orarootagent_root' disconnected from server. Details at (:CRSAGF00117:) in /u01/app/11.2.0/grid/log/orcl01/agent/crsd/orarootagent_root/orarootagent_root.log.
2014-05-19 18:38:42.652
[ctssd(9216)]CRS-2405:The Cluster Time Synchronization Service on host orcl01 is shutdown by user
2014-05-19 18:38:55.451
[cssd(8534)]CRS-1603:CSSD on node orcl01 shutdown by user.
[client(1355)]CRS-10001:ADVM/ACFS is not supported on SUSE


What should I do to configure clusterware on orcl01?

thanks in advance!

[Updated on: Tue, 20 May 2014 23:24]

Report message to a moderator

Re: Wrong step in Clusterware Removal [message #614341 is a reply to message #614329] Wed, 21 May 2014 01:27 Go to previous messageGo to next message
John Watson
Messages: 8922
Registered: January 2010
Location: Global Village
Senior Member
It looks as though you have de-configured the clusterware. Is that not what you wanted to do? If there are any Oracle-related files or directories left over, remove them.
Re: Wrong step in Clusterware Removal [message #614343 is a reply to message #614341] Wed, 21 May 2014 01:40 Go to previous messageGo to next message
juniordbanewbie
Messages: 250
Registered: April 2014
Senior Member
Yes I've deconfigured the clusterware, but unfortunately I deconfigure clusterware on both nodes, which is not what I want. I've probably unpinned both orcl01 and orcl02 as the olsnodes -s -t command shown. sorry for my poor English:
I'm currently removing a clusterware on a two node RAC installation.=> what I means is remove a clusterware node on a two node RAC. http://docs.oracle.com/cd/E11882_01/rac.112/e16794/adddelclusterware.htm#CWADD90992

have done a $GRID_HOME/crs/install/rootcrs.pl -h

it does not have a config option.

inside $GRID_HOME/root.sh

/u01/app/11.2.0/grid/perl/bin/perl -I/u01/app/11.2.0/grid/perl/lib -I/u01/app/11.2.0/grid/crs/install  /u01/app/11.2.0/grid/crs/install/rootcrs.pl

 if [ $? -eq 0 ] ; then echo Updating inventory properties for clusterware;
    su grid -c "/u01/app/11.2.0/grid/perl/bin/perl /u01/app/11.2.0/grid/install/utl/updnodelist.pl" ; else exit $?; fi


it again points to rootcrs.pl

I'm not too sure about how to go from here.

[Updated on: Wed, 21 May 2014 02:05]

Report message to a moderator

Re: Wrong step in Clusterware Removal [message #614346 is a reply to message #614343] Wed, 21 May 2014 02:03 Go to previous messageGo to next message
John Watson
Messages: 8922
Registered: January 2010
Location: Global Village
Senior Member
First, you haven't pinned or unpinned anything. The olsnodes utility reports things, it doesn't change them.

Second, if I understand the posts correctly, you need to tidy everything up and install again. IS that a problem?
Re: Wrong step in Clusterware Removal [message #614347 is a reply to message #614346] Wed, 21 May 2014 02:07 Go to previous messageGo to next message
juniordbanewbie
Messages: 250
Registered: April 2014
Senior Member
If I run $GRID_HOME/root.sh again, will I lost all the data in the ASM disks. Looks like I've learnt a very heavy lesson.

[Updated on: Wed, 21 May 2014 02:27]

Report message to a moderator

Re: Wrong step in Clusterware Removal [message #614355 is a reply to message #614347] Wed, 21 May 2014 02:37 Go to previous messageGo to next message
John Watson
Messages: 8922
Registered: January 2010
Location: Global Village
Senior Member
You should be able to keep the data on your ASM devices. I would create one new device for your OCR and vote file, create an external redundancy group for it during the install. The new ASM should recognize the existing disks and groups, so if you create new database instances they should be able to use existing database files.
An interesting exercise.
Re: Wrong step in Clusterware Removal [message #614369 is a reply to message #614355] Wed, 21 May 2014 03:57 Go to previous message
juniordbanewbie
Messages: 250
Registered: April 2014
Senior Member
when I run su -c '$GRID_HOME/root.sh'

2014-05-21 16:26:11: Successfully created CRS resources for cluster daemon and ASM
2014-05-21 16:26:11: Checking if initial configuration has been performed
2014-05-21 16:26:11: Starting CSS in exclusive mode
2014-05-21 16:26:45: CRS-2672: Attempting to start 'ora.gipcd' on 'orcl01'
2014-05-21 16:26:45: CRS-2672: Attempting to start 'ora.mdnsd' on 'orcl01'
2014-05-21 16:26:45: CRS-2676: Start of 'ora.gipcd' on 'orcl01' succeeded
2014-05-21 16:26:45: CRS-2676: Start of 'ora.mdnsd' on 'orcl01' succeeded
2014-05-21 16:26:45: CRS-2672: Attempting to start 'ora.gpnpd' on 'orcl01'
2014-05-21 16:26:45: CRS-2676: Start of 'ora.gpnpd' on 'orcl01' succeeded
2014-05-21 16:26:45: CRS-2672: Attempting to start 'ora.cssdmonitor' on 'orcl01'
2014-05-21 16:26:45: CRS-2676: Start of 'ora.cssdmonitor' on 'orcl01' succeeded
2014-05-21 16:26:45: CRS-2672: Attempting to start 'ora.cssd' on 'orcl01'
2014-05-21 16:26:45: CRS-2672: Attempting to start 'ora.diskmon' on 'orcl01'
2014-05-21 16:26:45: CRS-2676: Start of 'ora.diskmon' on 'orcl01' succeeded
2014-05-21 16:26:45: CRS-2676: Start of 'ora.cssd' on 'orcl01' succeeded
2014-05-21 16:26:45: Querying for existing CSS voting disks
2014-05-21 16:26:45: Found 3 configured voting files
2014-05-21 16:26:45: Existing configuration setup found
2014-05-21 16:26:45: Exiting exclusive mode
2014-05-21 16:26:45: Command return code of 1 (256) from command: /u01/app/11.2.0/grid/bin/crsctl stop resource ora.crsd -init
2014-05-21 16:26:45: Stop of resource "ora.crsd -init" failed
2014-05-21 16:26:45: Failed to stop CRSD
2014-05-21 16:26:45: Command return code of 1 (256) from command: /u01/app/11.2.0/grid/bin/crsctl stop resource ora.asm -init
2014-05-21 16:26:45: Stop of resource "ora.asm -init" failed
2014-05-21 16:26:45: Failed to stop ASM
2014-05-21 16:26:45: Command return code of 1 (256) from command: /u01/app/11.2.0/grid/bin/crsctl stop resource ora.ctssd -init
2014-05-21 16:26:45: Stop of resource "ora.ctssd -init" failed
2014-05-21 16:26:45: Failed to stop OCTSSD
2014-05-21 16:26:56: Initial cluster configuration failed.  See /u01/app/11.2.0/grid/cfgtoollogs/crsconfig/rootcrs_orcl01.log


I discover the following error from /u01/app/11.2.0/grid/log/orcl01/client/crsctl.log


2014-05-19 19:27:06.639: [ default][1111230240]utgdv:2:ocr loc file /etc/oracle/olr.loc cannot be opened. errno 2
2014-05-19 19:27:06.639: [ default][1111230240]utgdv:2:ocr loc file /etc/oracle/ocr.loc cannot be opened. errno 2
2014-05-19 19:27:06.639: [  CRSCTL][1111230240]Command::checkConfig: clsugetconf returned no configuration
2014-05-19 19:36:55.383: [ default][2132657952]utgdv:2:ocr loc file /etc/oracle/olr.loc cannot be opened. errno 2
2014-05-21 10:21:03.016: [ default][3936151328]utgdv:2:ocr loc file /etc/oracle/olr.loc cannot be opened. errno 2
2014-05-21 10:21:03.017: [ default][3936151328]utgdv:2:ocr loc file /etc/oracle/ocr.loc cannot be opened. errno 2
2014-05-21 10:21:03.017: [  CRSCTL][3936151328]Command::checkConfig: clsugetconf returned no configuratio


I don't know why it says cannot open ocr.loc
grid@ORCL01:~> ls -l /etc/oracle/ocr.loc
-rw-r--r-- 1 root oinstall 41 May 21 16:25 /etc/oracle/ocr.loc


grid@ORCL01:~> ls -l /dev/asm-ocr-vote*
lrwxrwxrwx 1 root root 5 May 21 16:26 /dev/asm-ocr-vote1 -> dm-22
lrwxrwxrwx 1 root root 5 May 21 16:26 /dev/asm-ocr-vote1_part1 -> dm-38
lrwxrwxrwx 1 root root 5 May 21 16:26 /dev/asm-ocr-vote2 -> dm-23
lrwxrwxrwx 1 root root 5 May 21 16:26 /dev/asm-ocr-vote2_part1 -> dm-39
lrwxrwxrwx 1 root root 5 May 21 16:26 /dev/asm-ocr-vote3 -> dm-24
lrwxrwxrwx 1 root root 5 May 21 16:26 /dev/asm-ocr-vote3_part1 -> dm-43

grid@ORCL01:~> ls -l /dev/dm-3[8-9] /dev/dm-43

brw-rw---- 1 grid oinstall 253, 38 May 21 16:26 /dev/dm-38
brw-rw---- 1 grid oinstall 253, 39 May 21 16:26 /dev/dm-39
brw-rw---- 1 grid oinstall 253, 43 May 21 16:26 /dev/dm-43



From the look of it, from the above log, can I salvage the existing configuration or do I have to reinstall again.

Existing Fact is orcl01 is deconfigured non intentionally but not not deinstall

If I've to reinstall everything but can I reinstall as grid2. because if I deinstall as grid then reinstall as grid I've high chances of getting deleting ASM disk right?

How about the existing Database installed on ORCL01. They won't be tied to the new GRID infrastructure right?

thanks

[Updated on: Wed, 21 May 2014 04:38]

Report message to a moderator

Previous Topic: ASM mirroring
Next Topic: clusterware
Goto Forum:
  


Current Time: Thu Mar 28 21:33:53 CDT 2024