RACnode communication problem?

  • From: "Walker, Jed S" <Jed_Walker@xxxxxxxxxxxxxxxxx>
  • To: "oracle-l@xxxxxxxxxxxxx" <oracle-l@xxxxxxxxxxxxx>
  • Date: Wed, 28 Sep 2011 15:51:35 +0000

Hi,
I'm trying to install RAC 11.2.0 - Linux-x86-64, RedHat 5.3, oracle 11.2.0.3.0 
(fyi - I also tried with 11.2.0.2.0 and got the same result)

Has anyone seen this before, and have a solution? runcluvfy.sh and runInstaller 
complain that the nodes can't talk to each other, but I can ssh (passwordless) 
between them and oracle has even copied files from node1 to the other nodes in 
/tmp. I believe there is nothing wrong, but need to figure out why oracle 
suddenly thinks something is wrong.

BTW, yesterday, it wasn't complaining at all, but the servers were rebooted 
last night. It works fine when I do things manually, but oracle doesn't seem to 
think it works. I'm hoping I'm either doing something "RAC newbie" or it is 
just a weird oracle thing

Here is what runcluvfy says (and runInstaller)
(note: yes, changed the a/b on network so I don't get in trouble)

[oracle@flux-rac-node-wcdp-01 grid]$ ./runcluvfy.sh stage -pre crsinst -n 
flux-rac-node-wcdp-01,flux-rac-node-wcdp-02

Performing pre-checks for cluster services setup

Checking node reachability...
Node reachability check passed from node "flux-rac-node-wcdp-01"


Checking user equivalence...
User equivalence check passed for user "oracle"

Checking node connectivity...

Checking hosts config file...

Verification of the hosts config file successful


WARNING:
Make sure IP address "bond0 : 99.99.230.195 [99.99.230.128] " is up and is a 
valid IP address on node "flux-rac-node-wcdp-02"

WARNING:
Make sure IP address "bond0 : 99.99.230.194 [99.99.230.128] " is up and is a 
valid IP address on node "flux-rac-node-wcdp-01"

ERROR:
PRVF-7616 : Node connectivity failed for subnet "99.99.230.128" between 
"flux-rac-node-wcdp-02 - bond0 : 99.99.230.195" and "flux-rac-node-wcdp-01 - 
bond0 : 99.99.230.194"
Checking multicast communication...

Checking subnet "99.99.230.128" for multicast communication with multicast 
group "230.0.1.0"...

I've checked ifconfig and everything looks fine, and here you can see I can ssh 
between them (and yes, I've done it both ways). Also, the runcluvfy.sh 
apparently is able to connect over because when I run it, node2 gets oracle 
files created in /tmp.

[oracle@flux-rac-node-wcdp-02 ~]$ ssh flux-rac-node-wcdp-02 ls -l /tmp
WARNING:

  This system is solely for  ...
total 4
drwxr-xr-x 3 oracle dba 4096 Sep 28 15:36 CVU_11.2.0.3.0_oracle


--
//www.freelists.org/webpage/oracle-l


Other related posts: