maandag 25 augustus 2014

Online Upgrading Oracle RAC from linux 5 to 6 (3/3)

11 Appendix 1: Output: ./cluvfy stage -pre nodeadd



#############################################################
Output: ./cluvfy stage -pre nodeadd -n m-lrkdb3
#############################################################


[grid@m-lrkdb2:bin]$ ./cluvfy stage -pre nodeadd -n m-lrkdb3

Performing pre-checks for node addition

Checking node reachability...
Node reachability check passed from node "m-lrkdb2"


Checking user equivalence...
User equivalence check passed for user "grid"

Checking CRS integrity...

CRS integrity check passed

Clusterware version consistency passed.

Checking shared resources...

Checking CRS home location...
Path "/u01/app/11.2.0/grid_3" either already exists or can be successfully created on nodes: "m-lrkdb3"
Shared resources check for node addition passed


Checking node connectivity...

Checking hosts config file...

Verification of the hosts config file successful

Check: Node connectivity using interfaces on subnet "10.19.62.0"
Node connectivity passed for subnet "10.19.62.0" with node(s) m-lrkdb3,m-lrkdb2
TCP connectivity check passed for subnet "10.19.62.0"


Check: Node connectivity using interfaces on subnet "192.168.1.0"
Node connectivity passed for subnet "192.168.1.0" with node(s) m-lrkdb2,m-lrkdb3
TCP connectivity check passed for subnet "192.168.1.0"

Checking subnet mask consistency...
Subnet mask consistency check passed for subnet "10.19.62.0".
Subnet mask consistency check passed for subnet "192.168.1.0".
Subnet mask consistency check passed.

Node connectivity check passed

Checking multicast communication...

Checking subnet "192.168.1.0" for multicast communication with multicast group "224.0.0.251"...
Check of subnet "192.168.1.0" for multicast communication with multicast group "224.0.0.251" passed.

Check of multicast communication passed.
Total memory check passed
Available memory check passed
Swap space check failed
Check failed on nodes:
m-lrkdb2,m-lrkdb3
Free disk space check failed for "m-lrkdb2:/usr,m-lrkdb2:/var,m-lrkdb2:/etc,m-lrkdb2:/u01/app/11.2.0/grid_3,m-lrkdb2:/sbin,m-lrkdb2:/tmp"
Check failed on nodes:
m-lrkdb2
Free disk space check failed for "m-lrkdb3:/usr,m-lrkdb3:/var,m-lrkdb3:/etc,m-lrkdb3:/u01/app/11.2.0/grid_3,m-lrkdb3:/sbin,m-lrkdb3:/tmp"
Check failed on nodes:
m-lrkdb3
Check for multiple users with UID value 502 passed
User existence check passed for "grid"
Run level check passed
Hard limits check passed for "maximum open file descriptors"
Soft limits check passed for "maximum open file descriptors"
Hard limits check passed for "maximum user processes"
Soft limits check passed for "maximum user processes"
System architecture check passed

WARNING:
PRVF-7524 : Kernel version is not consistent across all the nodes.
Kernel version = "2.6.18-164.el5" found on nodes: m-lrkdb2.
Kernel version = "2.6.39-200.24.1.el6uek.x86_64" found on nodes: m-lrkdb3.
Kernel version check passed
Kernel parameter check passed for "semmsl"
Kernel parameter check passed for "semmns"
Kernel parameter check passed for "semopm"
Kernel parameter check passed for "semmni"
Kernel parameter check passed for "shmmax"
Kernel parameter check passed for "shmmni"
Kernel parameter check passed for "shmall"
Kernel parameter check passed for "file-max"
Kernel parameter check passed for "ip_local_port_range"
Kernel parameter check passed for "rmem_default"
Kernel parameter check passed for "rmem_max"
Kernel parameter check passed for "wmem_default"
Kernel parameter check passed for "wmem_max"
Kernel parameter check passed for "aio-max-nr"
Package existence check passed for "make"
Package existence check passed for "binutils"
Package existence check passed for "gcc(x86_64)"
Package existence check passed for "libaio(x86_64)"
Package existence check passed for "glibc(x86_64)"
Package existence check passed for "compat-libstdc++-33(x86_64)"
Package existence check passed for "glibc-devel(x86_64)"
Package existence check passed for "gcc-c++(x86_64)"
Package existence check passed for "libaio-devel(x86_64)"
Package existence check passed for "libgcc(x86_64)"
Package existence check passed for "libstdc++(x86_64)"
Package existence check passed for "libstdc++-devel(x86_64)"
Package existence check passed for "sysstat"
Package existence check passed for "ksh"
Package existence check passed for "nfs-utils"
Check for multiple users with UID value 0 passed
Current group ID check passed

Starting check for consistency of primary group of root user

Check for consistency of root user's primary group passed
Group existence check passed for "asmadmin"
Group existence check passed for "asmdba"

Checking OCR integrity...

OCR integrity check passed

Checking Oracle Cluster Voting Disk configuration...

Oracle Cluster Voting Disk configuration check passed
Time zone consistency check passed

Starting Clock synchronization checks using Network Time Protocol(NTP)...
NTP configuration file "/etc/ntp.conf" existence check passed
Liveness check passed for "ntpd"
Check for NTP daemon or service alive passed on all nodes
Check of common NTP Time Server passed
Clock time offset check passed

Clock synchronization check using Network Time Protocol(NTP) passed


User "grid" is not part of "root" group. Check passed
Checking integrity of file "/etc/resolv.conf" across nodes

"domain" and "search" entries do not coexist in any "/etc/resolv.conf" file
All nodes have same "search" order defined in file "/etc/resolv.conf"
The DNS response time for an unreachable node is within acceptable limit on all nodes

Check for integrity of file "/etc/resolv.conf" passed

Checking integrity of name service switch configuration file "/etc/nsswitch.conf" ...
Check for integrity of name service switch configuration file "/etc/nsswitch.conf" passed


Pre-check for node addition was unsuccessful on all the nodes.


12 Appendix 2: Output: ./cluvfy stage -pre crsinst


#############################################################
Output: ./cluvfy stage -pre crsinst -n m-lrkdb2,m-lrkdb3 -verbose
#############################################################


[grid@m-lrkdb1:bin]$ ./cluvfy stage -pre crsinst -n m-lrkdb1,m-lrkdb3 -fixup -verbose

Performing pre-checks for cluster services setup

Checking node reachability...

Check: Node reachability from node "m-lrkdb1"
Destination Node Reachable?
------------------------------------ ------------------------
m-lrkdb1 yes
m-lrkdb3 yes
Result: Node reachability check passed from node "m-lrkdb1"


Checking user equivalence...

Check: User equivalence for user "grid"
Node Name Status
------------------------------------ ------------------------
m-lrkdb1 passed
m-lrkdb3 passed
Result: User equivalence check passed for user "grid"

Checking node connectivity...

Checking hosts config file...
Node Name Status
------------------------------------ ------------------------
m-lrkdb1 passed
m-lrkdb3 passed

Verification of the hosts config file successful


Interface information for node "m-lrkdb1"
Name IP Address Subnet Gateway Def. Gateway HW Address MTU
------ --------------- --------------- --------------- --------------- ----------------- ------
eth0 10.19.62.64 10.19.62.0 0.0.0.0 10.19.62.254 00:50:56:A6:00:3C 1500
eth0 10.19.62.66 10.19.62.0 0.0.0.0 10.19.62.254 00:50:56:A6:00:3C 1500
eth0 10.19.62.67 10.19.62.0 0.0.0.0 10.19.62.254 00:50:56:A6:00:3C 1500
eth0 10.19.62.49 10.19.62.0 0.0.0.0 10.19.62.254 00:50:56:A6:00:3C 1500
eth1 192.168.1.133 192.168.1.0 0.0.0.0 10.19.62.254 00:50:56:A6:00:3E 1500
eth1 169.254.162.124 169.254.0.0 0.0.0.0 10.19.62.254 00:50:56:A6:00:3E 1500
eth2 192.168.0.133 192.168.0.0 0.0.0.0 10.19.62.254 00:50:56:A6:00:3F 1500


Interface information for node "m-lrkdb3"
Name IP Address Subnet Gateway Def. Gateway HW Address MTU
------ --------------- --------------- --------------- --------------- ----------------- ------
eth0 10.19.62.53 10.19.62.0 0.0.0.0 10.19.62.254 00:50:56:8F:44:FC 1500
eth0 10.19.62.86 10.19.62.0 0.0.0.0 10.19.62.254 00:50:56:8F:44:FC 1500
eth1 192.168.1.135 192.168.1.0 0.0.0.0 10.19.62.254 00:50:56:8F:40:F8 1500
eth2 192.168.0.135 192.168.0.0 0.0.0.0 10.19.62.254 00:50:56:8F:F3:9A 1500


Check: Node connectivity using interfaces on subnet "10.19.62.0"

Check: Node connectivity of subnet "10.19.62.0"
Source Destination Connected?
------------------------------ ------------------------------ ----------------
m-lrkdb1[10.19.62.64] m-lrkdb1[10.19.62.67] yes
m-lrkdb1[10.19.62.64] m-lrkdb1[10.19.62.49] yes
m-lrkdb1[10.19.62.64] m-lrkdb1[10.19.62.66] yes
m-lrkdb1[10.19.62.64] m-lrkdb3[10.19.62.53] yes
m-lrkdb1[10.19.62.64] m-lrkdb3[10.19.62.86] yes
m-lrkdb1[10.19.62.67] m-lrkdb1[10.19.62.49] yes
m-lrkdb1[10.19.62.67] m-lrkdb1[10.19.62.66] yes
m-lrkdb1[10.19.62.67] m-lrkdb3[10.19.62.53] yes
m-lrkdb1[10.19.62.67] m-lrkdb3[10.19.62.86] yes
m-lrkdb1[10.19.62.49] m-lrkdb1[10.19.62.66] yes
m-lrkdb1[10.19.62.49] m-lrkdb3[10.19.62.53] yes
m-lrkdb1[10.19.62.49] m-lrkdb3[10.19.62.86] yes
m-lrkdb1[10.19.62.66] m-lrkdb3[10.19.62.53] yes
m-lrkdb1[10.19.62.66] m-lrkdb3[10.19.62.86] yes
m-lrkdb3[10.19.62.53] m-lrkdb3[10.19.62.86] yes
Result: Node connectivity passed for subnet "10.19.62.0" with node(s) m-lrkdb1,m-lrkdb3


Check: TCP connectivity of subnet "10.19.62.0"
Source Destination Connected?
------------------------------ ------------------------------ ----------------
m-lrkdb1:10.19.62.64 m-lrkdb1:10.19.62.67 passed
m-lrkdb1:10.19.62.64 m-lrkdb1:10.19.62.49 passed
m-lrkdb1:10.19.62.64 m-lrkdb1:10.19.62.66 passed
m-lrkdb1:10.19.62.64 m-lrkdb3:10.19.62.53 passed
m-lrkdb1:10.19.62.64 m-lrkdb3:10.19.62.86 passed
Result: TCP connectivity check passed for subnet "10.19.62.0"


Check: Node connectivity using interfaces on subnet "192.168.1.0"

Check: Node connectivity of subnet "192.168.1.0"
Source Destination Connected?
------------------------------ ------------------------------ ----------------
m-lrkdb3[192.168.1.135] m-lrkdb1[192.168.1.133] yes
Result: Node connectivity passed for subnet "192.168.1.0" with node(s) m-lrkdb3,m-lrkdb1


Check: TCP connectivity of subnet "192.168.1.0"
Source Destination Connected?
------------------------------ ------------------------------ ----------------
m-lrkdb3:192.168.1.135 m-lrkdb1:192.168.1.133 passed
Result: TCP connectivity check passed for subnet "192.168.1.0"

Checking subnet mask consistency...
Subnet mask consistency check passed for subnet "10.19.62.0".
Subnet mask consistency check passed for subnet "192.168.1.0".
Subnet mask consistency check passed.

Result: Node connectivity check passed

Checking multicast communication...

Checking subnet "192.168.1.0" for multicast communication with multicast group "224.0.0.251"...
Check of subnet "192.168.1.0" for multicast communication with multicast group "224.0.0.251" passed.

Check of multicast communication passed.

Check: Total memory
Node Name Available Required Status
------------ ------------------------ ------------------------ ----------
m-lrkdb1 11.7301GB (1.2299908E7KB) 4GB (4194304.0KB) passed
m-lrkdb3 11.7598GB (1.2331028E7KB) 4GB (4194304.0KB) passed
Result: Total memory check passed

Check: Available memory
Node Name Available Required Status
------------ ------------------------ ------------------------ ----------
m-lrkdb1 6.9644GB (7302676.0KB) 50MB (51200.0KB) passed
m-lrkdb3 11.5275GB (1.2087508E7KB) 50MB (51200.0KB) passed
Result: Available memory check passed

Check: Swap space
Node Name Available Required Status
------------ ------------------------ ------------------------ ----------
m-lrkdb1 9.9968GB (1.0482372E7KB) 11.7301GB (1.2299908E7KB) failed
m-lrkdb3 9.999GB (1.0484732E7KB) 11.7598GB (1.2331028E7KB) failed
Result: Swap space check failed

Check: Free disk space for "m-lrkdb1:/usr,m-lrkdb1:/var,m-lrkdb1:/etc,m-lrkdb1:/u01/app/11.2.0/grid_3,m-lrkdb1:/sbin,m-lrkdb1:/tmp"
Path Node Name Mount point Available Required Status
---------------- ------------ ------------ ------------ ------------ ------------
/usr m-lrkdb1 / 4.1719GB 7.9635GB failed
/var m-lrkdb1 / 4.1719GB 7.9635GB failed
/etc m-lrkdb1 / 4.1719GB 7.9635GB failed
/u01/app/11.2.0/grid_3 m-lrkdb1 / 4.1719GB 7.9635GB failed
/sbin m-lrkdb1 / 4.1719GB 7.9635GB failed
/tmp m-lrkdb1 / 4.1719GB 7.9635GB failed
Result: Free disk space check failed for "m-lrkdb1:/usr,m-lrkdb1:/var,m-lrkdb1:/etc,m-lrkdb1:/u01/app/11.2.0/grid_3,m-lrkdb1:/sbin,m-lrkdb1:/tmp"

Check: Free disk space for "m-lrkdb3:/usr,m-lrkdb3:/var,m-lrkdb3:/etc,m-lrkdb3:/u01/app/11.2.0/grid_3,m-lrkdb3:/sbin,m-lrkdb3:/tmp"
Path Node Name Mount point Available Required Status
---------------- ------------ ------------ ------------ ------------ ------------
/usr m-lrkdb3 / 6.4492GB 7.9635GB failed
/var m-lrkdb3 / 6.4492GB 7.9635GB failed
/etc m-lrkdb3 / 6.4492GB 7.9635GB failed
/u01/app/11.2.0/grid_3 m-lrkdb3 / 6.4492GB 7.9635GB failed
/sbin m-lrkdb3 / 6.4492GB 7.9635GB failed
/tmp m-lrkdb3 / 6.4492GB 7.9635GB failed
Result: Free disk space check failed for "m-lrkdb3:/usr,m-lrkdb3:/var,m-lrkdb3:/etc,m-lrkdb3:/u01/app/11.2.0/grid_3,m-lrkdb3:/sbin,m-lrkdb3:/tmp"

Check: User existence for "grid"
Node Name Status Comment
------------ ------------------------ ------------------------
m-lrkdb1 passed exists(502)
m-lrkdb3 passed exists(502)

Checking for multiple users with UID value 502
Result: Check for multiple users with UID value 502 passed
Result: User existence check passed for "grid"

Check: Group existence for "oinstall"
Node Name Status Comment
------------ ------------------------ ------------------------
m-lrkdb1 passed exists
m-lrkdb3 passed exists
Result: Group existence check passed for "oinstall"

Check: Group existence for "dba"
Node Name Status Comment
------------ ------------------------ ------------------------
m-lrkdb1 passed exists
m-lrkdb3 passed exists
Result: Group existence check passed for "dba"

Check: Membership of user "grid" in group "oinstall" [as Primary]
Node Name User Exists Group Exists User in Group Primary Status
---------------- ------------ ------------ ------------ ------------ ------------
m-lrkdb1 yes yes yes yes passed
m-lrkdb3 yes yes yes yes passed
Result: Membership check for user "grid" in group "oinstall" [as Primary] passed

Check: Membership of user "grid" in group "dba"
Node Name User Exists Group Exists User in Group Status
---------------- ------------ ------------ ------------ ----------------
m-lrkdb1 yes yes yes passed
m-lrkdb3 yes yes yes passed
Result: Membership check for user "grid" in group "dba" passed

Check: Run level
Node Name run level Required Status
------------ ------------------------ ------------------------ ----------
m-lrkdb1 3 3,5 passed
m-lrkdb3 3 3,5 passed
Result: Run level check passed

Check: Hard limits for "maximum open file descriptors"
Node Name Type Available Required Status
---------------- ------------ ------------ ------------ ----------------
m-lrkdb1 hard 65536 65536 passed
m-lrkdb3 hard 65536 65536 passed
Result: Hard limits check passed for "maximum open file descriptors"

Check: Soft limits for "maximum open file descriptors"
Node Name Type Available Required Status
---------------- ------------ ------------ ------------ ----------------
m-lrkdb1 soft 65536 1024 passed
m-lrkdb3 soft 65536 1024 passed
Result: Soft limits check passed for "maximum open file descriptors"

Check: Hard limits for "maximum user processes"
Node Name Type Available Required Status
---------------- ------------ ------------ ------------ ----------------
m-lrkdb1 hard 16384 16384 passed
m-lrkdb3 hard 96203 16384 passed
Result: Hard limits check passed for "maximum user processes"

Check: Soft limits for "maximum user processes"
Node Name Type Available Required Status
---------------- ------------ ------------ ------------ ----------------
m-lrkdb1 soft 2047 2047 passed
m-lrkdb3 soft 2047 2047 passed
Result: Soft limits check passed for "maximum user processes"

Check: System architecture
Node Name Available Required Status
------------ ------------------------ ------------------------ ----------
m-lrkdb1 x86_64 x86_64 passed
m-lrkdb3 x86_64 x86_64 passed
Result: System architecture check passed

Check: Kernel version
Node Name Available Required Status
------------ ------------------------ ------------------------ ----------
m-lrkdb1 2.6.18-164.el5 2.6.18 passed
m-lrkdb3 2.6.39-200.24.1.el6uek.x86_64 2.6.18 passed

WARNING:
PRVF-7524 : Kernel version is not consistent across all the nodes.
Kernel version = "2.6.18-164.el5" found on nodes: m-lrkdb1.
Kernel version = "2.6.39-200.24.1.el6uek.x86_64" found on nodes: m-lrkdb3.
Result: Kernel version check passed

Check: Kernel parameter for "semmsl"

Node Name Current Configured Required Status Comment
---------------- ------------ ------------ ------------ ------------ ------------
m-lrkdb1 250 250 250 passed
m-lrkdb3 250 250 250 passed
Result: Kernel parameter check passed for "semmsl"

Check: Kernel parameter for "semmns"
Node Name Current Configured Required Status Comment
---------------- ------------ ------------ ------------ ------------ ------------
m-lrkdb1 32000 32000 32000 passed
m-lrkdb3 32000 32000 32000 passed
Result: Kernel parameter check passed for "semmns"

Check: Kernel parameter for "semopm"
Node Name Current Configured Required Status Comment
---------------- ------------ ------------ ------------ ------------ ------------
m-lrkdb1 100 100 100 passed
m-lrkdb3 100 100 100 passed
Result: Kernel parameter check passed for "semopm"

Check: Kernel parameter for "semmni"
Node Name Current Configured Required Status Comment
---------------- ------------ ------------ ------------ ------------ ------------
m-lrkdb1 128 128 128 passed
m-lrkdb3 128 128 128 passed
Result: Kernel parameter check passed for "semmni"

Check: Kernel parameter for "shmmax"
Node Name Current Configured Required Status Comment
---------------- ------------ ------------ ------------ ------------ ------------
m-lrkdb1 68719476736 68719476736 6297552896 passed
m-lrkdb3 68719476736 68719476736 6313486336 passed
Result: Kernel parameter check passed for "shmmax"

Check: Kernel parameter for "shmmni"
Node Name Current Configured Required Status Comment
---------------- ------------ ------------ ------------ ------------ ------------
m-lrkdb1 4096 4096 4096 passed
m-lrkdb3 4096 4096 4096 passed
Result: Kernel parameter check passed for "shmmni"

Check: Kernel parameter for "shmall"
Node Name Current Configured Required Status Comment
---------------- ------------ ------------ ------------ ------------ ------------
m-lrkdb1 4294967296 4294967296 1229990 passed
m-lrkdb3 4294967296 4294967296 1233102 passed
Result: Kernel parameter check passed for "shmall"

Check: Kernel parameter for "file-max"
Node Name Current Configured Required Status Comment
---------------- ------------ ------------ ------------ ------------ ------------
m-lrkdb1 6815744 6815744 6815744 passed
m-lrkdb3 6815744 6815744 6815744 passed
Result: Kernel parameter check passed for "file-max"

Check: Kernel parameter for "ip_local_port_range"
Node Name Current Configured Required Status Comment
---------------- ------------ ------------ ------------ ------------ ------------
m-lrkdb1 between 9000 & 65500 between 9000 & 65500 between 9000 & 65535 passed
m-lrkdb3 between 9000 & 65500 between 9000 & 65500 between 9000 & 65535 passed
Result: Kernel parameter check passed for "ip_local_port_range"

Check: Kernel parameter for "rmem_default"
Node Name Current Configured Required Status Comment
---------------- ------------ ------------ ------------ ------------ ------------
m-lrkdb1 262144 262144 262144 passed
m-lrkdb3 262144 262144 262144 passed
Result: Kernel parameter check passed for "rmem_default"

Check: Kernel parameter for "rmem_max"
Node Name Current Configured Required Status Comment
---------------- ------------ ------------ ------------ ------------ ------------
m-lrkdb1 4194304 4194304 4194304 passed
m-lrkdb3 4194304 4194304 4194304 passed
Result: Kernel parameter check passed for "rmem_max"

Check: Kernel parameter for "wmem_default"
Node Name Current Configured Required Status Comment
---------------- ------------ ------------ ------------ ------------ ------------
m-lrkdb1 262144 262144 262144 passed
m-lrkdb3 262144 262144 262144 passed
Result: Kernel parameter check passed for "wmem_default"

Check: Kernel parameter for "wmem_max"
Node Name Current Configured Required Status Comment
---------------- ------------ ------------ ------------ ------------ ------------
m-lrkdb1 1048576 1048576 1048576 passed
m-lrkdb3 1048576 1048576 1048576 passed
Result: Kernel parameter check passed for "wmem_max"

Check: Kernel parameter for "aio-max-nr"
Node Name Current Configured Required Status Comment
---------------- ------------ ------------ ------------ ------------ ------------
m-lrkdb1 1048576 1048576 1048576 passed
m-lrkdb3 1048576 1048576 1048576 passed
Result: Kernel parameter check passed for "aio-max-nr"

Check: Package existence for "make"
Node Name Available Required Status
------------ ------------------------ ------------------------ ----------
m-lrkdb1 make-3.81-3.el5 make-3.81 passed
m-lrkdb3 make-3.81-20.el6 make-3.81 passed
Result: Package existence check passed for "make"

Check: Package existence for "binutils"
Node Name Available Required Status
------------ ------------------------ ------------------------ ----------
m-lrkdb1 binutils-2.17.50.0.6-14.el5 binutils-2.17.50.0.6 passed
m-lrkdb3 binutils-2.20.51.0.2-5.34.el6 binutils-2.17.50.0.6 passed
Result: Package existence check passed for "binutils"

Check: Package existence for "gcc(x86_64)"
Node Name Available Required Status
------------ ------------------------ ------------------------ ----------
m-lrkdb1 gcc(x86_64)-4.1.2-51.el5 gcc(x86_64)-4.1.2 passed
m-lrkdb3 gcc(x86_64)-4.4.6-4.el6 gcc(x86_64)-4.1.2 passed
Result: Package existence check passed for "gcc(x86_64)"

Check: Package existence for "libaio(x86_64)"
Node Name Available Required Status
------------ ------------------------ ------------------------ ----------
m-lrkdb1 libaio(x86_64)-0.3.106-5 libaio(x86_64)-0.3.106 passed
m-lrkdb3 libaio(x86_64)-0.3.107-10.el6 libaio(x86_64)-0.3.106 passed
Result: Package existence check passed for "libaio(x86_64)"

Check: Package existence for "glibc(x86_64)"
Node Name Available Required Status
------------ ------------------------ ------------------------ ----------
m-lrkdb1 glibc(x86_64)-2.5-65 glibc(x86_64)-2.5-58 passed
m-lrkdb3 glibc(x86_64)-2.12-1.132.el6_5.2 glibc(x86_64)-2.5-58 passed

Result: Package existence check passed for "glibc(x86_64)"

Check: Package existence for "compat-libstdc++-33(x86_64)"
Node Name Available Required Status
------------ ------------------------ ------------------------ ----------
m-lrkdb1 compat-libstdc++-33(x86_64)-3.2.3-61 compat-libstdc++-33(x86_64)-3.2.3 passed
m-lrkdb3 compat-libstdc++-33(x86_64)-3.2.3-69.el6 compat-libstdc++-33(x86_64)-3.2.3 passed
Result: Package existence check passed for "compat-libstdc++-33(x86_64)"

Check: Package existence for "glibc-devel(x86_64)"
Node Name Available Required Status
------------ ------------------------ ------------------------ ----------
m-lrkdb1 glibc-devel(x86_64)-2.5-65 glibc-devel(x86_64)-2.5 passed
m-lrkdb3 glibc-devel(x86_64)-2.12-1.132.el6_5.2 glibc-devel(x86_64)-2.5 passed
Result: Package existence check passed for "glibc-devel(x86_64)"

Check: Package existence for "gcc-c++(x86_64)"
Node Name Available Required Status
------------ ------------------------ ------------------------ ----------
m-lrkdb1 gcc-c++(x86_64)-4.1.2-51.el5 gcc-c++(x86_64)-4.1.2 passed
m-lrkdb3 gcc-c++(x86_64)-4.4.6-4.el6 gcc-c++(x86_64)-4.1.2 passed
Result: Package existence check passed for "gcc-c++(x86_64)"

Check: Package existence for "libaio-devel(x86_64)"
Node Name Available Required Status
------------ ------------------------ ------------------------ ----------
m-lrkdb1 libaio-devel(x86_64)-0.3.106-5 libaio-devel(x86_64)-0.3.106 passed
m-lrkdb3 libaio-devel(x86_64)-0.3.107-10.el6 libaio-devel(x86_64)-0.3.106 passed
Result: Package existence check passed for "libaio-devel(x86_64)"

Check: Package existence for "libgcc(x86_64)"
Node Name Available Required Status
------------ ------------------------ ------------------------ ----------
m-lrkdb1 libgcc(x86_64)-4.1.2-51.el5 libgcc(x86_64)-4.1.2 passed
m-lrkdb3 libgcc(x86_64)-4.4.6-4.el6 libgcc(x86_64)-4.1.2 passed
Result: Package existence check passed for "libgcc(x86_64)"

Check: Package existence for "libstdc++(x86_64)"
Node Name Available Required Status
------------ ------------------------ ------------------------ ----------
m-lrkdb1 libstdc++(x86_64)-4.1.2-51.el5 libstdc++(x86_64)-4.1.2 passed
m-lrkdb3 libstdc++(x86_64)-4.4.6-4.el6 libstdc++(x86_64)-4.1.2 passed
Result: Package existence check passed for "libstdc++(x86_64)"

Check: Package existence for "libstdc++-devel(x86_64)"
Node Name Available Required Status
------------ ------------------------ ------------------------ ----------
m-lrkdb1 libstdc++-devel(x86_64)-4.1.2-51.el5 libstdc++-devel(x86_64)-4.1.2 passed
m-lrkdb3 libstdc++-devel(x86_64)-4.4.6-4.el6 libstdc++-devel(x86_64)-4.1.2 passed
Result: Package existence check passed for "libstdc++-devel(x86_64)"

Check: Package existence for "sysstat"
Node Name Available Required Status
------------ ------------------------ ------------------------ ----------
m-lrkdb1 sysstat-7.0.2-11.el5 sysstat-7.0.2 passed
m-lrkdb3 sysstat-9.0.4-20.el6 sysstat-7.0.2 passed
Result: Package existence check passed for "sysstat"

Check: Package existence for "ksh"
Node Name Available Required Status
------------ ------------------------ ------------------------ ----------
m-lrkdb1 ksh-20100202-1.el5_6.6 ksh-... passed
m-lrkdb3 ksh-20120801-10.el6_5.6 ksh-... passed
Result: Package existence check passed for "ksh"

Check: Package existence for "nfs-utils"
Node Name Available Required Status
------------ ------------------------ ------------------------ ----------
m-lrkdb1 nfs-utils-1.0.9-66.el5 nfs-utils-1.0.9-60 passed
m-lrkdb3 nfs-utils-1.2.3-39.el6_5.3 nfs-utils-1.0.9-60 passed
Result: Package existence check passed for "nfs-utils"

Checking for multiple users with UID value 0
Result: Check for multiple users with UID value 0 passed

Check: Current group ID
Result: Current group ID check passed

Starting check for consistency of primary group of root user
Node Name Status
------------------------------------ ------------------------
m-lrkdb1 passed
m-lrkdb3 passed

Check for consistency of root user's primary group passed

Starting Clock synchronization checks using Network Time Protocol(NTP)...

Checking existence of NTP configuration file "/etc/ntp.conf" across nodes
Node Name File exists?
------------------------------------ ------------------------
m-lrkdb1 yes
m-lrkdb3 yes
The NTP configuration file "/etc/ntp.conf" is available on all nodes
NTP configuration file "/etc/ntp.conf" existence check passed

Checking daemon liveness...

Check: Liveness for "ntpd"
Node Name Running?
------------------------------------ ------------------------
m-lrkdb1 yes
m-lrkdb3 yes
Result: Liveness check passed for "ntpd"
Check for NTP daemon or service alive passed on all nodes

Checking whether NTP daemon or service is using UDP port 123 on all nodes

Check for NTP daemon or service using UDP port 123
Node Name Port Open?
------------------------------------ ------------------------
m-lrkdb1 yes
m-lrkdb3 yes

NTP common Time Server Check started...
NTP Time Server "193.67.79.202" is common to all nodes on which the NTP daemon is running
NTP Time Server "37.34.57.190" is common to all nodes on which the NTP daemon is running
Check of common NTP Time Server passed

Clock time offset check from NTP Time Server started...
Checking on nodes "[m-lrkdb1, m-lrkdb3]"...
Check: Clock time offset from NTP Time Server

Time Server: 193.67.79.202
Time Offset Limit: 1000.0 msecs
Node Name Time Offset Status
------------ ------------------------ ------------------------
m-lrkdb1 -0.518 passed
m-lrkdb3 -0.234 passed
Time Server "193.67.79.202" has time offsets that are within permissible limits for nodes "[m-lrkdb1, m-lrkdb3]".

Time Server: 37.34.57.190
Time Offset Limit: 1000.0 msecs
Node Name Time Offset Status
------------ ------------------------ ------------------------
m-lrkdb1 0.614 passed
m-lrkdb3 0.302 passed
Time Server "37.34.57.190" has time offsets that are within permissible limits for nodes "[m-lrkdb1, m-lrkdb3]".
Clock time offset check passed

Result: Clock synchronization check using Network Time Protocol(NTP) passed

Checking Core file name pattern consistency...
Core file name pattern consistency check passed.

Checking to make sure user "grid" is not in "root" group
Node Name Status Comment
------------ ------------------------ ------------------------
m-lrkdb1 passed does not exist
m-lrkdb3 passed does not exist
Result: User "grid" is not part of "root" group. Check passed

Check default user file creation mask
Node Name Available Required Comment
------------ ------------------------ ------------------------ ----------
m-lrkdb1 0022 0022 passed
m-lrkdb3 0022 0022 passed
Result: Default user file creation mask check passed
Checking integrity of file "/etc/resolv.conf" across nodes

Checking the file "/etc/resolv.conf" to make sure only one of domain and search entries is defined
"domain" and "search" entries do not coexist in any "/etc/resolv.conf" file
Checking if domain entry in file "/etc/resolv.conf" is consistent across the nodes...
"domain" entry does not exist in any "/etc/resolv.conf" file
Checking if search entry in file "/etc/resolv.conf" is consistent across the nodes...
Checking file "/etc/resolv.conf" to make sure that only one search entry is defined
More than one "search" entry does not exist in any "/etc/resolv.conf" file
All nodes have same "search" order defined in file "/etc/resolv.conf"
Checking DNS response time for an unreachable node
Node Name Status
------------------------------------ ------------------------
m-lrkdb1 passed
m-lrkdb3 passed
The DNS response time for an unreachable node is within acceptable limit on all nodes

Check for integrity of file "/etc/resolv.conf" passed

Check: Time zone consistency
Result: Time zone consistency check passed

Checking integrity of name service switch configuration file "/etc/nsswitch.conf" ...
Checking if "hosts" entry in file "/etc/nsswitch.conf" is consistent across nodes...
Checking file "/etc/nsswitch.conf" to make sure that only one "hosts" entry is defined
More than one "hosts" entry does not exist in any "/etc/nsswitch.conf" file
All nodes have same "hosts" entry defined in file "/etc/nsswitch.conf"
Check for integrity of name service switch configuration file "/etc/nsswitch.conf" passed


Checking daemon "avahi-daemon" is not configured and running

Check: Daemon "avahi-daemon" not configured
Node Name Configured Status
------------ ------------------------ ------------------------
m-lrkdb1 no passed
m-lrkdb3 no passed
Daemon not configured check passed for process "avahi-daemon"

Check: Daemon "avahi-daemon" not running
Node Name Running? Status
------------ ------------------------ ------------------------
m-lrkdb1 no passed
m-lrkdb3 no passed
Daemon not running check passed for process "avahi-daemon"

Starting check for /dev/shm mounted as temporary file system ...

Check for /dev/shm mounted as temporary file system passed

Starting check for /boot mount ...

Check for /boot mount passed

Starting check for zeroconf check ...

Check for zeroconf check passed


NOTE:
No fixable verification failures to fix

Pre-check for cluster services setup was unsuccessful on all the nodes.


13 Appendix 3: Output: addNode.sh



[grid@m-lrkdb2:bin]$ ./addNode.sh -silent "CLUSTER_NEW_NODES={m-lrkdb3}" "CLUSTER_NEW_VIRTUAL_HOSTNAMES={m-lrkdb3-vip}"
Starting Oracle Universal Installer...

Checking swap space: must be greater than 500 MB. Actual 9820 MB Passed
Oracle Universal Installer, Version 11.2.0.3.0 Production
Copyright (C) 1999, 2011, Oracle. All rights reserved.


Performing tests to see whether nodes m-lrkdb1,m-lrkdb3 are available
............................................................... 100% Done.

.
-----------------------------------------------------------------------------
Cluster Node Addition Summary
Global Settings
Source: /u01/app/11.2.0/grid_3
New Nodes
Space Requirements
New Nodes
m-lrkdb3
/u01: Required 17.54GB : Available 36.32GB
Installed Products
Product Names
Oracle Grid Infrastructure 11.2.0.3.0
Sun JDK 1.5.0.30.03
Installer SDK Component 11.2.0.3.0
Oracle One-Off Patch Installer 11.2.0.1.7
Oracle Universal Installer 11.2.0.3.0
Oracle USM Deconfiguration 11.2.0.3.0
Oracle Configuration Manager Deconfiguration 10.3.1.0.0
Enterprise Manager Common Core Files 10.2.0.4.4
Oracle DBCA Deconfiguration 11.2.0.3.0
Oracle RAC Deconfiguration 11.2.0.3.0
Oracle Quality of Service Management (Server) 11.2.0.3.0
Installation Plugin Files 11.2.0.3.0
Universal Storage Manager Files 11.2.0.3.0
Oracle Text Required Support Files 11.2.0.3.0
Automatic Storage Management Assistant 11.2.0.3.0
Oracle Database 11g Multimedia Files 11.2.0.3.0
Oracle Multimedia Java Advanced Imaging 11.2.0.3.0
Oracle Globalization Support 11.2.0.3.0
Oracle Multimedia Locator RDBMS Files 11.2.0.3.0
Oracle Core Required Support Files 11.2.0.3.0
Bali Share 1.1.18.0.0
Oracle Database Deconfiguration 11.2.0.3.0
Oracle Quality of Service Management (Client) 11.2.0.3.0
Expat libraries 2.0.1.0.1
Oracle Containers for Java 11.2.0.3.0
Perl Modules 5.10.0.0.1
Secure Socket Layer 11.2.0.3.0
Oracle JDBC/OCI Instant Client 11.2.0.3.0
Oracle Multimedia Client Option 11.2.0.3.0
LDAP Required Support Files 11.2.0.3.0
Character Set Migration Utility 11.2.0.3.0
Perl Interpreter 5.10.0.0.2
PL/SQL Embedded Gateway 11.2.0.3.0
OLAP SQL Scripts 11.2.0.3.0
Database SQL Scripts 11.2.0.3.0
Oracle Extended Windowing Toolkit 3.4.47.0.0
SSL Required Support Files for InstantClient 11.2.0.3.0
SQL*Plus Files for Instant Client 11.2.0.3.0
Oracle Net Required Support Files 11.2.0.3.0
Oracle Database User Interface 2.2.13.0.0
RDBMS Required Support Files for Instant Client 11.2.0.3.0
RDBMS Required Support Files Runtime 11.2.0.3.0
XML Parser for Java 11.2.0.3.0
Oracle Security Developer Tools 11.2.0.3.0
Oracle Wallet Manager 11.2.0.3.0
Enterprise Manager plugin Common Files 11.2.0.3.0
Platform Required Support Files 11.2.0.3.0
Oracle JFC Extended Windowing Toolkit 4.2.36.0.0
RDBMS Required Support Files 11.2.0.3.0
Oracle Ice Browser 5.2.3.6.0
Oracle Help For Java 4.2.9.0.0
Enterprise Manager Common Files 10.2.0.4.3
Deinstallation Tool 11.2.0.3.0
Oracle Java Client 11.2.0.3.0
Cluster Verification Utility Files 11.2.0.3.0
Oracle Notification Service (eONS) 11.2.0.3.0
Oracle LDAP administration 11.2.0.3.0
Cluster Verification Utility Common Files 11.2.0.3.0
Oracle Clusterware RDBMS Files 11.2.0.3.0
Oracle Locale Builder 11.2.0.3.0
Oracle Globalization Support 11.2.0.3.0
Buildtools Common Files 11.2.0.3.0
Oracle RAC Required Support Files-HAS 11.2.0.3.0
SQL*Plus Required Support Files 11.2.0.3.0
XDK Required Support Files 11.2.0.3.0
Agent Required Support Files 10.2.0.4.3
Parser Generator Required Support Files 11.2.0.3.0
Precompiler Required Support Files 11.2.0.3.0
Installation Common Files 11.2.0.3.0
Required Support Files 11.2.0.3.0
Oracle JDBC/THIN Interfaces 11.2.0.3.0
Oracle Multimedia Locator 11.2.0.3.0
Oracle Multimedia 11.2.0.3.0
HAS Common Files 11.2.0.3.0
Assistant Common Files 11.2.0.3.0
PL/SQL 11.2.0.3.0
HAS Files for DB 11.2.0.3.0
Oracle Recovery Manager 11.2.0.3.0
Oracle Database Utilities 11.2.0.3.0
Oracle Notification Service 11.2.0.3.0
SQL*Plus 11.2.0.3.0
Oracle Netca Client 11.2.0.3.0
Oracle Net 11.2.0.3.0
Oracle JVM 11.2.0.3.0
Oracle Internet Directory Client 11.2.0.3.0
Oracle Net Listener 11.2.0.3.0
Cluster Ready Services Files 11.2.0.3.0
Oracle Database 11g 11.2.0.3.0
-----------------------------------------------------------------------------

Instantiating scripts for add node (Tuesday, July 22, 2014 12:55:07 PM CEST)
. 1% Done.
Instantiation of add node scripts complete

Copying to remote nodes (Tuesday, July 22, 2014 12:55:11 PM CEST)
............................................................................................... 96% Done.
Home copied to new nodes

Saving inventory on nodes (Tuesday, July 22, 2014 1:33:55 PM CEST)
. 100% Done.
Save inventory complete
WARNING:A new inventory has been created on one or more nodes in this session. However, it has not yet been registered as the central inventory of this system.
To register the new inventory please run the script at '/u01/app/oraInventory/orainstRoot.sh' with root privileges on nodes 'm-lrkdb3'.
If you do not register the inventory, you may not be able to update or patch the products you installed.
The following configuration scripts need to be executed as the "root" user in each new cluster node. Each script in the list below is followed by a list of nodes.
/u01/app/oraInventory/orainstRoot.sh #On nodes m-lrkdb3
/u01/app/11.2.0/grid_3/root.sh #On nodes m-lrkdb3
To execute the configuration scripts:
1. Open a terminal window
2. Log in as "root"
3. Run the scripts in each cluster node

The Cluster Node Addition of /u01/app/11.2.0/grid_3 was successful.

14 Appendix 4: Output: cluvfy stage -post nodeadd



[grid@m-lrkdb2:bin]$ ./cluvfy stage -post nodeadd -n m-lrkdb3 -verbose

Performing post-checks for node addition

Checking node reachability...

Check: Node reachability from node "m-lrkdb2"
Destination Node Reachable?
------------------------------------ ------------------------
m-lrkdb3 yes
Result: Node reachability check passed from node "m-lrkdb2"


Checking user equivalence...

Check: User equivalence for user "grid"
Node Name Status
------------------------------------ ------------------------
m-lrkdb3 passed
Result: User equivalence check passed for user "grid"

Checking node connectivity...

Checking hosts config file...
Node Name Status
------------------------------------ ------------------------
m-lrkdb2 passed
m-lrkdb1 passed
m-lrkdb3 passed

Verification of the hosts config file successful


Interface information for node "m-lrkdb2"
Name IP Address Subnet Gateway Def. Gateway HW Address MTU
------ --------------- --------------- --------------- --------------- ----------------- ------
eth0 10.19.62.65 10.19.62.0 0.0.0.0 10.19.62.254 00:50:56:A6:00:3D 1500
eth0 10.19.62.51 10.19.62.0 0.0.0.0 10.19.62.254 00:50:56:A6:00:3D 1500
eth0 10.19.62.68 10.19.62.0 0.0.0.0 10.19.62.254 00:50:56:A6:00:3D 1500
eth1 192.168.1.134 192.168.1.0 0.0.0.0 10.19.62.254 00:50:56:A6:00:40 1500
eth1 169.254.177.108 169.254.0.0 0.0.0.0 10.19.62.254 00:50:56:A6:00:40 1500
eth2 192.168.0.134 192.168.0.0 0.0.0.0 10.19.62.254 00:50:56:A6:00:41 1500


Interface information for node "m-lrkdb1"
Name IP Address Subnet Gateway Def. Gateway HW Address MTU
------ --------------- --------------- --------------- --------------- ----------------- ------
eth0 10.19.62.64 10.19.62.0 0.0.0.0 10.19.62.254 00:50:56:A6:00:3C 1500
eth0 10.19.62.67 10.19.62.0 0.0.0.0 10.19.62.254 00:50:56:A6:00:3C 1500
eth0 10.19.62.49 10.19.62.0 0.0.0.0 10.19.62.254 00:50:56:A6:00:3C 1500
eth1 192.168.1.133 192.168.1.0 0.0.0.0 10.19.62.254 00:50:56:A6:00:3E 1500
eth1 169.254.162.124 169.254.0.0 0.0.0.0 10.19.62.254 00:50:56:A6:00:3E 1500
eth2 192.168.0.133 192.168.0.0 0.0.0.0 10.19.62.254 00:50:56:A6:00:3F 1500


Interface information for node "m-lrkdb3"
Name IP Address Subnet Gateway Def. Gateway HW Address MTU
------ --------------- --------------- --------------- --------------- ----------------- ------
eth0 10.19.62.53 10.19.62.0 0.0.0.0 10.19.62.254 00:50:56:8F:44:FC 1500
eth0 10.19.62.86 10.19.62.0 0.0.0.0 10.19.62.254 00:50:56:8F:44:FC 1500
eth0 10.19.62.66 10.19.62.0 0.0.0.0 10.19.62.254 00:50:56:8F:44:FC 1500
eth1 192.168.1.135 192.168.1.0 0.0.0.0 10.19.62.254 00:50:56:8F:40:F8 1500
eth1 169.254.55.52 169.254.0.0 0.0.0.0 10.19.62.254 00:50:56:8F:40:F8 1500
eth1 169.254.134.165 169.254.0.0 0.0.0.0 10.19.62.254 00:50:56:8F:40:F8 1500
eth2 192.168.0.135 192.168.0.0 0.0.0.0 10.19.62.254 00:50:56:8F:F3:9A 1500


Check: Node connectivity for interface "eth0"
Source Destination Connected?
------------------------------ ------------------------------ ----------------
m-lrkdb2[10.19.62.65] m-lrkdb2[10.19.62.51] yes
m-lrkdb2[10.19.62.65] m-lrkdb2[10.19.62.68] yes
m-lrkdb2[10.19.62.65] m-lrkdb1[10.19.62.64] yes
m-lrkdb2[10.19.62.65] m-lrkdb1[10.19.62.67] yes
m-lrkdb2[10.19.62.65] m-lrkdb1[10.19.62.49] yes
m-lrkdb2[10.19.62.65] m-lrkdb3[10.19.62.53] yes
m-lrkdb2[10.19.62.65] m-lrkdb3[10.19.62.86] yes
m-lrkdb2[10.19.62.65] m-lrkdb3[10.19.62.66] yes
m-lrkdb2[10.19.62.51] m-lrkdb2[10.19.62.68] yes
m-lrkdb2[10.19.62.51] m-lrkdb1[10.19.62.64] yes
m-lrkdb2[10.19.62.51] m-lrkdb1[10.19.62.67] yes
m-lrkdb2[10.19.62.51] m-lrkdb1[10.19.62.49] yes
m-lrkdb2[10.19.62.51] m-lrkdb3[10.19.62.53] yes
m-lrkdb2[10.19.62.51] m-lrkdb3[10.19.62.86] yes
m-lrkdb2[10.19.62.51] m-lrkdb3[10.19.62.66] yes
m-lrkdb2[10.19.62.68] m-lrkdb1[10.19.62.64] yes
m-lrkdb2[10.19.62.68] m-lrkdb1[10.19.62.67] yes
m-lrkdb2[10.19.62.68] m-lrkdb1[10.19.62.49] yes
m-lrkdb2[10.19.62.68] m-lrkdb3[10.19.62.53] yes
m-lrkdb2[10.19.62.68] m-lrkdb3[10.19.62.86] yes
m-lrkdb2[10.19.62.68] m-lrkdb3[10.19.62.66] yes
m-lrkdb1[10.19.62.64] m-lrkdb1[10.19.62.67] yes
m-lrkdb1[10.19.62.64] m-lrkdb1[10.19.62.49] yes
m-lrkdb1[10.19.62.64] m-lrkdb3[10.19.62.53] yes
m-lrkdb1[10.19.62.64] m-lrkdb3[10.19.62.86] yes
m-lrkdb1[10.19.62.64] m-lrkdb3[10.19.62.66] yes
m-lrkdb1[10.19.62.67] m-lrkdb1[10.19.62.49] yes
m-lrkdb1[10.19.62.67] m-lrkdb3[10.19.62.53] yes
m-lrkdb1[10.19.62.67] m-lrkdb3[10.19.62.86] yes
m-lrkdb1[10.19.62.67] m-lrkdb3[10.19.62.66] yes
m-lrkdb1[10.19.62.49] m-lrkdb3[10.19.62.53] yes
m-lrkdb1[10.19.62.49] m-lrkdb3[10.19.62.86] yes
m-lrkdb1[10.19.62.49] m-lrkdb3[10.19.62.66] yes
m-lrkdb3[10.19.62.53] m-lrkdb3[10.19.62.86] yes
m-lrkdb3[10.19.62.53] m-lrkdb3[10.19.62.66] yes
m-lrkdb3[10.19.62.86] m-lrkdb3[10.19.62.66] yes
Result: Node connectivity passed for interface "eth0"


Check: TCP connectivity of subnet "10.19.62.0"
Source Destination Connected?
------------------------------ ------------------------------ ----------------
m-lrkdb2:10.19.62.65 m-lrkdb2:10.19.62.51 passed
m-lrkdb2:10.19.62.65 m-lrkdb2:10.19.62.68 passed
m-lrkdb2:10.19.62.65 m-lrkdb1:10.19.62.64 passed
m-lrkdb2:10.19.62.65 m-lrkdb1:10.19.62.67 passed
m-lrkdb2:10.19.62.65 m-lrkdb1:10.19.62.49 passed
m-lrkdb2:10.19.62.65 m-lrkdb3:10.19.62.53 passed
m-lrkdb2:10.19.62.65 m-lrkdb3:10.19.62.86 passed
m-lrkdb2:10.19.62.65 m-lrkdb3:10.19.62.66 passed
Result: TCP connectivity check passed for subnet "10.19.62.0"

Checking subnet mask consistency...
Subnet mask consistency check passed for subnet "10.19.62.0".
Subnet mask consistency check passed.

Result: Node connectivity check passed

Checking multicast communication...

Checking subnet "10.19.62.0" for multicast communication with multicast group "230.0.1.0"...
Check of subnet "10.19.62.0" for multicast communication with multicast group "230.0.1.0" passed.

Check of multicast communication passed.

Checking cluster integrity...

Node Name
------------------------------------
m-lrkdb2
m-lrkdb1
m-lrkdb3

Cluster integrity check passed


Checking CRS integrity...

Clusterware version consistency passed
The Oracle Clusterware is healthy on node "m-lrkdb2"
The Oracle Clusterware is healthy on node "m-lrkdb1"
The Oracle Clusterware is healthy on node "m-lrkdb3"

CRS integrity check passed

Checking shared resources...

Checking CRS home location...
"/u01/app/11.2.0/grid_3" is not shared
Result: Shared resources check for node addition passed


Checking node connectivity...

Checking hosts config file...
Node Name Status
------------------------------------ ------------------------
m-lrkdb2 passed
m-lrkdb1 passed
m-lrkdb3 passed

Verification of the hosts config file successful


Interface information for node "m-lrkdb2"
Name IP Address Subnet Gateway Def. Gateway HW Address MTU
------ --------------- --------------- --------------- --------------- ----------------- ------
eth0 10.19.62.65 10.19.62.0 0.0.0.0 10.19.62.254 00:50:56:A6:00:3D 1500
eth0 10.19.62.51 10.19.62.0 0.0.0.0 10.19.62.254 00:50:56:A6:00:3D 1500
eth0 10.19.62.68 10.19.62.0 0.0.0.0 10.19.62.254 00:50:56:A6:00:3D 1500
eth1 192.168.1.134 192.168.1.0 0.0.0.0 10.19.62.254 00:50:56:A6:00:40 1500
eth1 169.254.177.108 169.254.0.0 0.0.0.0 10.19.62.254 00:50:56:A6:00:40 1500
eth2 192.168.0.134 192.168.0.0 0.0.0.0 10.19.62.254 00:50:56:A6:00:41 1500


Interface information for node "m-lrkdb1"
Name IP Address Subnet Gateway Def. Gateway HW Address MTU
------ --------------- --------------- --------------- --------------- ----------------- ------
eth0 10.19.62.64 10.19.62.0 0.0.0.0 10.19.62.254 00:50:56:A6:00:3C 1500
eth0 10.19.62.67 10.19.62.0 0.0.0.0 10.19.62.254 00:50:56:A6:00:3C 1500
eth0 10.19.62.49 10.19.62.0 0.0.0.0 10.19.62.254 00:50:56:A6:00:3C 1500
eth1 192.168.1.133 192.168.1.0 0.0.0.0 10.19.62.254 00:50:56:A6:00:3E 1500
eth1 169.254.162.124 169.254.0.0 0.0.0.0 10.19.62.254 00:50:56:A6:00:3E 1500
eth2 192.168.0.133 192.168.0.0 0.0.0.0 10.19.62.254 00:50:56:A6:00:3F 1500


Interface information for node "m-lrkdb3"
Name IP Address Subnet Gateway Def. Gateway HW Address MTU
------ --------------- --------------- --------------- --------------- ----------------- ------
eth0 10.19.62.53 10.19.62.0 0.0.0.0 10.19.62.254 00:50:56:8F:44:FC 1500
eth0 10.19.62.86 10.19.62.0 0.0.0.0 10.19.62.254 00:50:56:8F:44:FC 1500
eth0 10.19.62.66 10.19.62.0 0.0.0.0 10.19.62.254 00:50:56:8F:44:FC 1500
eth1 192.168.1.135 192.168.1.0 0.0.0.0 10.19.62.254 00:50:56:8F:40:F8 1500
eth1 169.254.55.52 169.254.0.0 0.0.0.0 10.19.62.254 00:50:56:8F:40:F8 1500
eth1 169.254.134.165 169.254.0.0 0.0.0.0 10.19.62.254 00:50:56:8F:40:F8 1500
eth2 192.168.0.135 192.168.0.0 0.0.0.0 10.19.62.254 00:50:56:8F:F3:9A 1500


Check: Node connectivity for interface "eth1"
Source Destination Connected?
------------------------------ ------------------------------ ----------------
m-lrkdb2[192.168.1.134] m-lrkdb1[192.168.1.133] yes
m-lrkdb2[192.168.1.134] m-lrkdb3[192.168.1.135] yes
m-lrkdb1[192.168.1.133] m-lrkdb3[192.168.1.135] yes
Result: Node connectivity passed for interface "eth1"


Check: TCP connectivity of subnet "192.168.1.0"
Source Destination Connected?
------------------------------ ------------------------------ ----------------
m-lrkdb2:192.168.1.134 m-lrkdb1:192.168.1.133 passed
m-lrkdb2:192.168.1.134 m-lrkdb3:192.168.1.135 passed
Result: TCP connectivity check passed for subnet "192.168.1.0"


Check: Node connectivity for interface "eth0"
Source Destination Connected?
------------------------------ ------------------------------ ----------------
m-lrkdb2[10.19.62.65] m-lrkdb2[10.19.62.51] yes
m-lrkdb2[10.19.62.65] m-lrkdb2[10.19.62.68] yes
m-lrkdb2[10.19.62.65] m-lrkdb1[10.19.62.64] yes
m-lrkdb2[10.19.62.65] m-lrkdb1[10.19.62.67] yes
m-lrkdb2[10.19.62.65] m-lrkdb1[10.19.62.49] yes
m-lrkdb2[10.19.62.65] m-lrkdb3[10.19.62.53] yes
m-lrkdb2[10.19.62.65] m-lrkdb3[10.19.62.86] yes
m-lrkdb2[10.19.62.65] m-lrkdb3[10.19.62.66] yes
m-lrkdb2[10.19.62.51] m-lrkdb2[10.19.62.68] yes
m-lrkdb2[10.19.62.51] m-lrkdb1[10.19.62.64] yes
m-lrkdb2[10.19.62.51] m-lrkdb1[10.19.62.67] yes
m-lrkdb2[10.19.62.51] m-lrkdb1[10.19.62.49] yes
m-lrkdb2[10.19.62.51] m-lrkdb3[10.19.62.53] yes
m-lrkdb2[10.19.62.51] m-lrkdb3[10.19.62.86] yes
m-lrkdb2[10.19.62.51] m-lrkdb3[10.19.62.66] yes
m-lrkdb2[10.19.62.68] m-lrkdb1[10.19.62.64] yes
m-lrkdb2[10.19.62.68] m-lrkdb1[10.19.62.67] yes
m-lrkdb2[10.19.62.68] m-lrkdb1[10.19.62.49] yes
m-lrkdb2[10.19.62.68] m-lrkdb3[10.19.62.53] yes
m-lrkdb2[10.19.62.68] m-lrkdb3[10.19.62.86] yes
m-lrkdb2[10.19.62.68] m-lrkdb3[10.19.62.66] yes
m-lrkdb1[10.19.62.64] m-lrkdb1[10.19.62.67] yes
m-lrkdb1[10.19.62.64] m-lrkdb1[10.19.62.49] yes
m-lrkdb1[10.19.62.64] m-lrkdb3[10.19.62.53] yes
m-lrkdb1[10.19.62.64] m-lrkdb3[10.19.62.86] yes
m-lrkdb1[10.19.62.64] m-lrkdb3[10.19.62.66] yes
m-lrkdb1[10.19.62.67] m-lrkdb1[10.19.62.49] yes
m-lrkdb1[10.19.62.67] m-lrkdb3[10.19.62.53] yes
m-lrkdb1[10.19.62.67] m-lrkdb3[10.19.62.86] yes
m-lrkdb1[10.19.62.67] m-lrkdb3[10.19.62.66] yes
m-lrkdb1[10.19.62.49] m-lrkdb3[10.19.62.53] yes
m-lrkdb1[10.19.62.49] m-lrkdb3[10.19.62.86] yes
m-lrkdb1[10.19.62.49] m-lrkdb3[10.19.62.66] yes
m-lrkdb3[10.19.62.53] m-lrkdb3[10.19.62.86] yes
m-lrkdb3[10.19.62.53] m-lrkdb3[10.19.62.66] yes
m-lrkdb3[10.19.62.86] m-lrkdb3[10.19.62.66] yes
Result: Node connectivity passed for interface "eth0"


Check: TCP connectivity of subnet "10.19.62.0"
Source Destination Connected?
------------------------------ ------------------------------ ----------------
m-lrkdb2:10.19.62.65 m-lrkdb2:10.19.62.51 passed
m-lrkdb2:10.19.62.65 m-lrkdb2:10.19.62.68 passed
m-lrkdb2:10.19.62.65 m-lrkdb1:10.19.62.64 passed
m-lrkdb2:10.19.62.65 m-lrkdb1:10.19.62.67 passed
m-lrkdb2:10.19.62.65 m-lrkdb1:10.19.62.49 passed
m-lrkdb2:10.19.62.65 m-lrkdb3:10.19.62.53 passed
m-lrkdb2:10.19.62.65 m-lrkdb3:10.19.62.86 passed
m-lrkdb2:10.19.62.65 m-lrkdb3:10.19.62.66 passed
Result: TCP connectivity check passed for subnet "10.19.62.0"

Checking subnet mask consistency...
Subnet mask consistency check passed for subnet "10.19.62.0".
Subnet mask consistency check passed for subnet "192.168.1.0".
Subnet mask consistency check passed.

Result: Node connectivity check passed

Checking multicast communication...

Checking subnet "10.19.62.0" for multicast communication with multicast group "230.0.1.0"...
Check of subnet "10.19.62.0" for multicast communication with multicast group "230.0.1.0" passed.

Checking subnet "192.168.1.0" for multicast communication with multicast group "230.0.1.0"...
Check of subnet "192.168.1.0" for multicast communication with multicast group "230.0.1.0" passed.

Check of multicast communication passed.

Checking node application existence...

Checking existence of VIP node application (required)
Node Name Required Running? Comment
------------ ------------------------ ------------------------ ----------
m-lrkdb2 yes yes passed
m-lrkdb1 yes yes passed
m-lrkdb3 yes yes passed
VIP node application check passed

Checking existence of NETWORK node application (required)
Node Name Required Running? Comment
------------ ------------------------ ------------------------ ----------
m-lrkdb2 yes yes passed
m-lrkdb1 yes yes passed
m-lrkdb3 yes yes passed
NETWORK node application check passed

Checking existence of GSD node application (optional)
Node Name Required Running? Comment
------------ ------------------------ ------------------------ ----------
m-lrkdb2 no no exists
m-lrkdb1 no no exists
m-lrkdb3 no no exists
GSD node application is offline on nodes "m-lrkdb2,m-lrkdb1,m-lrkdb3"

Checking existence of ONS node application (optional)
Node Name Required Running? Comment
------------ ------------------------ ------------------------ ----------
m-lrkdb2 no yes passed
m-lrkdb1 no yes passed
m-lrkdb3 no yes passed
ONS node application check passed


Checking Single Client Access Name (SCAN)...
SCAN Name Node Running? ListenerName Port Running?
---------------- ------------ ------------ ------------ ------------ ------------
m-lrkdbacc m-lrkdb2 true LISTENER_SCAN1 1521 true
m-lrkdbacc m-lrkdb3 true LISTENER_SCAN2 1521 true
m-lrkdbacc m-lrkdb1 true LISTENER_SCAN3 1521 true

Checking TCP connectivity to SCAN Listeners...
Node ListenerName TCP connectivity?
------------ ------------------------ ------------------------
m-lrkdb2 LISTENER_SCAN1 yes
m-lrkdb2 LISTENER_SCAN2 yes
m-lrkdb2 LISTENER_SCAN3 yes
TCP connectivity to SCAN Listeners exists on all cluster nodes

Checking name resolution setup for "m-lrkdbacc"...
SCAN Name IP Address Status Comment
------------ ------------------------ ------------------------ ----------
m-lrkdbacc 10.19.62.66 passed
m-lrkdbacc 10.19.62.67 passed
m-lrkdbacc 10.19.62.68 passed

Verification of SCAN VIP and Listener setup passed

Checking to make sure user "grid" is not in "root" group
Node Name Status Comment
------------ ------------------------ ------------------------
m-lrkdb3 passed does not exist
Result: User "grid" is not part of "root" group. Check passed

Checking if Clusterware is installed on all nodes...
Check of Clusterware install passed

Checking if CTSS Resource is running on all nodes...
Check: CTSS Resource running on all nodes
Node Name Status
------------------------------------ ------------------------
m-lrkdb3 passed
Result: CTSS resource check passed


Querying CTSS for time offset on all nodes...
Result: Query of CTSS for time offset passed

Check CTSS state started...
Check: CTSS state
Node Name State
------------------------------------ ------------------------
m-lrkdb3 Observer
CTSS is in Observer state. Switching over to clock synchronization checks using NTP


Starting Clock synchronization checks using Network Time Protocol(NTP)...

NTP Configuration file check started...
The NTP configuration file "/etc/ntp.conf" is available on all nodes
NTP Configuration file check passed

Checking daemon liveness...

Check: Liveness for "ntpd"
Node Name Running?
------------------------------------ ------------------------
m-lrkdb3 yes
Result: Liveness check passed for "ntpd"
Check for NTP daemon or service alive passed on all nodes

Checking NTP daemon command line for slewing option "-x"
Check: NTP daemon command line
Node Name Slewing Option Set?
------------------------------------ ------------------------
m-lrkdb3 yes
Result:
NTP daemon slewing option check passed

Checking NTP daemon's boot time configuration, in file "/etc/sysconfig/ntpd", for slewing option "-x"

Check: NTP daemon's boot time configuration
Node Name Slewing Option Set?
------------------------------------ ------------------------
m-lrkdb3 yes
Result:
NTP daemon's boot time configuration check for slewing option passed

Checking whether NTP daemon or service is using UDP port 123 on all nodes

Check for NTP daemon or service using UDP port 123
Node Name Port Open?
------------------------------------ ------------------------
m-lrkdb3 yes

NTP common Time Server Check started...
NTP Time Server "213.136.0.252" is common to all nodes on which the NTP daemon is running
NTP Time Server "193.79.237.14" is common to all nodes on which the NTP daemon is running
NTP Time Server "37.34.57.190" is common to all nodes on which the NTP daemon is running
Check of common NTP Time Server passed

Clock time offset check from NTP Time Server started...
Checking on nodes "[m-lrkdb3]"...
Check: Clock time offset from NTP Time Server

Time Server: 213.136.0.252
Time Offset Limit: 1000.0 msecs
Node Name Time Offset Status
------------ ------------------------ ------------------------
m-lrkdb3 -0.258 passed
Time Server "213.136.0.252" has time offsets that are within permissible limits for nodes "[m-lrkdb3]".

Time Server: 193.79.237.14
Time Offset Limit: 1000.0 msecs
Node Name Time Offset Status
------------ ------------------------ ------------------------
m-lrkdb3 -0.794 passed
Time Server "193.79.237.14" has time offsets that are within permissible limits for nodes "[m-lrkdb3]".

Time Server: 37.34.57.190
Time Offset Limit: 1000.0 msecs
Node Name Time Offset Status
------------ ------------------------ ------------------------
m-lrkdb3 -250.64 passed
Time Server "37.34.57.190" has time offsets that are within permissible limits for nodes "[m-lrkdb3]".
Clock time offset check passed

Result: Clock synchronization check using Network Time Protocol(NTP) passed


Oracle Cluster Time Synchronization Services check passed

Post-check for node addition was successful. 
 
_______________________________________ END_____________________________________

Page 1, 2, 3

Geen opmerkingen:

Een reactie posten