- Linux, Virtualbox Installation
- Check OS setting
- DNS Setup including BIND, NTP, DHCP in a LAN on a separate VirtualBox VM
- Setup User Accounts
- Setup ASM disks
- Attach disk to VM
- Reboot and format disks
- Clone VirtualBox Image
- Install Clusterware Software
- Fixing a failed GRID Installation
- Install RDBMS and create RAC database
- Reference
Linux, Virtualbox Installation
Check the following link for Linux/VirtualBox installation details: http://www.oracle-base.com/articles/11g/oracle-db-11gr2-rac-installation-on-oracle-linux-6-using-virtualbox.php
- Install Virtualbox Guest Additons
- Install package : # yum install oracle-rdbms-server-11gR2-preinstall
- Update the installation: : # yum update
- Install Wireshark: # yum install wireshark # yum install wireshark-gnome
- Install ASMlib
- Install cluvfy as user grid – download here and extract files under user grid
- Extract grid software to folder grid and install rpm from folder: grid/rpm
# cd /media/sf_kits/Oracle/11.2.0.4/grid/rpm
# rpm -iv cvuqdisk-1.0.9-1.rpm
Preparing packages for installation...
Using default group oinstall to install package
cvuqdisk-1.0.9-1
- Verify the currrent OS status by running : $ ./bin/cluvfy stage -pre crsinst -n grac41
Check OS setting
Install X11 applications like xclock # yum install xorg-x11-apps Turn off and disable the firewall IPTables and disable SELinux # service iptables stop iptables: Flushing firewall rules: [ OK ] iptables: Setting chains to policy ACCEPT: filter [ OK ] iptables: Unloading modules: [ OK ] # chkconfig iptables off # chkconfig --list iptables iptables 0:off 1:off 2:off 3:off 4:off 5:off 6:off Disable SELinux. Open the config file and change the SELINUX variable from enforcing to disabled. # vim /etc/selinux/config # This file controls the state of SELinux on the system. # SELINUX= can take one of these three values: # enforcing - SELinux security policy is enforced. # permissive - SELinux prints warnings instead of enforcing. # disabled - No SELinux policy is loaded. SELINUX=disabled
DNS Setup including BIND, NTP, DHCP in a LAN on a separate VirtualBox VM
Even if you are using a DNS, Oracle recommends to list the public IP, VIP and private addresses
for each node in the hosts file on each node.
Domain: example.com Name Server: ns1.example.com 192.168.1.50
RAC Sub-Domain: grid.example.com Name Server: gns.example.com 192.168.1.55
DHCP Server: ns1.example.com
NTP Server: ns1.example.com
DHCP adresses: 192.168.1.100 ... 192.168.1.254
Configure DNS:
Identity Home Node Host Node Given Name Type Address Address Assigned By Resolved By
GNS VIP None Selected by Oracle Clusterware gns.example.com Virtual 192.168.1.55 Net administrator DNS + GNS
Node 1 Public Node 1 grac1 grac1.example.com Public 192.168.1.61 Fixed DNS
Node 1 VIP Node 1 Selected by Oracle Clusterware grac1-vip.grid.example.com Private Dynamic DHCP GNS
Node 1 Private Node 1 grac1int grac1int.example.com Private 192.168.2.71 Fixed DNS
Node 2 Public Node 2 grac2 grac2.example.com Public 192.168.1.62 Fixed DNS
Node 2 VIP Node 2 Selected by Oracle Clusterware grac2-vip.grid.example.com Private Dynamic DHCP GNS
Node 2 Private Node 2 grac2int grac2int.example.com Private 192.168.2.72 Fixed DNS
SCAN VIP 1 none Selected by Oracle Clusterware GRACE2-scan.grid.example.com Virtual Dynamic DHCP GNS
SCAN VIP 2 none Selected by Oracle Clusterware GRACE2-scan.grid.example.com Virtual Dynamic DHCP GNS
SCAN VIP 3 none Selected by Oracle Clusterware GRACE2-scan.grid.example.com Virtual Dynamic DHCP GNS
Note: the cluster node VIPs and SCANs are obtained via DHCP and if GNS is up all DHCP adresses should be found with nslookup. If you have problems with zone delegation add your GNS name server to /etc/resolv.conf
Install BIND – Make sure the following rpms are installed
Install – Make sure the following rpms are installed:
dhcp-common-4.1.1-34.P1.0.1.el6
dhcp-common-4.1.1-34.P1.0.1.el6.x86_64
bind-9.8.2-0.17.rc1.0.2.el6_4.4.x86_64.rpm
bind-libs-9.8.2-0.17.rc1.0.2.el6_4.4.x86_64.rpm
bind-utils-9.8.2-0.17.rc1.0.2.el6_4.4.x86_64.rpm
Install Bind packages
# rpm -Uvh bind-9.8.2-0.17.rc1.0.2.el6_4.4.x86_64.rpm bind-libs-9.8.2-0.17.rc1.0.2.el6_4.4.x86_64.rpm
bind-utils-9.8.2-0.17.rc1.0.2.el6_4.4.x86_64.rpm
For a detailed describtion using zone delegations check following link:
Configure DNS: -> named.conf options { listen-on port 53 { 192.168.1.50; 127.0.0.1; }; directory "/var/named"; dump-file "/var/named/data/cache_dump.db"; statistics-file "/var/named/data/named_stats.txt"; memstatistics-file "/var/named/data/named_mem_stats.txt"; allow-query { any; }; allow-recursion { any; }; recursion yes; dnssec-enable no; dnssec-validation no; }; logging { channel default_debug { file "data/named.run"; severity dynamic; }; }; zone "." IN { type hint; file "named.ca"; }; zone "1.168.192.in-addr.arpa" IN { // Reverse zone type master; file "192.168.1.db"; allow-transfer { any; }; allow-update { none; }; }; zone "2.168.192.in-addr.arpa" IN { // Reverse zone type master; file "192.168.2.db"; allow-transfer { any; }; allow-update { none; }; }; zone "example.com" IN { type master; notify no; file "example.com.db"; }; -> Forward zone: example.com.db ; ; see http://www.zytrax.com/books/dns/ch9/delegate.html ; $TTL 1H ; Time to live $ORIGIN example.com. @ IN SOA ns1.example.com. hostmaster.example.com. ( 2009011202 ; serial (todays date + todays serial #) 3H ; refresh 3 hours 1H ; retry 1 hour 1W ; expire 1 week 1D ) ; minimum 24 hour ; IN A 192.168.1.50 IN NS ns1.example.com. ; name server for example.com ns1 IN A 192.168.1.50 grac1 IN A 192.168.1.61 grac2 IN A 192.168.1.62 grac3 IN A 192.168.1.63 ; $ORIGIN grid.example.com. @ IN NS gns.grid.example.com. ; NS grid.example.com IN NS ns1.example.com. ; NS example.com gns IN A 192.168.1.55 ; glue record -> Reverse zone: 192.168.1.db $TTL 1H @ IN SOA ns1.example.com. hostmaster.example.com. ( 2009011201 ; serial (todays date + todays serial #) 3H ; refresh 3 hours 1H ; retry 1 hour 1W ; expire 1 week 1D ) ; minimum 24 hour ; NS ns1.example.com. NS gns.grid.example.com. 50 PTR ns1.example.com. 55 PTR gns.grid.example.com. ; reverse mapping for GNS 61 PTR grac1.example.com. ; reverse mapping for GNS 62 PTR grac2.example.com. ; reverse mapping for GNS 63 PTR grac3.example.com. ; reverse mapping for GNS -> Reverse zone: 192.168.2.db $TTL 1H @ IN SOA ns1.example.com. hostmaster.example.com. ( 2009011201 ; serial (todays date + todays serial #) 3H ; refresh 3 hours 1H ; retry 1 hour 1W ; expire 1 week 1D ) ; minimum 24 hour ; NS ns1.example.com. 71 PTR grac1int.example.com. ; reverse mapping for GNS 72 PTR grac2int.example.com. ; reverse mapping for GNS 73 PTR grac3int.example.com. ; reverse mapping for GNS ->/etc/resolv.conf # Generated by NetworkManager search example.com nameserver 192.168.1.50 Verify DNS ( Note: Commands where execute with a running GNS - means we already have install GRID ) Check the current GNS status # /u01/app/11203/grid/bin/srvctl config gns -a -l GNS is enabled. GNS is listening for DNS server requests on port 53 GNS is using port 5353 to connect to mDNS GNS status: OK Domain served by GNS: grid3.example.com GNS version: 11.2.0.3.0 GNS VIP network: ora.net1.network Name Type Value grac3-scan A 192.168.1.220 grac3-scan A 192.168.1.221 grac3-scan A 192.168.1.222 grac3-scan1-vip A 192.168.1.220 grac3-scan2-vip A 192.168.1.221 grac3-scan3-vip A 192.168.1.222 grac31-vip A 192.168.1.219 grac32-vip A 192.168.1.224 grac33-vip A 192.168.1.226 $ nslookup grac1.example.com Name: grac1.example.com Address: 192.168.1.61 $ nslookup grac1.example.com Name: grac1.example.com Address: 192.168.1.61 $ nslookup grac1.example.com Name: grac1.example.com Address: 192.168.1.61 $ nslookup grac1int.example.com Name: grac1int.example.com Address: 192.168.2.71 $ nslookup grac1int.example.com Name: grac1int.example.com Address: 192.168.2.71 $ nslookup grac1int.example.com Name: grac1int.example.com Address: 192.168.2.71 $ nslookup 192.168.2.71 71.2.168.192.in-addr.arpa name = grac1int.example.com. $ nslookup 192.168.2.72 72.2.168.192.in-addr.arpa name = grac2int.example.com. $ nslookup 192.168.2.73 73.2.168.192.in-addr.arpa name = grac3int.example.com. $ nslookup 192.168.1.61 61.1.168.192.in-addr.arpa name = grac1.example.com. $ nslookup 192.168.1.62 62.1.168.192.in-addr.arpa name = grac2.example.com. $ nslookup 192.168.1.63 63.1.168.192.in-addr.arpa name = grac3.example.com. $ nslookup grac1-vip.grid.example.com Non-authoritative answer: Name: grac1-vip.grid.example.com Address: 192.168.1.107 $ nslookup grac2-vip.grid.example.com Non-authoritative answer: Name: grac2-vip.grid.example.com Address: 192.168.1.112 $ nslookup GRACE2-scan.grid.example.com Non-authoritative answer: Name: GRACE2-scan.grid.example.com Address: 192.168.1.108 Name: GRACE2-scan.grid.example.com Address: 192.168.1.110 Name: GRACE2-scan.grid.example.com Address: 192.168.1.109 Use dig against DNS name server - DNS name server should use Zone Delegation $ dig @192.168.1.50 GRACE2-scan.grid.example.com ; <<>> DiG 9.8.2rc1-RedHat-9.8.2-0.10.rc1.el6 <<>> @192.168.1.50 GRACE2-scan.grid.example.com ; (1 server found) ;; global options: +cmd ;; Got answer: ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 64626 ;; flags: qr rd ra; QUERY: 1, ANSWER: 3, AUTHORITY: 2, ADDITIONAL: 1 ;; QUESTION SECTION: ;GRACE2-scan.grid.example.com. IN A ;; ANSWER SECTION: GRACE2-scan.grid.example.com. 1 IN A 192.168.1.108 GRACE2-scan.grid.example.com. 1 IN A 192.168.1.109 GRACE2-scan.grid.example.com. 1 IN A 192.168.1.110 ;; AUTHORITY SECTION: grid.example.com. 3600 IN NS ns1.example.com. grid.example.com. 3600 IN NS gns.grid.example.com. ;; ADDITIONAL SECTION: ns1.example.com. 3600 IN A 192.168.1.50 ;; Query time: 0 msec ;; SERVER: 192.168.1.50#53(192.168.1.50) ;; WHEN: Sun Jul 28 13:50:26 2013 ;; MSG SIZE rcvd: 146 Use dig against GNS name server $ dig @192.168.1.55 GRACE2-scan.grid.example.com ; <<>> DiG 9.8.2rc1-RedHat-9.8.2-0.10.rc1.el6 <<>> @192.168.1.55 GRACE2-scan.grid.example.com ; (1 server found) ;; global options: +cmd ;; Got answer: ;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 32138 ;; flags: qr aa; QUERY: 1, ANSWER: 3, AUTHORITY: 1, ADDITIONAL: 1 ;; QUESTION SECTION: ;GRACE2-scan.grid.example.com. IN A ;; ANSWER SECTION: GRACE2-scan.grid.example.com. 120 IN A 192.168.1.108 GRACE2-scan.grid.example.com. 120 IN A 192.168.1.109 GRACE2-scan.grid.example.com. 120 IN A 192.168.1.110 ;; AUTHORITY SECTION: grid.example.com. 10800 IN SOA GRACE2-gns-vip.grid.example.com. GRACE2-gns-vip.grid.example.com. 3173463 10800 10800 30 120 ;; ADDITIONAL SECTION: GRACE2-gns-vip.grid.example.com. 10800 IN A 192.168.1.55 ;; Query time: 15 msec ;; SERVER: 192.168.1.55#53(192.168.1.55) ;; WHEN: Sun Jul 28 13:50:26 2013 ;; MSG SIZE rcvd: 161
Start the DNS server
# service named restart
Starting named: [ OK ]
Ensure DNS service restart on the reboot
# chkconfig named on
# chkconfig –list named
named 0:off 1:off 2:on 3:on 4:on 5:on 6:off
Display all records for zone example.com with dig
$ dig example.com AXFR
$ dig @192.168.1.55 AXFR
$ dig GRACE2-scan.grid.example.com
Configure DHCP server
- dhclient is recreate /etc/resolv,conf . Run $ service network restart after testing dhclient that to have a consistent /etc/resolv,conf on all cluster nodes
Verify that you don't use any DHCP server from a vbriged network - Note If using Virtualbox briged network devices using same Network address as our local Router the Virtualbox DHCP server is used ( of course you can disable M:\VM> vboxmanage list bridgedifs Name: Realtek PCIe GBE Family Controller GUID: 7e0af9ff-ea37-4e63-b2e5-5128c60ab300 DHCP: Enabled IPAddress: 192.168.1.4 NetworkMask: 255.255.255.0 M:\VM\GRAC_OEL64_11203>ipconfig Windows-IP-Konfiguration Ethernet-Adapter LAN-Verbindung: Verbindungsspezifisches DNS-Suffix: speedport.ip Verbindungslokale IPv6-Adresse . : fe80::c52f:f681:bb0b:c358%11 IPv4-Adresse . . . . . . . . . . : 192.168.1.4 Subnetzmaske . . . . . . . . . . : 255.255.255.0 Standardgateway . . . . . . . . . : 192.168.1.1 Solution: Use Internal Network devices instead of Bridged Network devices for the Virtulbox Network setup -> /etc/sysconfig/dhcpd Command line options here DHCPDARGS="eth0" -> /etc/dhcp/dhcpd.conf ( don't used domain-name as this will create a new resolv.conf ) ddns-update-style interim; ignore client-updates; subnet 192.168.1.0 netmask 255.255.255.0 { option routers 192.168.1.1; # Default gateway to be used by DHCP clients option subnet-mask 255.255.255.0; # Default subnet mask to be used by DHCP clients. option ip-forwarding off; # Do not forward DHCP requests. option broadcast-address 192.168.1.255; # Default broadcast address to be used by DHCP client. # option domain-name "grid.example.com"; option domain-name-servers 192.168.1.50; # IP address of the DNS server. In this document it will be oralab1 option time-offset -19000; # Central Standard Time option ntp-servers 0.pool.ntp.org; # Default NTP server to be used by DHCP clients range 192.168.1.100 192.168.1.254; # Range of IP addresses that can be issued to DHCP client default-lease-time 21600; # Amount of time in seconds that a client may keep the IP address max-lease-time 43200; } # service dhcpd restart # chkconfig dhcpd on Test on all cluster instances: # dhclient eth0 Check /var/log/messages # tail -f /var/log/messages Jul 8 12:46:09 gns dhclient[3909]: DHCPDISCOVER on eth0 to 255.255.255.255 port 67 interval 7 (xid=0x6fb12d80) Jul 8 12:46:09 gns dhcpd: DHCPDISCOVER from 08:00:27:e6:71:54 via eth0 Jul 8 12:46:10 gns dhcpd: 0.pool.ntp.org: temporary name server failure Jul 8 12:46:10 gns dhcpd: DHCPOFFER on 192.168.1.100 to 08:00:27:e6:71:54 via eth0 Jul 8 12:46:10 gns dhclient[3909]: DHCPOFFER from 192.168.1.50 Jul 8 12:46:10 gns dhclient[3909]: DHCPREQUEST on eth0 to 255.255.255.255 port 67 (xid=0x6fb12d80) Jul 8 12:46:10 gns dhcpd: DHCPREQUEST for 192.168.1.100 (192.168.1.50) from 08:00:27:e6:71:54 via eth0 Jul 8 12:46:10 gns dhcpd: DHCPACK on 192.168.1.100 to 08:00:27:e6:71:54 via eth0 Jul 8 12:46:10 gns dhclient[3909]: DHCPACK from 192.168.1.50 (xid=0x6fb12d80) Jul 8 12:46:12 gns avahi-daemon[1407]: Registering new address record for 192.168.1.100 on eth0.IPv4. Jul 8 12:46:12 gns NET[3962]: /sbin/dhclient-script : updated /etc/resolv.conf Jul 8 12:46:12 gns dhclient[3909]: bound to 192.168.1.100 -- renewal in 9071 seconds. Jul 8 12:46:13 gns ntpd[2051]: Listening on interface #6 eth0, 192.168.1.100#123 Enabled
- Verify that the right DHCP server is in use ( at least check the bound an renwal values )
NTP Setup – Server: gns.example.com
# cat /etc/ntp.conf restrict default nomodify notrap noquery restrict 127.0.0.1 # -- CLIENT NETWORK ------- restrict 192.168.1.0 mask 255.255.255.0 nomodify notrap # --- OUR TIMESERVERS ----- can't reach NTP servers - build my own server #server 0.pool.ntp.org iburst #server 1.pool.ntp.org iburst server 127.127.1.0 # --- NTP MULTICASTCLIENT --- # --- GENERAL CONFIGURATION --- # Undisciplined Local Clock. fudge 127.127.1.0 stratum 9 # Drift file. driftfile /var/lib/ntp/drift broadcastdelay 0.008 # Keys file. keys /etc/ntp/keys # chkconfig ntpd on # ntpq -p remote refid st t when poll reach delay offset jitter ============================================================================== *LOCAL(0) .LOCL. 9 l 11 64 377 0.000 0.000 0.000 NTP Setup - Clients: grac1.example.com, grac2.example.com, ... # cat /etc/ntp.conf restrict default nomodify notrap noquery restrict 127.0.0.1 # -- CLIENT NETWORK ------- # --- OUR TIMESERVERS ----- # 192.168.1.2 is the address for my timeserver, # use the address of your own, instead: server 192.168.1.50 server 127.127.1.0 # --- NTP MULTICASTCLIENT --- # --- GENERAL CONFIGURATION --- # Undisciplined Local Clock. fudge 127.127.1.0 stratum 12 # Drift file. driftfile /var/lib/ntp/drift broadcastdelay 0.008 # Keys file. keys /etc/ntp/keys # ntpq -p remote refid st t when poll reach delay offset jitter ============================================================================== gns.example.com LOCAL(0) 10 u 22 64 1 2.065 -11.015 0.000 LOCAL(0) .LOCL. 12 l 21 64 1 0.000 0.000 0.000 Verify setup with cluvfy : Add to our /etc/rc.local # service ntpd stop ntpdate -u 192.168.1.50 service ntpd start
Verify GNS setup with cluvfy:
$ ./bin/cluvfy comp gns -precrsinst -domain grid.example.com -vip 192.168.2.100 -verbose -n grac1,grac2
Verifying GNS integrity
Checking GNS integrity...
Checking if the GNS subdomain name is valid...
The GNS subdomain name "grid.example.com" is a valid domain name
Checking if the GNS VIP is a valid address...
GNS VIP "192.168.2.100" resolves to a valid IP address
Checking the status of GNS VIP...
GNS integrity check passed
Verification of GNS integrity was successful.
Setup User Accounts
NOTE: Oracle recommend different users for the installation of the Grid Infrastructure (GI) and the Oracle RDBMS home. The GI will be installed in a separate Oracle base, owned by user ‘grid.’ After the grid install the GI home will be owned by root, and inaccessible to unauthorized users.
Create OS groups using the command below. Enter these commands as the 'root' user: #/usr/sbin/groupadd -g 501 oinstall #/usr/sbin/groupadd -g 502 dba #/usr/sbin/groupadd -g 504 asmadmin #/usr/sbin/groupadd -g 506 asmdba #/usr/sbin/groupadd -g 507 asmoper Create the users that will own the Oracle software using the commands: #/usr/sbin/useradd -u 501 -g oinstall -G asmadmin,asmdba,asmoper grid #/usr/sbin/useradd -u 502 -g oinstall -G dba,asmdba oracle $ id uid=501(grid) gid=54321(oinstall) groups=54321(oinstall),504(asmadmin),506(asmdba),507(asmoper) $ id uid=54321(oracle) gid=54321(oinstall) groups=54321(oinstall),501(vboxsf),506(adba),54322(dba) For the C shell (csh or tcsh), add the following lines to the /etc/csh.login file: if ( $USER = "oracle" || $USER = "grid" ) then limit maxproc 16384 limit descriptors 65536 endif Modify /etc/security/limits.conf # oracle-rdbms-server-11gR2-preinstall setting for nofile soft limit is 1024 oracle soft nofile 1024 grid soft nofile 1024 # oracle-rdbms-server-11gR2-preinstall setting for nofile hard limit is 65536 oracle hard nofile 65536 grid hard nofile 65536 # oracle-rdbms-server-11gR2-preinstall setting for nproc soft limit is 2047 oracle soft nproc 2047 grid soft nproc 2047 # oracle-rdbms-server-11gR2-preinstall setting for nproc hard limit is 16384 oracle hard nproc 16384 grid hard nproc 16384 # oracle-rdbms-server-11gR2-preinstall setting for stack soft limit is 10240KB oracle soft stack 10240 grid soft stack 10240 # oracle-rdbms-server-11gR2-preinstall setting for stack hard limit is 32768KB oracle hard stack 32768 grid hard stack 32768 Create Directories: - Have a separate ORACLE_BASE for both GRID and RDBMS install ! Create the Oracle Inventory Directory ( needed or 11.2.0.3 will ) To create the Oracle Inventory directory, enter the following commands as the root user: # mkdir -p /u01/app/oraInventory # chown -R grid:oinstall /u01/app/oraInventory Creating the Oracle Grid Infrastructure Home Directory To create the Grid Infrastructure home directory, enter the following commands as the root user: # mkdir -p /u01/app/grid # chown -R grid:oinstall /u01/app/grid # chmod -R 775 /u01/app/grid # mkdir -p /u01/app/11203/grid # chown -R grid:oinstall /u01//app/11203/grid # chmod -R 775 /u01/app/11203/grid Creating the Oracle Base Directory To create the Oracle Base directory, enter the following commands as the root user: # mkdir -p /u01/app/oracle # chown -R oracle:oinstall /u01/app/oracle # chmod -R 775 /u01/app/oracle Creating the Oracle RDBMS Home Directory To create the Oracle RDBMS Home directory, enter the following commands as the root user: # mkdir -p /u01/app/oracle/product/11203/racdb # chown -R oracle:oinstall /u01/app/oracle/product/11203/racdb # chmod -R 775 /u01/app/oracle/product/11203/racdb Add divider=10″ to /boot/grub/grub.conf Finally, add “divider=10″ to the boot parameters in grub.conf to improve VM performance. This is often recommended as a way to reduce host CPU utilization when a VM is idle, but it also improves overall guest performance. When I tried my first run-through of this process without this parameter enabled, the cluster configuration script bogged down terribly, and failed midway through creating the database Verify Initial Virtualbox Image using cluvfy Install the cluvfy as Grid Owner ( grid ) in ~/cluvfy112 Check the minimum system for our 1.st Virtualbox image by running: cluvfy -p crs $ ./bin/cluvfy comp sys -p crs -n grac1 Verifying system requirement Total memory check passed Available memory check passed Swap space check passed Free disk space check passed for "grac1:/u01/app/11203/grid,grac1:/tmp" Check for multiple users with UID value 501 passed User existence check passed for "grid" Group existence check passed for "oinstall" Group existence check passed for "dba" Membership check for user "grid" in group "oinstall" [as Primary] passed Membership check for user "grid" in group "dba" passed Run level check passed Hard limits check passed for "maximum open file descriptors" Hard limits check passed for "maximum user processes" System architecture check passed Kernel version check passed Kernel parameter check passed for "semmsl" Kernel parameter check passed for "semmns" Kernel parameter check passed for "semopm" Kernel parameter check passed for "semmni" Kernel parameter check passed for "shmmax" Kernel parameter check passed for "shmmni" Kernel parameter check passed for "shmall" Kernel parameter check passed for "file-max" Kernel parameter check passed for "ip_local_port_range" Kernel parameter check passed for "rmem_default" Kernel parameter check passed for "rmem_max" Kernel parameter check passed for "wmem_default" Kernel parameter check passed for "wmem_max" Kernel parameter check passed for "aio-max-nr" Package existence check passed for "binutils" Package existence check passed for "compat-libcap1" Package existence check passed for "compat-libstdc++-33(x86_64)" Package existence check passed for "libgcc(x86_64)" Package existence check passed for "libstdc++(x86_64)" Package existence check passed for "libstdc++-devel(x86_64)" Package existence check passed for "sysstat" Package existence check passed for "gcc" Package existence check passed for "gcc-c++" Package existence check passed for "ksh" Package existence check passed for "make" Package existence check passed for "glibc(x86_64)" Package existence check passed for "glibc-devel(x86_64)" Package existence check passed for "libaio(x86_64)" Package existence check passed for "libaio-devel(x86_64)" Check for multiple users with UID value 0 passed Starting check for consistency of primary group of root user Check for consistency of root user's primary group passed Time zone consistency check passed Verification of system requirement was successful.
Setup ASM disks
Create ASM disks Note : Create all ASM disks on my SSD device ( C:\VM\GRACE2\ASM ) Create 6 ASM disks : 3 disks with 5 Gbyte each 3 disks with 2 Gbyte each D:\VM>set_it D:\VM>set path="d:\Program Files\Oracle\VirtualBox";D:\Windows\system32;D:\Windo ws;D:\Windows\System32\Wbem;D:\Windows\System32\WindowsPowerShell\v1.0\;D:\Progr am Files (x86)\IDM Computer Solutions\UltraEdit\ D:\VM>VBoxManage createhd --filename C:\VM\GRACE2\ASM\asm1_5G.vdi --size 5120 -- format VDI --variant Fixed 0%...10%...20%...30%...40%...50%...60%...70%...80%...90%...100% Disk image created. UUID: 7c9711c7-14e9-4bc4-8390-3e7dbb2ad130 D:\VM>VBoxManage createhd --filename C:\VM\GRACE2\ASM\asm2_5G.vdi --size 5120 -- format VDI --variant Fixed 0%...10%...20%...30%...40%...50%...60%...70%...80%...90%...100% Disk image created. UUID: 5c801291-7083-4030-9221-cfab1460f527 D:\VM>VBoxManage createhd --filename C:\VM\GRACE2\ASM\asm3_5G.vdi --size 5120 -- format VDI --variant Fixed 0%...10%...20%...30%...40%...50%...60%...70%...80%...90%...100% Disk image created. UUID: 28b0e0b4-c9ae-474e-b339-d742a10bb120 D:\VM>VBoxManage createhd --filename C:\VM\GRACE2\ASM\asm1_2G.vdi --size 2048 -- format VDI --variant Fixed 0%...10%...20%...30%...40%...50%...60%...70%...80%...90%...100% Disk image created. UUID: acc2b925-fa58-4d5f-966f-1c9cac014d1b D:\VM>VBoxManage createhd --filename C:\VM\GRACE2\ASM\asm2_2G.vdi --size 2048 -- format VDI --variant Fixed 0%...10%...20%...30%...40%...50%...60%...70%...80%...90%...100% Disk image created. UUID: a93f5fd8-bb10-4421-af07-3dfe4fc0d740 D:\VM>VBoxManage createhd --filename C:\VM\GRACE2\ASM\asm3_2G.vdi --size 2048 -- format VDI --variant Fixed 0%...10%...20%...30%...40%...50%...60%...70%...80%...90%...100% Disk image created. UUID: 89c0f4cd-569e-4a30-9b6e-5ce3044fcde5 D:\VM>dir C:\VM\GRACE2\ASM\* Volume in Laufwerk C: hat keine Bezeichnung. Volumeseriennummer: 20BF-FC17 Verzeichnis von C:\VM\GRACE2\ASM 13.07.2013 13:00 2.147.495.936 asm1_2G.vdi 13.07.2013 12:56 5.368.733.696 asm1_5G.vdi 13.07.2013 13:00 2.147.495.936 asm2_2G.vdi 13.07.2013 12:57 5.368.733.696 asm2_5G.vdi 13.07.2013 13:00 2.147.495.936 asm3_2G.vdi 13.07.2013 12:59 5.368.733.696 asm3_5G.vdi
Attach disk to VM
D:\VM>VBoxManage storageattach grac1 --storagectl "SATA" --port 1 --device 0 --type hdd --medium C:\VM\GRACE2\ASM\asm1_5G.vdi D:\VM>VBoxManage storageattach grac1 --storagectl "SATA" --port 2 --device 0 --type hdd --medium C:\VM\GRACE2\ASM\asm2_5G.vdi D:\VM>VBoxManage storageattach grac1 --storagectl "SATA" --port 3 --device 0 --type hdd --medium C:\VM\GRACE2\ASM\asm3_5G.vdi D:\VM>VBoxManage storageattach grac1 --storagectl "SATA" --port 4 --device 0 --type hdd --medium C:\VM\GRACE2\ASM\asm1_2G.vdi D:\VM>VBoxManage storageattach grac1 --storagectl "SATA" --port 5 --device 0 --type hdd --medium C:\VM\GRACE2\ASM\asm2_2G.vdi D:\VM>VBoxManage storageattach grac1 --storagectl "SATA" --port 6 --device 0 --type hdd --medium C:\VM\GRACE2\ASM\asm3_2G.vdi Change disk type to sharable disks: D:\VM>VBoxManage modifyhd C:\VM\GRACE2\ASM\asm1_5G.vdi --type shareable D:\VM>VBoxManage modifyhd C:\VM\GRACE2\ASM\asm2_5G.vdi --type shareable D:\VM>VBoxManage modifyhd C:\VM\GRACE2\ASM\asm3_5G.vdi --type shareable D:\VM>VBoxManage modifyhd C:\VM\GRACE2\ASM\asm1_2G.vdi --type shareable D:\VM>VBoxManage modifyhd C:\VM\GRACE2\ASM\asm2_2G.vdi --type shareable D:\VM>VBoxManage modifyhd C:\VM\GRACE2\ASM\asm3_2G.vdi --type shareable
Reboot and format disks
# ls /dev/sd* /dev/sda /dev/sda2 /dev/sdb /dev/sdd /dev/sdf /dev/sda1 /dev/sda3 /dev/sdc /dev/sde /dev/sdg # fdisk /dev/sdb Command (m for help): n Command action e extended p primary partition (1-4) p Partition number (1-4): 1 First sector (2048-10485759, default 2048): Using default value 2048 Last sector, +sectors or +size{K,M,G} (2048-10485759, default 10485759): Using default value 10485759 Command (m for help): w The partition table has been altered! In each case, the sequence of answers is "n", "p", "1", "Return", "Return" and "w". Repeat steps for : /dev/sdb -> /dev/sdg # ls /dev/sd* /dev/sda /dev/sda3 /dev/sdc /dev/sdd1 /dev/sdf /dev/sdg1 /dev/sda1 /dev/sdb /dev/sdc1 /dev/sde /dev/sdf1 /dev/sda2 /dev/sdb1 /dev/sdd /dev/sde1 /dev/sdg
Configure ASMLib and Disks
# /usr/sbin/oracleasm configure -i # /etc/init.d/oracleasm createdisk data1 /dev/sdb1 Marking disk "data1" as an ASM disk: [ OK ] # /etc/init.d/oracleasm createdisk data2 /dev/sdc1 Marking disk "data2" as an ASM disk: [ OK ] # /etc/init.d/oracleasm createdisk data3 /dev/sdd1 Marking disk "data3" as an ASM disk: [ OK ] # /etc/init.d/oracleasm createdisk ocr1 /dev/sde1 Marking disk "ocr1" as an ASM disk: [ OK ] # /etc/init.d/oracleasm createdisk ocr2 /dev/sdf1 Marking disk "ocr2" as an ASM disk: [ OK ] [root@grac1 Desktop]# /etc/init.d/oracleasm createdisk ocr3 /dev/sdg1 Marking disk "ocr3" as an ASM disk: [ OK ] # /etc/init.d/oracleasm listdisks DATA1 DATA2 DATA3 OCR1 OCR2 OCR3 # ls -l /dev/oracleasm/disks total 0 brw-rw---- 1 grid asmadmin 8, 17 Jul 13 16:32 DATA1 brw-rw---- 1 grid asmadmin 8, 33 Jul 13 16:32 DATA2 brw-rw---- 1 grid asmadmin 8, 49 Jul 13 16:33 DATA3 brw-rw---- 1 grid asmadmin 8, 65 Jul 13 16:33 OCR1 brw-rw---- 1 grid asmadmin 8, 81 Jul 13 16:33 OCR2 brw-rw---- 1 grid asmadmin 8, 97 Jul 13 16:33 OCR3 # /etc/init.d/oracleasm status Checking if ASM is loaded: yes Checking if /dev/oracleasm is mounted: yes [root@grac1 Desktop]# /etc/init.d/oracleasm listdisks DATA1 DATA2 DATA3 OCR1 OCR2 OCR3 # /etc/init.d/oracleasm querydisk -d DATA1 Disk "DATA1" is a valid ASM disk on device [8, 17] # /etc/init.d/oracleasm querydisk -d DATA2 Disk "DATA2" is a valid ASM disk on device [8, 33] # /etc/init.d/oracleasm querydisk -d DATA3 Disk "DATA3" is a valid ASM disk on device [8, 49] # /etc/init.d/oracleasm querydisk -d OCR1 Disk "OCR1" is a valid ASM disk on device [8, 65] # /etc/init.d/oracleasm querydisk -d OCR2 # /etc/init.d/oracleasm querydisk -d OCR3 Disk "OCR3" is a valid ASM disk on device [8, 97] # /etc/init.d/oracleasm scandisks Scanning the system for Oracle ASMLib disks: [ OK ]
Clone VirtualBox Image
Shutdown Virtualbox image 1 and manually clone the "grac1.vdi" disk using the following commands on the host server. D:\VM> set_it D:\VM> md D:\VM\GNS_RACE2\grac2 D:\VM> VBoxManage clonehd D:\VM\GNS_RACE2\grac1\grac1.vdi d:\VM\GNS_RACE2\grac2\grac2.vdi 0%...10%...20%...30%...40%...50%...60%...70%...80%...90%...100% Clone hard disk created in format 'VDI'. UUID: 0d626e95-9354-4f65-8fc0-e40ba44e1 Manually clone the "ol6-112-rac1.vdi" disk using the following commands on the host server. Create new VM grac2 by using disk grac2.vdi Attach disk to VM: grac2 D:\VM>VBoxManage storageattach grac2 --storagectl "SATA" --port 1 --device 0 --type hdd --medium C:\VM\GRACE2\ASM\asm1_5G.vdi D:\VM>VBoxManage storageattach grac2 --storagectl "SATA" --port 2 --device 0 --type hdd --medium C:\VM\GRACE2\ASM\asm2_5G.vdi D:\VM>VBoxManage storageattach grac2 --storagectl "SATA" --port 3 --device 0 --type hdd --medium C:\VM\GRACE2\ASM\asm3_5G.vdi D:\VM>VBoxManage storageattach grac2 --storagectl "SATA" --port 4 --device 0 --type hdd --medium C:\VM\GRACE2\ASM\asm1_2G.vdi D:\VM>VBoxManage storageattach grac2 --storagectl "SATA" --port 5 --device 0 --type hdd --medium C:\VM\GRACE2\ASM\asm2_2G.vdi D:\VM>VBoxManage storageattach grac2 --storagectl "SATA" --port 6 --device 0 --type hdd --medium C:\VM\GRACE2\ASM\asm3_2G.vdi Start the "grac2" virtual machine by clicking the "Start" button on the toolbar. Ignore any network errors during the startup. Log in to the "grac2" virtual machine as the "root" user so we can reconfigure the network settings to match the following. hostname: grac2.example.com IP Address eth0: 192.168.1.62 (public address) Default Gateway eth0: 192.168.1.1 (public address) IP Address eth1: 192.168.2.72 (private address) Default Gateway eth1: none Amend the hostname in the "/etc/sysconfig/network" file. NETWORKING=yes HOSTNAME=grac2.example.com Check the MAC address of each of the available network connections. Don't worry that they are listed as "eth2" and "eth3". These are dynamically created connections because the MAC address of the "eth0" and "eth1" connections is incorrect. # ifconfig -a | grep eth eth2 Link encap:Ethernet HWaddr 08:00:27:1F:2E:33 eth3 Link encap:Ethernet HWaddr 08:00:27:8E:6D:24 Edit the "/etc/sysconfig/network-scripts/ifcfg-eth0", amending only the IPADDR and HWADDR settings as follows and deleting the UUID entry. Note, the HWADDR value comes from the "eth2" interface displayed above. IPADDR=192.168.1.62 HWADDR=08:00:27:1F:2E:33 Edit the "/etc/sysconfig/network-scripts/ifcfg-eth1", amending only the IPADDR and HWADDR settings as follows and deleting the UUID entry. Note, the HWADDR value comes from the "eth3" interface displayed above. HWADDR=08:00:27:8E:6D:24 IPADDR=192.168.2.102 Change .login for grid user setenv ORACLE_SID +ASM2 Remove udev rules: # rm /etc/udev/rules.d/70-persistent-net.rules # reboot Verify network devices ( use graphical tool if needed for changes ) # ifconfig eth0 Link encap:Ethernet HWaddr 08:00:27:1F:2E:33 inet addr:192.168.1.62 Bcast:192.168.1.255 Mask:255.255.255.0 .. eth1 Link encap:Ethernet HWaddr 08:00:27:8E:6D:24 inet addr:192.168.2.72 Bcast:192.168.2.255 Mask:255.255.255.0 .. Check Ntp $ ntpq -p remote refid st t when poll reach delay offset jitter ============================================================================== gns.example.com LOCAL(0) 10 u 30 64 1 0.462 2233.72 0.000 LOCAL(0) .LOCL. 12 l 29 64 1 0.000 0.000 0.000 Check DHCP $ grep -i dhcp /var/log/messages Jul 15 19:12:21 grac1 NetworkManager[1528]: <info> Activation (eth2) Beginning DHCPv4 transaction Jul 15 19:12:21 grac1 NetworkManager[1528]: <info> Activation (eth2) DHCPv4 will time out in 45 seconds Jul 15 19:12:21 grac1 NetworkManager[1528]: <info> Activation (eth3) Beginning DHCPv4 transaction Jul 15 19:12:21 grac1 NetworkManager[1528]: <info> Activation (eth3) DHCPv4 will time out in 45 seconds Jul 15 19:12:21 grac1 dhclient[1547]: Internet Systems Consortium DHCP Client 4.1.1-P1 Jul 15 19:12:21 grac1 dhclient[1547]: For info, please visit https://www.isc.org/software/dhcp/ Jul 15 19:12:21 grac1 dhclient[1537]: Internet Systems Consortium DHCP Client 4.1.1-P1 Jul 15 19:12:21 grac1 dhclient[1537]: For info, please visit https://www.isc.org/software/dhcp/ Jul 15 19:12:21 grac1 NetworkManager[1528]: <info> (eth2): DHCPv4 state changed nbi -> preinit Jul 15 19:12:21 grac1 NetworkManager[1528]: <info> (eth3): DHCPv4 state changed nbi -> preinit Jul 15 19:12:22 grac1 dhclient[1537]: DHCPDISCOVER on eth2 to 255.255.255.255 port 67 interval 4 (xid=0x5ddfdccc) Jul 15 19:12:23 grac1 dhclient[1547]: DHCPDISCOVER on eth3 to 255.255.255.255 port 67 interval 5 (xid=0x5c751799) Jul 15 19:12:26 grac1 dhclient[1537]: DHCPDISCOVER on eth2 to 255.255.255.255 port 67 interval 11 (xid=0x5ddfdccc) Jul 15 19:12:28 grac1 dhclient[1547]: DHCPDISCOVER on eth3 to 255.255.255.255 port 67 interval 11 (xid=0x5c751799) Jul 15 19:12:32 grac1 dhclient[1537]: DHCPOFFER from 192.168.1.50 Jul 15 19:12:32 grac1 dhclient[1537]: DHCPREQUEST on eth2 to 255.255.255.255 port 67 (xid=0x5ddfdccc) Jul 15 19:12:32 grac1 dhclient[1537]: DHCPACK from 192.168.1.50 (xid=0x5ddfdccc) Jul 15 19:12:32 grac1 NetworkManager[1528]: <info> (eth2): DHCPv4 state changed preinit -> bound Jul 15 19:12:33 grac1 dhclient[1547]: DHCPOFFER from 192.168.1.50 Jul 15 19:12:33 grac1 dhclient[1547]: DHCPREQUEST on eth3 to 255.255.255.255 port 67 (xid=0x5c751799) Jul 15 19:12:33 grac1 dhclient[1547]: DHCPACK from 192.168.1.50 (xid=0x5c751799) Jul 15 19:12:33 grac1 NetworkManager[1528]: <info> (eth3): DHCPv4 state changed preinit -> bound Jul 15 19:27:53 grac2 NetworkManager[1617]: <info> Activation (eth2) Beginning DHCPv4 transaction Jul 15 19:27:53 grac2 NetworkManager[1617]: <info> Activation (eth2) DHCPv4 will time out in 45 seconds Jul 15 19:27:53 grac2 dhclient[1637]: Internet Systems Consortium DHCP Client 4.1.1-P1 Jul 15 19:27:53 grac2 dhclient[1637]: For info, please visit https://www.isc.org/software/dhcp/ Jul 15 19:27:53 grac2 dhclient[1637]: DHCPDISCOVER on eth2 to 255.255.255.255 port 67 interval 4 (xid=0x44e12e9) Jul 15 19:27:53 grac2 NetworkManager[1617]: <info> (eth2): DHCPv4 state changed nbi -> preinit Jul 15 19:27:57 grac2 dhclient[1637]: DHCPDISCOVER on eth2 to 255.255.255.255 port 67 interval 10 (xid=0x44e12e9) Jul 15 19:28:03 grac2 dhclient[1637]: DHCPOFFER from 192.168.1.50 Jul 15 19:28:03 grac2 dhclient[1637]: DHCPREQUEST on eth2 to 255.255.255.255 port 67 (xid=0x44e12e9) Jul 15 19:28:03 grac2 dhclient[1637]: DHCPACK from 192.168.1.50 (xid=0x44e12e9) Jul 15 19:28:03 grac2 NetworkManager[1617]: <info> (eth2): DHCPv4 state changed preinit -> bound Jul 15 19:32:52 grac2 NetworkManager[1690]: <info> Activation (eth0) Beginning DHCPv4 transaction Jul 15 19:32:52 grac2 NetworkManager[1690]: <info> Activation (eth0) DHCPv4 will time out in 45 seconds Jul 15 19:32:52 grac2 dhclient[1703]: Internet Systems Consortium DHCP Client 4.1.1-P1 Jul 15 19:32:52 grac2 dhclient[1703]: For info, please visit https://www.isc.org/software/dhcp/ Jul 15 19:32:52 grac2 dhclient[1703]: DHCPDISCOVER on eth0 to 255.255.255.255 port 67 interval 6 (xid=0x6781ea4f) Jul 15 19:32:52 grac2 NetworkManager[1690]: <info> (eth0): DHCPv4 state changed nbi -> preinit Jul 15 19:32:58 grac2 dhclient[1703]: DHCPDISCOVER on eth0 to 255.255.255.255 port 67 interval 12 (xid=0x6781ea4f) Jul 15 19:33:02 grac2 dhclient[1703]: DHCPOFFER from 192.168.1.50 Jul 15 19:33:02 grac2 dhclient[1703]: DHCPREQUEST on eth0 to 255.255.255.255 port 67 (xid=0x6781ea4f) Jul 15 19:33:02 grac2 dhclient[1703]: DHCPACK from 192.168.1.50 (xid=0x6781ea4f) Jul 15 19:33:02 grac2 NetworkManager[1690]: <info> (eth0): DHCPv4 state changed preinit -> bound Jul 15 19:37:56 grac2 NetworkManager[1690]: <info> (eth0): canceled DHCP transaction, DHCP client pid 1703 Rerun clufify for 2.nd node and test GNS connectivity: Verify GNS: $ ./bin/cluvfy comp gns -precrsinst -domain oracle-gns.example.com -vip 192.168.2.72 -verbose -n grac2 Verifying GNS integrity Checking GNS integrity... Checking if the GNS subdomain name is valid... The GNS subdomain name "oracle-gns.example.com" is a valid domain name Checking if the GNS VIP is a valid address... GNS VIP "192.168.2.72" resolves to a valid IP address Checking the status of GNS VIP... GNS integrity check passed Verification of GNS integrity was successful. Verify CRS for both nodes using newly created ASM disk and asmadmin group $ ./bin/cluvfy stage -pre crsinst -n grac1,grac2 -asm -asmgrp asmadmin -asmdev /dev/oracleasm/disks/DATA1,/dev/oracleasm/disks/DATA2,/dev/oracleasm/disks/DATA3 Performing pre-checks for cluster services setup Checking node reachability... Node reachability check passed from node "grac1" Checking user equivalence... User equivalence check passed for user "grid" Checking node connectivity... Checking hosts config file... Verification of the hosts config file successful Node connectivity passed for subnet "192.168.1.0" with node(s) grac2,grac1 TCP connectivity check passed for subnet "192.168.1.0" Node connectivity passed for subnet "192.168.2.0" with node(s) grac2,grac1 TCP connectivity check passed for subnet "192.168.2.0" Node connectivity passed for subnet "169.254.0.0" with node(s) grac2,grac1 TCP connectivity check passed for subnet "169.254.0.0" Interfaces found on subnet "169.254.0.0" that are likely candidates for VIP are: grac2 eth1:169.254.86.205 grac1 eth1:169.254.168.215 Interfaces found on subnet "192.168.2.0" that are likely candidates for a private interconnect are: grac2 eth1:192.168.2.102 grac1 eth1:192.168.2.101 Checking subnet mask consistency... Subnet mask consistency check passed for subnet "192.168.1.0". Subnet mask consistency check passed for subnet "192.168.2.0". Subnet mask consistency check passed for subnet "169.254.0.0". Subnet mask consistency check passed. Node connectivity check passed Checking ASMLib configuration. Check for ASMLib configuration passed. Total memory check passed Available memory check passed Swap space check passed Free disk space check passed for "grac2:/u01/app/11203/grid,grac2:/tmp" Free disk space check passed for "grac1:/u01/app/11203/grid,grac1:/tmp" Check for multiple users with UID value 501 passed User existence check passed for "grid" Group existence check passed for "oinstall" Group existence check passed for "dba" Group existence check passed for "asmadmin" Membership check for user "grid" in group "oinstall" [as Primary] passed Membership check for user "grid" in group "dba" passed Membership check for user "grid" in group "asmadmin" passed Run level check passed Hard limits check passed for "maximum open file descriptors" Hard limits check passed for "maximum user processes" System architecture check passed Kernel version check passed Kernel parameter check passed for "semmsl" Kernel parameter check passed for "semmns" Kernel parameter check passed for "semopm" Kernel parameter check passed for "semmni" Kernel parameter check passed for "shmmax" Kernel parameter check passed for "shmmni" Kernel parameter check passed for "shmall" Kernel parameter check passed for "file-max" Kernel parameter check passed for "ip_local_port_range" Kernel parameter check passed for "rmem_default" Kernel parameter check passed for "rmem_max" Kernel parameter check passed for "wmem_default" Kernel parameter check passed for "wmem_max" Kernel parameter check passed for "aio-max-nr" Package existence check passed for "binutils" Package existence check passed for "compat-libcap1" Package existence check passed for "compat-libstdc++-33(x86_64)" Package existence check passed for "libgcc(x86_64)" Package existence check passed for "libstdc++(x86_64)" Package existence check passed for "libstdc++-devel(x86_64)" Package existence check passed for "sysstat" Package existence check passed for "gcc" Package existence check passed for "gcc-c++" Package existence check passed for "ksh" Package existence check passed for "make" Package existence check passed for "glibc(x86_64)" Package existence check passed for "glibc-devel(x86_64)" Package existence check passed for "libaio(x86_64)" Package existence check passed for "libaio-devel(x86_64)" Check for multiple users with UID value passed Current group ID check passed Starting check for consistency of primary group of root user Check for consistency of root user's primary group passed Package existence check passed for "cvuqdisk" Checking Devices for ASM... Checking for shared devices... Device Device Type ------------------------------------ ------------------------ /dev/oracleasm/disks/DATA3 Disk /dev/oracleasm/disks/DATA2 Disk /dev/oracleasm/disks/DATA1 Disk Checking consistency of device owner across all nodes... Consistency check of device owner for "/dev/oracleasm/disks/DATA3" PASSED Consistency check of device owner for "/dev/oracleasm/disks/DATA1" PASSED Consistency check of device owner for "/dev/oracleasm/disks/DATA2" PASSED Checking consistency of device group across all nodes... Consistency check of device group for "/dev/oracleasm/disks/DATA3" PASSED Consistency check of device group for "/dev/oracleasm/disks/DATA1" PASSED Consistency check of device group for "/dev/oracleasm/disks/DATA2" PASSED Checking consistency of device permissions across all nodes... Consistency check of device permissions for "/dev/oracleasm/disks/DATA3" PASSED Consistency check of device permissions for "/dev/oracleasm/disks/DATA1" PASSED Consistency check of device permissions for "/dev/oracleasm/disks/DATA2" PASSED Checking consistency of device size across all nodes... Consistency check of device size for "/dev/oracleasm/disks/DATA3" PASSED Consistency check of device size for "/dev/oracleasm/disks/DATA1" PASSED Consistency check of device size for "/dev/oracleasm/disks/DATA2" PASSED UDev attributes check for ASM Disks started... ERROR: PRVF-9802 : Attempt to get udev info from node "grac2" failed ERROR: PRVF-9802 : Attempt to get udev info from node "grac1" failed UDev attributes check failed for ASM Disks Devices check for ASM passed Starting Clock synchronization checks using Network Time Protocol(NTP)... NTP Configuration file check started... NTP Configuration file check passed Checking daemon liveness... Liveness check passed for "ntpd" Check for NTP daemon or service alive passed on all nodes NTP daemon slewing option check passed NTP daemon's boot time configuration check for slewing option passed NTP common Time Server Check started... Check of common NTP Time Server passed Clock time offset check from NTP Time Server started... Clock time offset check passed Clock synchronization check using Network Time Protocol(NTP) passed Core file name pattern consistency check passed. User "grid" is not part of "root" group. Check passed Default user file creation mask check passed Checking consistency of file "/etc/resolv.conf" across nodes File "/etc/resolv.conf" does not have both domain and search entries defined domain entry in file "/etc/resolv.conf" is consistent across nodes search entry in file "/etc/resolv.conf" is consistent across nodes All nodes have one search entry defined in file "/etc/resolv.conf" PRVF-5636 : The DNS response time for an unreachable node exceeded "15000" ms on following nodes: grac2,grac1 File "/etc/resolv.conf" is not consistent across nodes Time zone consistency check passed Pre-check for cluster services setup was unsuccessful on all the nodes. Ignore PRVF-9802 , PRVF-5636. For details check the following link.
Install Clusterware Software
As user Root # xhost + access control disabled, clients can connect from any host As user Grid $ xclock ( Testing X connection ) $ cd /KITS/Oracle/11.2.0.3/Linux_64/grid ( your grid staging area ) $ ./runInstaller --> Important : Select Installation type : Advanced Installation Cluster name GRACE2 Scan name: GRACE2-scan.grid.example.com Scan port: 1521 Configure GNS GNS sub domain: grid.example.com GNS VIP address: 192.168.1.55 ( This address shouldn't be in use: # ping 192.168.1.55 should fail ) Hostname: grac1.example.com Virtual hostnames : AUTO Hostname: grac1.example.com Virtual hostnames : AUTO Test and configure SSH connectivity Configure ASM disk string: /dev/oracleasm/disks/* ASM password: sys Don't user IPM Dont't change groups ORACLE_BASE: /u01/app/grid Sofware Location : /u01/app/11.2.0/grid --> Check OUI Prerequisites Check -> Ignore the wellknown PRVF-5636, PRVF-9802 errors/warnings ( see the former clufvfy reports ) Install software and run the related root.sh scripts Run on grac1: /u01/app/11203/grid/root.sh Performing root user operation for Oracle 11g The following environment variables are set as: ORACLE_OWNER= grid ORACLE_HOME= /u01/app/11203/grid Enter the full pathname of the local bin directory: [/usr/local/bin]: The contents of "dbhome" have not changed. No need to overwrite. The contents of "oraenv" have not changed. No need to overwrite. The contents of "coraenv" have not changed. No need to overwrite. Entries will be added to the /etc/oratab file as needed by Database Configuration Assistant when a database is created Finished running generic part of root script. Now product-specific root actions will be performed. Using configuration parameter file: /u01/app/11203/grid/crs/install/crsconfig_params Creating trace directory User ignored Prerequisites during installation OLR initialization - successful root wallet root wallet cert root cert export peer wallet profile reader wallet pa wallet peer wallet keys pa wallet keys peer cert request pa cert request peer cert pa cert peer root cert TP profile reader root cert TP pa root cert TP peer pa cert TP pa peer cert TP profile reader pa cert TP profile reader peer cert TP peer user cert pa user cert Adding Clusterware entries to upstart CRS-2672: Attempting to start 'ora.mdnsd' on 'grac1' CRS-2676: Start of 'ora.mdnsd' on 'grac1' succeeded CRS-2672: Attempting to start 'ora.gpnpd' on 'grac1' CRS-2676: Start of 'ora.gpnpd' on 'grac1' succeeded CRS-2672: Attempting to start 'ora.cssdmonitor' on 'grac1' CRS-2672: Attempting to start 'ora.gipcd' on 'grac1' CRS-2676: Start of 'ora.cssdmonitor' on 'grac1' succeeded CRS-2676: Start of 'ora.gipcd' on 'grac1' succeeded CRS-2672: Attempting to start 'ora.cssd' on 'grac1' CRS-2672: Attempting to start 'ora.diskmon' on 'grac1' CRS-2676: Start of 'ora.diskmon' on 'grac1' succeeded CRS-2676: Start of 'ora.cssd' on 'grac1' succeeded ASM created and started successfully. Disk Group DATA created successfully. clscfg: -install mode specified Successfully accumulated necessary OCR keys. Creating OCR keys for user 'root', privgrp 'root'.. Operation successful. CRS-4256: Updating the profile Successful addition of voting disk 3ee007b399cc4f59bfa0fc80ff3fa9ff. Successful addition of voting disk 7a73147a81dc4f71bfc8757343aee181. Successful addition of voting disk 25fcfbdb854a4f49bf0addd0fa32d0a2. Successfully replaced voting disk group with +DATA. CRS-4256: Updating the profile CRS-4266: Voting file(s) successfully replaced ## STATE File Universal Id File Name Disk group -- ----- ----------------- --------- --------- 1. ONLINE 3ee007b399cc4f59bfa0fc80ff3fa9ff (/dev/oracleasm/disks/DATA1) [DATA] 2. ONLINE 7a73147a81dc4f71bfc8757343aee181 (/dev/oracleasm/disks/DATA2) [DATA] 3. ONLINE 25fcfbdb854a4f49bf0addd0fa32d0a2 (/dev/oracleasm/disks/DATA3) [DATA] Located 3 voting disk(s). CRS-2672: Attempting to start 'ora.asm' on 'grac1' CRS-2676: Start of 'ora.asm' on 'grac1' succeeded CRS-2672: Attempting to start 'ora.DATA.dg' on 'grac1' CRS-2676: Start of 'ora.DATA.dg' on 'grac1' succeeded Configure Oracle Grid Infrastructure for a Cluster ... succeeded Run on grac2: /u01/app/11203/grid/root.sh # /u01/app/11203/grid/root.sh Performing root user operation for Oracle 11g The following environment variables are set as: ORACLE_OWNER= grid ORACLE_HOME= /u01/app/11203/grid Enter the full pathname of the local bin directory: [/usr/local/bin]: The contents of "dbhome" have not changed. No need to overwrite. The contents of "oraenv" have not changed. No need to overwrite. The contents of "coraenv" have not changed. No need to overwrite. Entries will be added to the /etc/oratab file as needed by Database Configuration Assistant when a database is created Finished running generic part of root script. Now product-specific root actions will be performed. Using configuration parameter file: /u01/app/11203/grid/crs/install/crsconfig_params Creating trace directory User ignored Prerequisites during installation OLR initialization - successful Adding Clusterware entries to upstart CRS-4402: The CSS daemon was started in exclusive mode but found an active CSS daemon on node grac1, number 1, and is terminating An active cluster was found during exclusive startup, restarting to join the cluster Configure Oracle Grid Infrastructure for a Cluster ... succeeded Run cluvfyy and crsctl to verify Oracle Grid Installation $ ./bin/cluvfy stage -post crsinst -n grac1,grac2 -verbose Performing post-checks for cluster services setup Checking node reachability... Check: Node reachability from node "grac1" Destination Node Reachable? ------------------------------------ ------------------------ grac2 yes grac1 yes Result: Node reachability check passed from node "grac1" Checking user equivalence... Check: User equivalence for user "grid" Node Name Status ------------------------------------ ------------------------ grac2 passed grac1 passed Result: User equivalence check passed for user "grid" Checking node connectivity... Checking hosts config file... Node Name Status ------------------------------------ ------------------------ grac2 passed grac1 passed Verification of the hosts config file successful Interface information for node "grac2" Name IP Address Subnet Gateway Def. Gateway HW Address MTU ------ --------------- --------------- --------------- --------------- ----------------- ------ eth0 192.168.1.62 192.168.1.0 0.0.0.0 168.1.0.1 08:00:27:1F:2E:33 1500 eth0 192.168.1.112 192.168.1.0 0.0.0.0 168.1.0.1 08:00:27:1F:2E:33 1500 eth0 192.168.1.108 192.168.1.0 0.0.0.0 168.1.0.1 08:00:27:1F:2E:33 1500 eth1 192.168.2.102 192.168.2.0 0.0.0.0 168.1.0.1 08:00:27:8E:6D:24 1500 eth1 169.254.86.205 169.254.0.0 0.0.0.0 168.1.0.1 08:00:27:8E:6D:24 1500 Interface information for node "grac1" Name IP Address Subnet Gateway Def. Gateway HW Address MTU ------ --------------- --------------- --------------- --------------- ----------------- ------ eth0 192.168.1.61 192.168.1.0 0.0.0.0 192.168.1.1 08:00:27:6E:17:DB 1500 eth0 192.168.1.55 192.168.1.0 0.0.0.0 192.168.1.1 08:00:27:6E:17:DB 1500 eth0 192.168.1.110 192.168.1.0 0.0.0.0 192.168.1.1 08:00:27:6E:17:DB 1500 eth0 192.168.1.109 192.168.1.0 0.0.0.0 192.168.1.1 08:00:27:6E:17:DB 1500 eth0 192.168.1.107 192.168.1.0 0.0.0.0 192.168.1.1 08:00:27:6E:17:DB 1500 eth1 192.168.2.101 192.168.2.0 0.0.0.0 192.168.1.1 08:00:27:F5:31:22 1500 eth1 169.254.168.215 169.254.0.0 0.0.0.0 192.168.1.1 08:00:27:F5:31:22 1500 Check: Node connectivity for interface "eth0" Source Destination Connected? ------------------------------ ------------------------------ ---------------- grac2[192.168.1.62] grac2[192.168.1.112] yes grac2[192.168.1.62] grac2[192.168.1.108] yes grac2[192.168.1.62] grac1[192.168.1.61] yes grac2[192.168.1.62] grac1[192.168.1.55] yes grac2[192.168.1.62] grac1[192.168.1.110] yes grac2[192.168.1.62] grac1[192.168.1.109] yes grac2[192.168.1.62] grac1[192.168.1.107] yes grac2[192.168.1.112] grac2[192.168.1.108] yes grac2[192.168.1.112] grac1[192.168.1.61] yes grac2[192.168.1.112] grac1[192.168.1.55] yes grac2[192.168.1.112] grac1[192.168.1.110] yes grac2[192.168.1.112] grac1[192.168.1.109] yes grac2[192.168.1.112] grac1[192.168.1.107] yes grac2[192.168.1.108] grac1[192.168.1.61] yes grac2[192.168.1.108] grac1[192.168.1.55] yes grac2[192.168.1.108] grac1[192.168.1.110] yes grac2[192.168.1.108] grac1[192.168.1.109] yes grac2[192.168.1.108] grac1[192.168.1.107] yes grac1[192.168.1.61] grac1[192.168.1.55] yes grac1[192.168.1.61] grac1[192.168.1.110] yes grac1[192.168.1.61] grac1[192.168.1.109] yes grac1[192.168.1.61] grac1[192.168.1.107] yes grac1[192.168.1.55] grac1[192.168.1.110] yes grac1[192.168.1.55] grac1[192.168.1.109] yes grac1[192.168.1.55] grac1[192.168.1.107] yes grac1[192.168.1.110] grac1[192.168.1.109] yes grac1[192.168.1.110] grac1[192.168.1.107] yes grac1[192.168.1.109] grac1[192.168.1.107] yes Result: Node connectivity passed for interface "eth0" Check: TCP connectivity of subnet "192.168.1.0" Source Destination Connected? ------------------------------ ------------------------------ ---------------- grac1:192.168.1.61 grac2:192.168.1.62 passed grac1:192.168.1.61 grac2:192.168.1.112 passed grac1:192.168.1.61 grac2:192.168.1.108 passed grac1:192.168.1.61 grac1:192.168.1.55 passed grac1:192.168.1.61 grac1:192.168.1.110 passed grac1:192.168.1.61 grac1:192.168.1.109 passed grac1:192.168.1.61 grac1:192.168.1.107 passed Result: TCP connectivity check passed for subnet "192.168.1.0" Check: Node connectivity for interface "eth1" Source Destination Connected? ------------------------------ ------------------------------ ---------------- grac2[192.168.2.102] grac1[192.168.2.101] yes Result: Node connectivity passed for interface "eth1" Check: TCP connectivity of subnet "192.168.2.0" Source Destination Connected? ------------------------------ ------------------------------ ---------------- grac1:192.168.2.101 grac2:192.168.2.102 passed Result: TCP connectivity check passed for subnet "192.168.2.0" Checking subnet mask consistency... Subnet mask consistency check passed for subnet "192.168.1.0". Subnet mask consistency check passed for subnet "192.168.2.0". Subnet mask consistency check passed. Result: Node connectivity check passed Checking multicast communication... Checking subnet "192.168.1.0" for multicast communication with multicast group "230.0.1.0"... Check of subnet "192.168.1.0" for multicast communication with multicast group "230.0.1.0" passed. Checking subnet "192.168.2.0" for multicast communication with multicast group "230.0.1.0"... Check of subnet "192.168.2.0" for multicast communication with multicast group "230.0.1.0" passed. Check of multicast communication passed. Check: Time zone consistency Result: Time zone consistency check passed Checking Oracle Cluster Voting Disk configuration... ASM Running check passed. ASM is running on all specified nodes Oracle Cluster Voting Disk configuration check passed Checking Cluster manager integrity... Checking CSS daemon... Node Name Status ------------------------------------ ------------------------ grac2 running grac1 running Oracle Cluster Synchronization Services appear to be online. Cluster manager integrity check passed UDev attributes check for OCR locations started... Result: UDev attributes check passed for OCR locations UDev attributes check for Voting Disk locations started... Result: UDev attributes check passed for Voting Disk locations Check default user file creation mask Node Name Available Required Comment ------------ ------------------------ ------------------------ ---------- grac2 22 0022 passed grac1 22 0022 passed Result: Default user file creation mask check passed Checking cluster integrity... Node Name ------------------------------------ grac1 grac2 Cluster integrity check passed Checking OCR integrity... Checking the absence of a non-clustered configuration... All nodes free of non-clustered, local-only configurations ASM Running check passed. ASM is running on all specified nodes Checking OCR config file "/etc/oracle/ocr.loc"... OCR config file "/etc/oracle/ocr.loc" check successful Disk group for ocr location "+DATA" available on all the nodes NOTE: This check does not verify the integrity of the OCR contents. Execute 'ocrcheck' as a privileged user to verify the contents of OCR. OCR integrity check passed Checking CRS integrity... Clusterware version consistency passed The Oracle Clusterware is healthy on node "grac2" The Oracle Clusterware is healthy on node "grac1" CRS integrity check passed Checking node application existence... Checking existence of VIP node application (required) Node Name Required Running? Comment ------------ ------------------------ ------------------------ ---------- grac2 yes yes passed grac1 yes yes passed VIP node application check passed Checking existence of NETWORK node application (required) Node Name Required Running? Comment ------------ ------------------------ ------------------------ ---------- grac2 yes yes passed grac1 yes yes passed NETWORK node application check passed Checking existence of GSD node application (optional) Node Name Required Running? Comment ------------ ------------------------ ------------------------ ---------- grac2 no no exists grac1 no no exists GSD node application is offline on nodes "grac2,grac1" Checking existence of ONS node application (optional) Node Name Required Running? Comment ------------ ------------------------ ------------------------ ---------- grac2 no yes passed grac1 no yes passed ONS node application check passed Checking Single Client Access Name (SCAN)... SCAN Name Node Running? ListenerName Port Running? ---------------- ------------ ------------ ------------ ------------ ------------ GRACE2-scan.grid.example.com grac2 true LISTENER_SCAN1 1521 true GRACE2-scan.grid.example.com grac1 true LISTENER_SCAN2 1521 true GRACE2-scan.grid.example.com grac1 true LISTENER_SCAN3 1521 true Checking TCP connectivity to SCAN Listeners... Node ListenerName TCP connectivity? ------------ ------------------------ ------------------------ grac1 LISTENER_SCAN1 yes grac1 LISTENER_SCAN2 yes grac1 LISTENER_SCAN3 yes TCP connectivity to SCAN Listeners exists on all cluster nodes Checking name resolution setup for "GRACE2-scan.grid.example.com"... SCAN Name IP Address Status Comment ------------ ------------------------ ------------------------ ---------- GRACE2-scan.grid.example.com 192.168.1.110 passed GRACE2-scan.grid.example.com 192.168.1.109 passed GRACE2-scan.grid.example.com 192.168.1.108 passed Verification of SCAN VIP and Listener setup passed Checking OLR integrity... Checking OLR config file... OLR config file check successful Checking OLR file attributes... OLR file check successful WARNING: This check does not verify the integrity of the OLR contents. Execute 'ocrcheck -local' as a privileged user to verify the contents of OLR. OLR integrity check passed Checking GNS integrity... Checking if the GNS subdomain name is valid... The GNS subdomain name "grid.example.com" is a valid domain name Checking if the GNS VIP belongs to same subnet as the public network... Public network subnets "192.168.1.0" match with the GNS VIP "192.168.1.0" Checking if the GNS VIP is a valid address... GNS VIP "192.168.1.55" resolves to a valid IP address Checking the status of GNS VIP... Checking if FDQN names for domain "grid.example.com" are reachable GNS resolved IP addresses are reachable GNS resolved IP addresses are reachable GNS resolved IP addresses are reachable GNS resolved IP addresses are reachable GNS resolved IP addresses are reachable Checking status of GNS resource... Node Running? Enabled? ------------ ------------------------ ------------------------ grac2 no yes grac1 yes yes GNS resource configuration check passed Checking status of GNS VIP resource... Node Running? Enabled? ------------ ------------------------ ------------------------ grac2 no yes grac1 yes yes GNS VIP resource configuration check passed. GNS integrity check passed Checking to make sure user "grid" is not in "root" group Node Name Status Comment ------------ ------------------------ ------------------------ grac2 passed does not exist grac1 passed does not exist Result: User "grid" is not part of "root" group. Check passed Checking if Clusterware is installed on all nodes... Check of Clusterware install passed Checking if CTSS Resource is running on all nodes... Check: CTSS Resource running on all nodes Node Name Status ------------------------------------ ------------------------ grac2 passed grac1 passed Result: CTSS resource check passed Querying CTSS for time offset on all nodes... Result: Query of CTSS for time offset passed Check CTSS state started... Check: CTSS state Node Name State ------------------------------------ ------------------------ grac2 Observer grac1 Observer CTSS is in Observer state. Switching over to clock synchronization checks using NTP Starting Clock synchronization checks using Network Time Protocol(NTP)... NTP Configuration file check started... The NTP configuration file "/etc/ntp.conf" is available on all nodes NTP Configuration file check passed Checking daemon liveness... Check: Liveness for "ntpd" Node Name Running? ------------------------------------ ------------------------ grac2 yes grac1 yes Result: Liveness check passed for "ntpd" Check for NTP daemon or service alive passed on all nodes Checking NTP daemon command line for slewing option "-x" Check: NTP daemon command line Node Name Slewing Option Set? ------------------------------------ ------------------------ grac2 yes grac1 yes Result: NTP daemon slewing option check passed Checking NTP daemon's boot time configuration, in file "/etc/sysconfig/ntpd", for slewing option "-x" Check: NTP daemon's boot time configuration Node Name Slewing Option Set? ------------------------------------ ------------------------ grac2 yes grac1 yes Result: NTP daemon's boot time configuration check for slewing option passed Checking whether NTP daemon or service is using UDP port 123 on all nodes Check for NTP daemon or service using UDP port 123 Node Name Port Open? ------------------------------------ ------------------------ grac2 yes grac1 yes NTP common Time Server Check started... NTP Time Server ".LOCL." is common to all nodes on which the NTP daemon is running Check of common NTP Time Server passed Clock time offset check from NTP Time Server started... Checking on nodes "[grac2, grac1]"... Check: Clock time offset from NTP Time Server Time Server: .LOCL. Time Offset Limit: 1000.0 msecs Node Name Time Offset Status ------------ ------------------------ ------------------------ grac2 0.0 passed grac1 0.0 passed Time Server ".LOCL." has time offsets that are within permissible limits for nodes "[grac2, grac1]". Clock time offset check passed Result: Clock synchronization check using Network Time Protocol(NTP) passed Oracle Cluster Time Synchronization Services check passed Checking VIP configuration. Checking VIP Subnet configuration. Check for VIP Subnet configuration passed. Checking VIP reachability Check for VIP reachability passed. Post-check for cluster services setup was successful. Checking CRS status after installation] $ my_crs_stat NAME TARGET STATE SERVER STATE_DETAILS ------------------------- ---------- ---------- ------------ ------------------ ora.DATA.dg ONLINE ONLINE grac1 ora.DATA.dg ONLINE ONLINE grac2 ora.LISTENER.lsnr ONLINE ONLINE grac1 ora.LISTENER.lsnr ONLINE ONLINE grac2 ora.asm ONLINE ONLINE grac1 Started ora.asm ONLINE ONLINE grac2 Started ora.gsd OFFLINE OFFLINE grac1 ora.gsd OFFLINE OFFLINE grac2 ora.net1.network ONLINE ONLINE grac1 ora.net1.network ONLINE ONLINE grac2 ora.ons ONLINE ONLINE grac1 ora.ons ONLINE ONLINE grac2 ora.LISTENER_SCAN1.lsnr ONLINE ONLINE grac2 ora.LISTENER_SCAN2.lsnr ONLINE ONLINE grac1 ora.LISTENER_SCAN3.lsnr ONLINE ONLINE grac1 ora.cvu ONLINE ONLINE grac1 ora.gns ONLINE ONLINE grac1 ora.gns.vip ONLINE ONLINE grac1 ora.grac1.vip ONLINE ONLINE grac1 ora.grac2.vip ONLINE ONLINE grac2 ora.oc4j ONLINE ONLINE grac1 ora.scan1.vip ONLINE ONLINE grac2 ora.scan2.vip ONLINE ONLINE grac1 ora.scan3.vip ONLINE ONLINE grac1 Grid post installation - Ologgerd process cosumes high CPU time It had been noticed that after a while, the ologgerd process can consume excessive CPU resource. The ologgerd is part of Oracle Cluster Health Monitor and used by Oracle Support to troubleshoot RAC problems. You can check that by starting top: (sometime up we see up to 60% WA states ) top - 15:02:38 up 15 min, 6 users, load average: 3.70, 2.54, 1.78 Tasks: 215 total, 2 running, 213 sleeping, 0 stopped, 0 zombie Cpu(s): 3.6%us, 8.9%sy, 0.0%ni, 55.4%id, 31.4%wa, 0.0%hi, 0.8%si, 0.0%st Mem: 3234376k total, 2512568k used, 721808k free, 108508k buffers Swap: 3227644k total, 0k used, 3227644k free, 1221196k cached PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND 5602 root RT 0 501m 145m 60m S 48.1 4.6 0:31.29 ologgerd If ologgerd process is consuming a lot of CPU, you can stop it by executing: # crsctl stop resource ora.crf -init Now top looks good as IDLE CPU time increases from 55 % to 95 % ! hrac1: top - 15:07:56 up 20 min, 6 users, load average: 2.57, 3.33, 2.41 Tasks: 212 total, 1 running, 211 sleeping, 0 stopped, 0 zombie Cpu(s): 1.3%us, 4.2%sy, 0.0%ni, 94.3%id, 0.1%wa, 0.0%hi, 0.2%si, 0.0%st Mem: 3234376k total, 2339268k used, 895108k free, 132604k buffers Swap: 3227644k total, 0k used, 3227644k free, 1126964k cached hrac2: top - 15:48:37 up 33 min, 3 users, load average: 2.63, 2.40, 2.13 Tasks: 204 total, 1 running, 203 sleeping, 0 stopped, 0 zombie Cpu(s): 0.9%us, 3.3%sy, 0.0%ni, 95.6%id, 0.1%wa, 0.0%hi, 0.2%si, 0.0%st Mem: 2641484k total, 1975444k used, 666040k free, 158212k buffers Swap: 3227644k total, 0k used, 3227644k free, 993328k cached If you want to disable ologgerd permanently, then execute: # crsctl delete resource ora.crf -init
Fixing a failed GRID Installation
Fixing a failed Grid Installation ( runt this commands on all instances ) [grid@grac31 ~]$ rm -rf /u01/app/11203/grid/* [grid@grac31 ~]$ rm /u01/app/oraInventory/*
Install RDBMS and create RAC database
Login as Oracle user and verify the accout $ id uid=54321(oracle) gid=54321(oinstall) groups=54321(oinstall),501(vboxsf),506(asmdba),54322(dba) $ env | grep ORA ORACLE_BASE=/u01/app/oracle ORACLE_SID=RACE2 ORACLE_HOME=/u01/app/oracle/product/11203/racdb Verfiy system by running cluvfy with: stage -pre dbinst $ ./bin/cluvfy stage -pre dbinst -n grac1,grac2 Performing pre-checks for database installation Checking node reachability... Node reachability check passed from node "grac1" Checking user equivalence... User equivalence check passed for user "oracle" Checking node connectivity... Checking hosts config file... Verification of the hosts config file successful Check: Node connectivity for interface "eth0" Node connectivity passed for interface "eth0" TCP connectivity check passed for subnet "192.168.1.0" Check: Node connectivity for interface "eth1" Node connectivity passed for interface "eth1" TCP connectivity check passed for subnet "192.168.2.0" Checking subnet mask consistency... Subnet mask consistency check passed for subnet "192.168.1.0". Subnet mask consistency check passed for subnet "192.168.2.0". Subnet mask consistency check passed. Node connectivity check passed Checking multicast communication... Checking subnet "192.168.1.0" for multicast communication with multicast group "230.0.1.0"... Check of subnet "192.168.1.0" for multicast communication with multicast group "230.0.1.0" passed. Checking subnet "192.168.2.0" for multicast communication with multicast group "230.0.1.0"... Check of subnet "192.168.2.0" for multicast communication with multicast group "230.0.1.0" passed. Check of multicast communication passed. Total memory check passed Available memory check passed Swap space check passed Free disk space check passed for "grac2:/tmp" Free disk space check passed for "grac1:/tmp" Check for multiple users with UID value 54321 passed User existence check passed for "oracle" Group existence check passed for "oinstall" Group existence check passed for "dba" Membership check for user "oracle" in group "oinstall" [as Primary] passed Membership check for user "oracle" in group "dba" passed Run level check passed Hard limits check passed for "maximum open file descriptors" Hard limits check passed for "maximum user processes" System architecture check passed Kernel version check passed ... Check for multiple users with UID value 0 passed Current group ID check passed Starting check for consistency of primary group of root user Check for consistency of root user's primary group passed Default user file creation mask check passed Checking CRS integrity... Clusterware version consistency passed CRS integrity check passed Checking Cluster manager integrity... Checking CSS daemon... Oracle Cluster Synchronization Services appear to be online. Cluster manager integrity check passed Checking node application existence... Checking existence of VIP node application (required) VIP node application check passed Checking existence of NETWORK node application (required) NETWORK node application check passed Checking existence of GSD node application (optional) GSD node application is offline on nodes "grac2,grac1" Checking existence of ONS node application (optional) ONS node application check passed Checking if Clusterware is installed on all nodes... Check of Clusterware install passed Checking if CTSS Resource is running on all nodes... CTSS resource check passed Querying CTSS for time offset on all nodes... Query of CTSS for time offset passed Check CTSS state started... CTSS is in Observer state. Switching over to clock synchronization checks using NTP Starting Clock synchronization checks using Network Time Protocol(NTP)... NTP Configuration file check started... NTP Configuration file check passed Checking daemon liveness... Liveness check passed for "ntpd" Check for NTP daemon or service alive passed on all nodes NTP daemon slewing option check passed NTP daemon's boot time configuration check for slewing option passed NTP common Time Server Check started... Check of common NTP Time Server passed Clock time offset check from NTP Time Server started... Clock time offset check passed Clock synchronization check using Network Time Protocol(NTP) passed Oracle Cluster Time Synchronization Services check passed Checking consistency of file "/etc/resolv.conf" across nodes File "/etc/resolv.conf" does not have both domain and search entries defined domain entry in file "/etc/resolv.conf" is consistent across nodes search entry in file "/etc/resolv.conf" is consistent across nodes All nodes have one search entry defined in file "/etc/resolv.conf" PRVF-5636 : The DNS response time for an unreachable node exceeded "15000" ms on following nodes: grac2,grac1 File "/etc/resolv.conf" is not consistent across nodes Time zone consistency check passed Checking Single Client Access Name (SCAN)... Checking TCP connectivity to SCAN Listeners... TCP connectivity to SCAN Listeners exists on all cluster nodes Checking name resolution setup for "GRACE2-scan.grid.example.com"... Verification of SCAN VIP and Listener setup passed Checking VIP configuration. Checking VIP Subnet configuration. Check for VIP Subnet configuration passed. Checking VIP reachability Check for VIP reachability passed. ASM and CRS versions are compatible Database Clusterware version compatibility passed Pre-check for database installation was unsuccessful on all the nodes. Run cluvfy with: stage -pre dbcfg $ ./bin/cluvfy stage -pre dbcfg -n grac1,grac2 -d $ORACLE_HOME Performing pre-checks for database configuration ERROR: Unable to determine OSDBA group from Oracle Home "/u01/app/oracle/product/11203/racdb" Checking node reachability... Node reachability check passed from node "grac1" Checking user equivalence... User equivalence check passed for user "oracle" Checking node connectivity... Checking hosts config file... Verification of the hosts config file successful Check: Node connectivity for interface "eth0" Node connectivity passed for interface "eth0" ERROR: PRVF-7617 : Node connectivity between "grac1 : 192.168.1.61" and "grac2 : 192.168.1.108" failed TCP connectivity check failed for subnet "192.168.1.0" Check: Node connectivity for interface "eth1" Node connectivity passed for interface "eth1" TCP connectivity check passed for subnet "192.168.2.0" Checking subnet mask consistency... Subnet mask consistency check passed for subnet "192.168.1.0". Subnet mask consistency check passed for subnet "192.168.2.0". Subnet mask consistency check passed. Node connectivity check failed Checking multicast communication... Checking subnet "192.168.1.0" for multicast communication with multicast group "230.0.1.0"... Check of subnet "192.168.1.0" for multicast communication with multicast group "230.0.1.0" passed. Checking subnet "192.168.2.0" for multicast communication with multicast group "230.0.1.0"... Check of subnet "192.168.2.0" for multicast communication with multicast group "230.0.1.0" passed. Check of multicast communication passed. Total memory check passed Available memory check passed Swap space check passed Free disk space check passed for "grac2:/u01/app/oracle/product/11203/racdb,grac2:/tmp" Free disk space check passed for "grac1:/u01/app/oracle/product/11203/racdb,grac1:/tmp" Check for multiple users with UID value 54321 passed User existence check passed for "oracle" Group existence check passed for "oinstall" Membership check for user "oracle" in group "oinstall" [as Primary] passed Run level check passed Hard limits check passed for "maximum open file descriptors" Hard limits check passed for "maximum user processes" System architecture check passed Kernel version check passed ... Package existence check passed for "libaio-devel(x86_64)" Check for multiple users with UID value 0 passed Current group ID check passed Starting check for consistency of primary group of root user Check for consistency of root user's primary group passed Checking CRS integrity... Clusterware version consistency passed CRS integrity check passed Checking node application existence... Checking existence of VIP node application (required) VIP node application check passed Checking existence of NETWORK node application (required) NETWORK node application check passed Checking existence of GSD node application (optional) GSD node application is offline on nodes "grac2,grac1" Checking existence of ONS node application (optional) ONS node application check passed Time zone consistency check passed Pre-check for database configuration was unsuccessful on all the nodes. Ignore ERROR: Unable to determine OSDBA group from Oracle Home "/u01/app/oracle/product/11203/racdb" -> Oracle software isn't software installed yet and cluvfy can't find $ORACLE_HOME/bin/osdbagrp stat("/u01/app/oracle/product/11203/racdb/bin/osdbagrp", 0x7fff2fd6e530) = -1 ENOENT (No such file or directory) Run only cluvfy stage -pre dbcfg only after you have installed the software and before you have created the database. Run Installer As user Root # xhost + access control disabled, clients can connect from any host As user Oracle $ xclock ( Testing X connection ) $ cd /KITS/Oracle/11.2.0.3/Linux_64/database ( rdbms staging area ) $ ./runInstaller ( select SERVER class ) Node Name : grac1,grac2 Storage type : ASM Location : DATA OSDBDBA group : asmdba Global database name: GRACE2 On grac1 run: /u01/app/oracle/product/11203/racdb/root.sh On grac2 run: /u01/app/oracle/product/11203/racdb/root.sh Enterprise Manager Database Control URL - (RACE2) : https://hrac1.de.oracle.com:1158/em Verify Rac Install $ my_crs_stat NAME TARGET STATE SERVER STATE_DETAILS ------------------------- ---------- ---------- ------------ ------------------ ora.DATA.dg ONLINE ONLINE grac1 ora.DATA.dg ONLINE ONLINE grac2 ora.LISTENER.lsnr ONLINE ONLINE grac1 ora.LISTENER.lsnr ONLINE ONLINE grac2 ora.asm ONLINE ONLINE grac1 Started ora.asm ONLINE ONLINE grac2 Started ora.gsd OFFLINE OFFLINE grac1 ora.gsd OFFLINE OFFLINE grac2 ora.net1.network ONLINE ONLINE grac1 ora.net1.network ONLINE ONLINE grac2 ora.ons ONLINE ONLINE grac1 ora.ons ONLINE ONLINE grac2 ora.LISTENER_SCAN1.lsnr ONLINE ONLINE grac2 ora.LISTENER_SCAN2.lsnr ONLINE ONLINE grac1 ora.LISTENER_SCAN3.lsnr ONLINE ONLINE grac1 ora.cvu ONLINE ONLINE grac1 ora.gns ONLINE ONLINE grac1 ora.gns.vip ONLINE ONLINE grac1 ora.grac1.vip ONLINE ONLINE grac1 ora.grac2.vip ONLINE ONLINE grac2 ora.grace2.db ONLINE ONLINE grac1 Open ora.grace2.db ONLINE ONLINE grac2 Open ora.oc4j ONLINE ONLINE grac1 ora.scan1.vip ONLINE ONLINE grac2 ora.scan2.vip ONLINE ONLINE grac1 ora.scan3.vip ONLINE ONLINE grac1 $ srvctl status database -d GRACE2 Instance GRACE21 is running on node grac1 Instance GRACE22 is running on node grac2 $GRID_HOME/bin/olsnodes -n grac1 1 grac2 2
所有评论(0)