You are on page 1of 111

Installing Oracle Database 10g Rel.2 (10.2.0.1.

0) RAC on Red Hat Enterprise Linux AS Version 4 Update 5 using VMware Server Introduction
One of the biggest obstacles preventing people from setting up test RAC environments is the requirement for shared storage. In a production environment, shared storage is often provided by a SAN or high-end NAS device, but both of these options are very expensive when all you want to do is get some experience installing and using RAC. A cheaper alternative is to use a FireWire disk enclosure to allow two machines to access the same disk(s), but that still costs money and requires two servers. A third option is to use VMware Server to fake the shared storage. Using VMware Server you can run multiple Virtual Machines on a single server, allowing you to run both RAC nodes on a single machine. In addition, it allows you to set up shared virtual disks, overcoming the obstacle of expensive shared storage.

1. VMware Server Installation


For this article, I will use Windows XP Professional with Service Pack 2 as the host OS and Red Hat Enterprise Linux AS Version 4 Update 5 as the guest OS. I have demonstrated the installation process with screen shots. Detailed explanation will be added where necessary.

Click the OK button and continue.

Enter the serial number.

Double-click the VMware Server Console icon on your desktop.

Click the OK button.

2. Virtual Machine Setup


Click File New Virtual Machine.

Uncheck Make this virtual machine private.

Uncheck Allocate all disk space now and check Split disk into 2 GB files.

Click Edit virtual machine settings.

Click the Add button.

Select Ethernet Adapter and click the Next button.

Again click Edit, select the CD-ROM, browse the ISO image and click OK button.

3. Guest Operating System Installation


Click the Start this virtual machine.

Click the Yes button.

Click the Proceed button.

Hint: The date & time should be smaller than the host machine. This will help to synchronize time later on.

Click the Continue button.

4. Oracle Installation Prerequisites


Perform the following steps as the root user.

The /etc/hosts file must contain the following information. 127.0.0.1 localhost.localdomain localhost # Public 192.168.2.101 rac1 192.168.2.102 rac2 #Private 192.168.0.101 rac1-priv 192.168.0.102 rac2-priv #Virtual 192.168.2.111 rac1-vip 192.168.2.112 rac2-vip Run these commands. # service sendmail stop # chkconfig --level 345 sendmail off Add the following lines to the /etc/sysctl.conf file. kernel.shmall = 2097152 kernel.shmmax = 2147483648 kernel.shmmni = 4096 kernel.sem = 250 32000 100 128 fs.file-max = 65536 net.ipv4.ip_local_port_range = 1024 65000 net.core.rmem_default=262144 net.core.rmem_max=262144 net.core.wmem_default=262144 net.core.wmem_max=262144 Run the following command to change the current kernel parameters. /sbin/sysctl -p Add the following lines to the /etc/security/limits.conf file. * * * * soft nproc 2047 hard nproc 16384 soft nofile 1024 hard nofile 65536

Add the following line to the /etc/pam.d/login file. session required /lib/security/pam_limits.so

Disable secure linux by editing the /etc/selinux/config file, making sure the SELINUX flag is set as follows. SELINUX=disabled Alternatively, this alteration can be done using the GUI tool (Applications > System Settings > Security Level). Click on the SELinux tab and disable the feature. Set the hangcheck kernel module parameters by adding the following line to the /etc/modprobe.conf file.

options hangcheck-timer hangcheck_tick=30 hangcheck_margin=180 To load the module immediately, execute: modprobe -v hangcheck-timer Create the new groups and users. groupadd oinstall groupadd dba groupadd oper useradd -g oinstall -G dba oracle passwd oracle Create the directories in which the Oracle software will be installed. mkdir -p /u01/crs/oracle/product/10.2.0/crs mkdir -p /u01/app/oracle/product/10.2.0/db_1 mkdir /u02 chown -R oracle:oinstall /u01 /u02 chmod -R 775 /u01 /u02 During the installation, both RSH and RSH-Server were installed. Enable remote shell and rlogin by doing the following. chkconfig rsh on chkconfig rlogin on service xinetd reload Create the /etc/hosts.equiv file as the root user. touch /etc/hosts.equiv chmod 600 /etc/hosts.equiv chown root:root /etc/hosts.equiv Edit the /etc/hosts.equiv file to include all the RAC nodes: +rac1 oracle +rac2 oracle +rac1-priv oracle +rac2-priv oracle Login as the oracle user and add the following lines at the end of the .bash_profile file. # Oracle Settings TMP=/tmp; export TMP TMPDIR=$TMP; export TMPDIR ORACLE_BASE=/u01/app/oracle; export ORACLE_BASE ORACLE_HOME=$ORACLE_BASE/product/10.2.0/db_1; export ORACLE_HOME ORACLE_SID=RAC1; export ORACLE_SID ORACLE_TERM=xterm; export ORACLE_TERM PATH=/usr/sbin:$PATH; export PATH PATH=$ORACLE_HOME/bin:$PATH; export PATH LD_LIBRARY_PATH=$ORACLE_HOME/lib:/lib:/usr/lib; export LD_LIBRARY_PATH

CLASSPATH=$ORACLE_HOME/JRE:$ORACLE_HOME/jlib:$ORACLE_HOME/rdbms/jlib; export CLASSPATH if [ $USER = "oracle" ]; then if [ $SHELL = "/bin/ksh" ]; then ulimit -p 16384 ulimit -n 65536 else ulimit -u 16384 -n 65536 fi fi

5. Install VMware Client Tools VMware client tools are now installed.
Login as the root user on the rac1 virtual machine, then select the "VM --> Install VMware Tools..." option from the main VMware Server Console menu.

This should mount a virtual CD containing the VMware Tools software. Double-click on the CD icon labeled "VMware Tools" to open the CD. Right-click on the ".rpm" package and select the "Open with 'Install Packages'" menu option.

Click the "Continue" button on the "Completed System Preparation" screen and wait for the installation to complete.

Once the package is loaded, the CD should unmount automatically. You must then run the "vmware-config-tools.pl" script as the root user. The following listing is an example of the output you should expect. # vmware-config-tools.pl Stopping VMware Tools services in the virtual machine: Guest operating system daemon: [ OK ] Trying to find a suitable vmhgfs module for your running kernel. The module bld-2.6.9-11.EL-i686up-RHEL4 loads perfectly in the running kernel. pcnet32 30409 0 Unloading pcnet32 module Trying to find a suitable vmxnet module for your running kernel. The module bld-2.6.9-11.EL-i686up-RHEL4 loads perfectly in the running kernel.

Detected X.org version 6.8.

Please choose one of the following display sizes (1 - 13): [1] "640x480" [2] "800x600" [3] "1024x768" [4] "1152x864" [5] "1280x800" [6] "1152x900" [7] "1280x1024" [8] "1376x1032" [9] "1400x1050" [10] "1680x1050" [11] "1600x1200" [12]< "1920x1200" [13] "2364x1773" Please enter a number between 1 and 13: [12] 3

X Window System Version 6.8.2 Release Date: 9 February 2005 X Protocol Version 11, Revision 0, Release 6.8.2 Build Operating System: Linux 2.6.9-11.EL i686 [ELF] Current Operating System: Linux rac1.localdomain 2.6.9-22.EL #1 Sat Oct 8 17:48:27 CDT 2005 i686 Build Date: 07 October 2005 Build Host: x8664-build.home.local Before reporting problems, check http://wiki.X.Org to make sure that you have the latest version. Module Loader present

OS Kernel: Linux version 2.6.9-22.EL (buildcentos@louisa.home.local) (gcc version 3.4.4 20050721 (Red Hat 3.4.4-2)) #1 Sat Oct 8 17:48:27 CDT 2005 P Markers: (--) probed, (**) from config file, (==) default setting, (++) from command line, (!!) notice, (II) informational, (WW) warning, (EE) error, (NI) not implemented, (??) unknown. (++) Log file: "/tmp/vmware-config0/XF86ConfigLog.3674", Time: Thu Apr 13 21:17:37 2006 (++) Using config file: "/tmp/vmware-config0/XF86Config.3674" X is running fine with the new config file. (WW) VMWARE(0): Failed to set up write-combining range (0xf0000000,0x1000000) Starting VMware Tools services in the virtual machine: Switching to guest configuration: [ OK ] Guest filesystem driver: [ OK ] Guest vmxnet fast network device: [ OK ] DMA setup: [ OK ] Guest operating system daemon: [ OK ] The configuration of VMware Tools e.x.p build-22874 for Linux for this running kernel completed successfully. You must restart your X session before any mouse or graphics changes take effect. You can now run VMware Tools by invoking the following command: "/usr/bin/vmware-toolbox" during an XFree86 session. To use the vmxnet driver, restart networking using the following commands: /etc/init.d/network stop rmmod pcnet32 rmmod vmxnet depmod -a modprobe vmxnet /etc/init.d/network start Enjoy, --the VMware team # The VMware client tools are now installed. 6. Time Synchronization a) As root on rac1 run vmware-toolbox and Select the Time synchrononization between the virtual machine and the host operating system option. This is the sample screen shot of rac2 machine just for demonstaration.

b) Edit the /boot/grub/grub.conf file and enter clock=pit nosmp noapic nolapic to the kernel line.

c) Reboot the machine. Note: Time Zone of the host and guest operating systems should match.

7. Create Shared Disks Shut down the rac1 virtual machine using the following command. # shutdown -h now

Create a directory E:\rac\shared on the host system to hold the shared virtual disks. On the VMware Server Console, click the "Edit virtual machine settings" button. On the "Virtual Machine Settings" screen, click the "Add..." button.

Click the Next button.

Select the hardware type of "Hard Disk" and click the "Next" button.

Accept the "Create a new virtual disk" option by clicking the "Next" button.

Accept the "SCSI" option by clicking the "Next" button.

Set the disk size to "2.0" GB and uncheck the "Allocate all disk space now" option, then click the "Next" button.

Set the disk name to "E:\rac\shared\ocr.vmdk" and click the "Advanced" button.

Set the virtual device node to "SCSI 1:0" and the mode to "Independent" and "Persistent", then click the "Finish" button.

Repeat the previous hard disk creation steps 2 more times, using the following values: # 2.0 GB File Name: E:\rac\shared\votingdisk.vmdk Virtual Device Node: SCSI 1:1

Mode: Independent and Persistent # 30.0 GB File Name: E:\rac\shared\shareddisk.vmdk Virtual Device Node: SCSI 1:2 Mode: Independent and Persistent At the end of this process, the virtual machine should look something like the picture below.

Edit the contents of the "E:\rac\rac1\Red Hat Enterprise Linux 4.vmx" file using a text editor, making sure the following entries are present. Some of the tries will already be present, some will not.

disk.locking = "FALSE" diskLib.dataCacheMaxSize = "0" diskLib.dataCacheMaxReadAheadSize = "0" diskLib.dataCacheMinReadAheadSize = "0" diskLib.dataCachePageSize = "4096" diskLib.maxUnsyncedWrites = "0" scsi1.present = "TRUE" scsi1.virtualDev = "lsilogic" scsi1.sharedBus = "VIRTUAL" scsi1:0.present = "TRUE" scsi1:0.mode = "independent-persistent" scsi1:0.fileName = E:\rac\shared\ocr.vmdk" scsi1:0.deviceType = "plainDisk" scsi1:0.redo = "" scsi1:1.present = "TRUE" scsi1:1.mode = "independent-persistent" scsi1:1.fileName = E:\rac\shared\votingdisk.vmdk" scsi1:1.deviceType = "plainDisk" scsi1:1.redo = "" scsi1:2.present = "TRUE" scsi1:2.mode = "independent-persistent" scsi1:2.fileName = E:\rac\shared\shareddisk.vmdk" scsi1:2.deviceType = "plainDisk" scsi1:2.redo = "" Start the rac1 virtual machine by clicking the "Start this virtual machine" button on the VMware Server Console. When the server has started, log in as the root user so you can partition the disks. The current disks can be seen by issueing the following commands. # cd /dev # ls sd* sda sda1 sda2 sdb sdc sdd Use the "fdisk" command to partition the disks sdb to sdd. The following output shows the expected fdisk output for the sdb disk. # fdisk /dev/sdb Device contains neither a valid DOS partition table, nor Sun, SGI or OSF disklabel Building a new DOS disklabel. Changes will remain in memory only, until you decide to write them. After that, of course, the previous content won't be recoverable. The number of cylinders for this disk is set to 1305. There is nothing wrong with that, but this is larger than 1024, and could in certain setups cause problems with: 1) software that runs at boot time (e.g., old versions of LILO) 2) booting and partitioning software from other OSs (e.g., DOS FDISK, OS/2 FDISK) Warning: invalid flag 0x0000 of partition table 4 will be corrected by w(rite) Command (m for help): n Command action

e extended p primary partition (1-4) p Partition number (1-4): 1 First cylinder (1-1305, default 1): Using default value 1 Last cylinder or +size or +sizeM or +sizeK (1-1305, default 1305): Using default value 1305 Command (m for help): p Disk /dev/sdb: 10.7 GB, 10737418240 bytes 255 heads, 63 sectors/track, 1305 cylinders Units = cylinders of 16065 * 512 = 8225280 bytes Device Boot /dev/sdb1 Start 1 End Blocks Id System 1305 10482381 83 Linux

Command (m for help): w The partition table has been altered! Calling ioctl() to re-read partition table. Syncing disks. # In each case, the sequence of answers is "n", "p", "1", "Return", "Return", "p" and "w". Once all the disks are partitioned, the results can be seen by repeating the previous "ls" command. # cd /dev # ls sd* sda sda1 sda2 sdb sdb1 sdc sdc1 sdd sdd1 Edit the /etc/sysconfig/rawdevices file, adding the following lines. /dev/raw/raw1 /dev/sdb1 /dev/raw/raw2 /dev/sdc1 Restart the rawdevices service using the following command. service rawdevices restart Run the following commands: chown root:oinstall /dev/raw/raw1 chmod 640 /dev/raw/raw1 chown oracle:oinstall /dev/raw/raw2 chown 640 /dev/raw/raw2 Create the file /etc/udev/permissions.d/49-oracle.permissions and add the following lines to it: # OCR raw/raw1:root:oinstall:0640 # Voting Disks raw/raw2:oracle:oinstall:0640

8. Clone the Virtual Machine Shut down the rac1 virtual machine using the following command. # shutdown -h now Copy the contents of the rac1 virtual machine into "E:\rac\rac2". Edit the contents of the "E:\rac\rac2\Red Hat Enterprise Linux 4.vmx" file, making the following change. displayName = "rac2" In the VMware Server Console, select the File > Open menu options and browse for the "E:\rac\rac2\Red Hat Enterprise Linux 4.vmx" file. Once opened, the rac2 virtual machine is visible on the console. Start the rac2 virtual machine by clicking the "Start this virtual machine" button and click the "Always Create" button on the subsequent "Question" screen. Ignore any errors during the server startup. We are expecting the networking components to fail at this point. Log in to the rac2 virtual machine as the root user and start the "Network Configuration" tool (Applications > System Settings > Network). Highlight the "eth0" interface and click the "Edit" button on the toolbar and alter the IP address to "192.168.2.102" in the resulting screen. Click on the "Hardware Device" tab and click the "Probe" button. Then accept the changes by clicking the "OK" button. Repeat the process for the "eth1" interface, this time setting the IP Address to "192.168.0.102". Click on the "DNS" tab and change the host name to "rac2", then click on the "Devices" tab. Once you have finished, save the changes (File > Save) and activate the network interfaces by highlighting them and clicking the "Activate" button. 9. Install and Configure Oracle Cluster File System (OCFS2) 1) Install OCFS2 --> both nodes I will install the OCFS2 rpms onto two rac nodes. The installation process is simply a matter of running the following command on both Oracle RAC nodes in the cluster as the root user: # rpm -ivh ocfs2-2.6.9-55.ELsmp-1.2.7-1.el4.i686.rpm \ ocfs2console-1.2.7-1.el4.i386.rpm \ ocfs2-tools-1.2.7-1.el4.i386.rpm Preparing... ########################################### [100%] 1:ocfs2-tools ########################################### [ 33%] 2:ocfs2-2.6.9-55.ELsmp ########################################### [ 67%] 3:ocfs2console ########################################### [100%] 2)Disable SELinux (RHEL4 U2 and higher) --> both nodes

a) /usr/bin/system-config-securitylevel b) Now, click the SELinux tab and check off the "Enabled" checkbox. After clicking [OK], you will be presented with a warning dialog. Simply acknowledge this warning by clicking "Yes". c) After making this change on both nodes in the cluster, each node will need to be rebooted to implement the change 3) Configure OCFS2 --> both nodes This will need to be done on both Oracle RAC nodes in the cluster as the root user: # ocfs2console Using the ocfs2console GUI tool, perform the following steps: 1)Select [Cluster] -> [Configure Nodes...]. This will start the OCFS2 Cluster Stack . Acknowledge this Information dialog box by clicking [Close]. You will then be presented with the "Node Configuration" dialog. 2)On the "Node Configurtion" dialog, click the [Add] button. This will bring up the "Add Node" dialog. 3)In the "Add Node" dialog, enter the Host name and IP address for the first node in the cluster. Leave the IP Port set to its default value of 7777. In my example, I added both nodes using rac1 / 192.168.0.101 for the first node and rac2 / 192.168.0.102 for the second node. Note: The node name you enter "must" match the hostname of the machine and the IP addresses will use the private interconnect. Click [Apply] on the "Node Configuration" dialog - All nodes should now be "Active" . Click [Close] on the "Node Configuration" dialog. After verifying all values are correct, exit the application using [File] -> [Quit]. This needs to be performed on both Oracle RAC nodes in the cluster 4)After exiting the ocfs2console, you will have a /etc/ocfs2/cluster.conf similar to the following. This process needs to be completed on both Oracle RAC nodes in the cluster and the OCFS2 configuration file should be exactly the same for both of the nodes: /etc/ocfs2/cluster.conf node: ip_port = 7777 ip_address = 192.168.0.101 number = 0 name = rac1 cluster = ocfs2 node: ip_port = 7777 ip_address = 192.168.0.102 number = 1 name = rac2 cluster = ocfs2 cluster:

node_count = 2 name = ocfs2 4) O2CB Cluster Service --> both nodes Before we can do anything with OCFS2 like formatting or mounting the file system, we need to first have OCFS2's cluster stack, O2CB, running (which it will be as a result of the configuration process performed above). The stack includes the following services: NM: Node Manager that keep track of all the nodes in the cluster.conf HB: Heart beat service that issues up/down notifications when nodes join or leave the cluster TCP: Handles communication between the nodes DLM: Distributed lock manager that keeps track of all locks, its owners and status CONFIGFS: User space driven configuration file system mounted at /config DLMFS: User space interface to the kernel space DLM /etc/init.d/o2cb status Module "configfs": Loaded Filesystem "configfs": Mounted Module "ocfs2_nodemanager": Loaded Module "ocfs2_dlm": Loaded Module "ocfs2_dlmfs": Loaded Filesystem "ocfs2_dlmfs": Mounted Checking O2CB cluster ocfs2: Online Heartbeat dead threshold: 31 Network idle timeout: 30000 Network keepalive delay: 2000 Network reconnect delay: 2000 Checking O2CB heartbeat: Not active 5) Configure O2CB to Start on Boot and Adjust O2CB Heartbeat Threshold --> both nodes All of the tasks within this section will need to be performed on both nodes in the cluster. Set the on-boot properties as follows: # /etc/init.d/o2cb offline ocfs2 # /etc/init.d/o2cb unload # /etc/init.d/o2cb configure Configuring the O2CB driver. This will configure the on-boot properties of the O2CB driver. The following questions will determine whether the driver is loaded on boot. The current values will be shown in brackets ('[]'). Hitting <ENTER> without typing an answer will keep that current value. Ctrl-C will abort. Load O2CB driver on boot (y/n) [n]: y Cluster to start on boot (Enter "none" to clear) [ocfs2]: ocfs2 Specify heartbeat dead threshold (>=7) [31]: 61 Specify network idle timeout in ms (>=5000) [30000]: 30000 Specify network keepalive delay in ms (>=1000) [2000]: 2000 Specify network reconnect delay in ms (>=2000) [2000]: 2000 Writing O2CB configuration: OK Loading module "configfs": OK

Mounting configfs filesystem at /config: OK Loading module "ocfs2_nodemanager": OK Loading module "ocfs2_dlm": OK Loading module "ocfs2_dlmfs": OK Mounting ocfs2_dlmfs filesystem at /dlm: OK Starting O2CB cluster ocfs2: OK Now activate it # /etc/init.d/o2cb load # /etc/init.d/o2cb online ocfs2 6) Format the OCFS2 File System --> 1-Node (rac1) a) Unlike the other tasks in this section, creating the OCFS2 file system should only be executed on one of nodes in the RAC cluster. I will be executing all commands in this section from rac1 only. b) If the O2CB cluster is offline, start it. The format operation needs the cluster to be online, as it needs to ensure that the volume is not mounted on some node in the cluster. # /etc/init.d/o2cb load # /etc/init.d/o2cb online ocfs2 c) # mkfs.ocfs2 -b 4K -C 256K -N 4 -L dbfiles /dev/sdd1

7)Mount the OCFS2 File System --> both nodes Mounting the file system will need to be performed on both nodes in the Oracle RAC cluster as the root user account using the OCFS2 label dbfiles! First, here is how to manually mount the OCFS2 file system from the command-line. Remember that this needs to be performed as the root user account: # mount -t ocfs2 -o datavolume,nointr -L "dbfiles" /u02 If the mount was successful, you will simply get your prompt back. We should, however, run the following checks to ensure the file system is mounted correctly. Let's use the mount command to ensure that the new file system is really mounted. This should be performed on both nodes in the RAC cluster: # mount /dev/mapper/VolGroup00-LogVol00 on / type ext3 (rw) none on /proc type proc (rw) none on /sys type sysfs (rw) none on /dev/pts type devpts (rw,gid=5,mode=620) usbfs on /proc/bus/usb type usbfs (rw) /dev/hda1 on /boot type ext3 (rw) none on /dev/shm type tmpfs (rw) none on /proc/sys/fs/binfmt_misc type binfmt_misc (rw) sunrpc on /var/lib/nfs/rpc_pipefs type rpc_pipefs (rw) cartman:SHARE2 on /cartman type nfs (rw,addr=192.168.1.120) configfs on /config type configfs (rw)

ocfs2_dlmfs on /dlm type ocfs2_dlmfs (rw) /dev/sdd1 on /u02 type ocfs2 (rw,_netdev,datavolume,nointr,heartbeat=local) 8)Configure OCFS2 to Mount Automatically at Startup --> both nodes We start by adding the following line to the vi --> /etc/fstab file on both nodes in the RAC cluster: LABEL=dbfiles /u02 ocfs2 _netdev,datavolume,nointr 00

9)Check Permissions on New OCFS2 File System --> both nodes

Use the ls command to check ownership. The permissions should be set to 0775 with owner "oracle" and group "oinstall". Let's first check the permissions: # ls -ld /u02 drwxr-xr-x 3 root root 4096 Sep 3 00:42 /u02 As we can see from the listing above, the oracle user account (and the oinstall group) will not be able to write to this directory. Let's fix that: # chown oracle:oinstall /u02 # chmod 775 /u02 Let's now go back and re-check that the permissions are correct for both Oracle RAC nodes in the cluster: # ls -ld /u02 drwxrwxr-x 3 oracle oinstall 4096 Sep 3 00:42 /u02 10)Create Directory for Oracle Clusterware Files --> 1 node (rac1) The following tasks only need to be executed on one of nodes in the RAC cluster. I will be executing all commands in this section from rac1 only. # mkdir -p /u02/oradata # chown -R oracle:oinstall /u02/oradata # chmod -R 775 /u02/oradata # ls -l /u02/oradata total 4 drwxrwxr-x 2 oracle oinstall 4096 Sep 3 00:45 orcl 11)Reboot Both Nodes --> both nodes Before starting the next section, this would be a good place to reboot both of the nodes in the RAC cluster. When the machines come up, ensure that the cluster stack services are being loaded and the new OCFS2

file system is being mounted: # mount /dev/mapper/VolGroup00-LogVol00 on / type ext3 (rw) none on /proc type proc (rw) none on /sys type sysfs (rw) none on /dev/pts type devpts (rw,gid=5,mode=620) usbfs on /proc/bus/usb type usbfs (rw) /dev/hda1 on /boot type ext3 (rw) none on /dev/shm type tmpfs (rw) none on /proc/sys/fs/binfmt_misc type binfmt_misc (rw) sunrpc on /var/lib/nfs/rpc_pipefs type rpc_pipefs (rw) cartman:SHARE2 on /cartman type nfs (rw,addr=192.168.1.120) configfs on /config type configfs (rw) ocfs2_dlmfs on /dlm type ocfs2_dlmfs (rw) /dev/sdd1 on /u02 type ocfs2 (rw,_netdev,datavolume,nointr,heartbeat=local) If you modified the O2CB heartbeat threshold, you should verify that it is set correctly: # cat /proc/fs/ocfs2_nodemanager/hb_dead_threshold 61 12)How to Determine OCFS2 Version --> both nodes To determine which version of OCFS2 is running, use: # cat /proc/fs/ocfs2/version OCFS2 1.2.7 Tue Oct 9 16:15:42 PDT 2007 (build d443ce77532cea8d1e167ab2de51b8c8) The shared disks are now configured. Edit the /home/oracle/.bash_profile file on the rac2 node to correct the ORACLE_SID value. ORACLE_SID=RAC2; export ORACLE_SID Start the rac1 virtual machine and restart the rac2 virtual machine. While starting up, the "Kudzu" detection screen may be displayed. Press a key and accept the configuration change on the following screen. When both nodes have started, check they can both ping all the public and private IP addresses using the following commands. ping -c 3 rac1 ping -c 3 rac1-priv ping -c 3 rac2 ping -c 3 rac2-priv At this point the virtual IP addresses defined in the /etc/hosts file will not work, so don't bother testing them. It is a good idea to make a consistent backup of this virtual environment. Shutdown both the RAC nodes and compress the main rac folder in E drive. The virtual machine setup is

now complete. Note: You can also configure ocfs2 on one node before cloning the virtual machine.
10. Oracle Clusterware and DB Installation

OCR home: /u01/crs/oracle/product/10.2.0/crs OCR Location: /dev/raw/raw1 Voting Disk Location:/dev/raw/raw2 Oracle Software Home: /u01/app/oracle/product/10.2.0/db_1 Database Files location: /u02/oradata

You might also like