Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x


Oracle Solutions Engineering Wes Vaske Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x This document providesinstructions for deploying Oracle Database on Red Hat Enterprise Linux 6 or Oracle Linux 6 Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x ii This document is for informational purposes only and may contain typographical errors and technical inaccuracies. The content is provided as is, without express or implied warranties of any kind. © 2012 Dell Inc. All rights reserved. Dell and its affiliates cannot be responsible for errors or omissions in typography or photography. Dell, the Dell logo, and PowerEdge are trademarks of Dell Inc. Intel and Xeon are registered trademarks of Intel Corporation in the U.S. and other countries. Microsoft, Windows, and Windows Server are either trademarks or registered trademarks of Microsoft Corporation in the United States and/or other countries. Other trademarks and trade names may be used in this document to refer to either the entities claiming the marks and names or their products. Dell disclaims proprietary interest in the marks and names of others. January 2012| Version 1.0 April 2013| Version 1.1 Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x iii Contents Overview ................................................................................................................. 5 Hardware and software requirements .............................................................................. 5 Hardware requirements .............................................................................................. 5 Memory requirements ............................................................................................... 5 Disk space requirements ............................................................................................ 5 Network Requirements .............................................................................................. 6 Operating System Requirements .................................................................................... 6 Preparing Nodes for Oracle Installation ............................................................................. 6 Attaching to RHN/ULN Repository .................................................................................. 6 Installing the Dell Oracle Preinstall RPM .......................................................................... 8 Installing the Dell Oracle Utilities RPM ............................................................................ 8 Oracle Software Binary Location .................................................................................... 9 Setting up the Network ............................................................................................... 9 Public Network ....................................................................................................... 9 Private Network .................................................................................................... 10 IP Address and Name Resolution Requirements ................................................................ 12 For a Cluster Using DNS ........................................................................................... 12 Preparing Shared Storage for Oracle RAC Installation ........................................................ 13 Setting up Device Mapper Multipath ............................................................................ 14 Partitioning the Shared Disk ..................................................................................... 16 Identifying ASM Disks to Setup Udev Rules **Applies to RHEL 6 and Oracle Linux 6 (if not using ASMlib)** ............................................................................................................. 17 Installing and Configuring ASMLib **Applies ONLY to Oracle Linux 6 and if not setting udev rules** ........................................................................................................................ 18 Using ASMLib to Mark the Shared Disks as Candidate Disks **Applies ONLY to Oracle Linux 6 and if not setting udev rules** ........................................................................................... 18 Installing Oracle 11g R2 Grid Infrastructure ..................................................................... 19 Before You Begin .................................................................................................... 19 Configure the System Clock Settings for All Nodes ............................................................ 19 Configuring Node One ............................................................................................... 20 Configuring Shared Oracle Home using ACFS..................................................................... 28 Installing Oracle 11g R2 Database (RDBMS) Software .......................................................... 31 Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x iv Creating Diskgroup Using ASM Configuration Assistant (ASMCA) ............................................ 37 Creating Database Using DBCA .................................................................................... 39 Tables Table 1. Swap Space Requirements ................................................................................ 5 Table 2. Disk Space Requirements ................................................................................. 5 Table 3. Cluster Requirements for DNS ......................................................................... 12 Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 5 Overview This document applies to Oracle Database 11g R2 running on Red Hat Enterprise Linux 6.x AS x86_64 or Oracle Enterprise Linux 6.x AS x86_64. Hardware and software requirements Hardware requirements For Dell supported hardware configurations, see the Software Deliverable List (SDL) for each Dell Validated Component at the Current Release Page: en.community.dell.com/techcenter/enterprise- solutions/w/oracle_solutions/1429.current-release.aspx. Memory requirements 1. Oracle requires 1.5 gigabytes (GB) of physical memory. 2. Swap space is proportional to the amount of RAM allocated to the system as indicated in the table below. Swap Space Requirements Table 1. Ram Swap Space Up to 2GB 1.5 times the size of RAM Up to 16GB Equal to the size of RAM 16GB or more 16GB 3. Oracle's temporary space (/tmp) must be at least 1 GB in size. 4. A monitor that supports resolution of 1024 x 768 to correctly display the Oracle Universal Installer (OUI) Disk space requirements The following table describes the disk space required for an Oracle Database installation: Disk Space Requirements Table 2. Software Installation Location Size Required Grid Infrastructure Home 4.5GB Oracle Database Home 4 GB Shared Storage Disk Database size + Flashback Recovery Area (FRA) Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 6 Network Requirements 1. It is recommended that each node contain at least three network interface cards (NICs). One NIC for public network and two NICs for private network to ensure high availability of the Oracle RAC cluster. 2. Public and private interface names must be the same on all nodes. For example, if em1 is used as the public interface on node one, then all other nodes require em1 as the public interface. 3. All public interfaces for each node should be able to communicate with all nodes within the cluster. 4. All private interfaces for each node should be able to communicate with all nodes within the cluster. 5. The hostname of each node must follow the RFC 952 standard (ietf.org/rfc/rfc952.txt). Hostnames that include an underscore ("_") are not permitted. 6. Each node in the cluster requires the following IP address: o One public IP address o Two private IP address o One virtual IP address o Three Single Client Access Name (SCAN) addresses for the cluster Operating System Requirements • Red Hat Enterprise Linux 6.x AS x86_64 (Kernel 2.6.32-71 or higher) • Oracle Linux 6.x AS x86_64 (Kernel 2.6.32-71 or higher) Preparing Nodes for Oracle Installation Attaching to RHN/ULN Repository NOTE: The documentation provided below discusses how to setup a local yum repository using your operating system installation media. If you would like to connect to the RHN/ULN channels, see the appropriate documentation. For Red Hat see, redhat.com/red_hat_network. For information relating to ULN network see, linux.oracle.com. The recommended configuration is to serve the files over http using an Apache server (package name: httpd). This section discusses hosting the repository files from a local filesystem storage. While other options to host repository files exist, they are outside of the scope of this document. It is highly recommended to use local filesystem storage for speed and simplicity of maintenance. 1. One of the requirements is to have the DVD image mounted either by physical media or by ISO image. a. To mount the DVD, insert the DVD into the server and it should auto-mount into the /media directory. b. To mount an ISO image we will need to run the following command as root, substituting the path name of your ISO image for the field myISO.iso: Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 7 [ root@localhost ~]# mkdir /media/myISO [ root@localhost ~]# mount -o loop myISO.iso /media/myISO 2. To install and configure the http daemon, configure the machine that will host the repository for all other machines to use the DVD image locally. Create the file /etc/yum.repos.d/local.repo and enter the following: [local] name=Local Repository baseurl=file:///media/myISO/Server gpgcheck=0 enabled=0 3. Now we will install the Apache service daemon with the following command which will also temporarily enable the local repository for dependency resolution: [ root@localhost ~]# yum -y install httpd --enablerepo=local 4. After the Apache service daemon is installed, start the service and set it to start up for us next time we reboot. Run the following commands as root: [ root@localhost ~]# service httpd start [ root@localhost ~]# chkconfig httpd on 5. To use Apache to serve out the repository, copy the contents of the DVD into a published web directory. Run the following commands as root (make sure to switch myISO with the name of your ISO)commands: [ root@localhost ~]# mkdir /var/www/html/myISO [ root@localhost ~]# cp -R /media/myISO/* /var/www/html/myISO a. This step is only necessary if you are running SELinux on the server that hosts the repository. The following command should be run as root and will restore the appropriate SELinux context to the copied files: [ root@localhost ~]# restorecon -Rvv /var/www/html/. 6. The final step is to gather the DNS name or IP of the server that is hosting the repository. The DNS name or IP of the hosting server will be used to configure your yum repository repo file on the client server. The following is the listing of an example configuration using the RHEL 6.x Server media and is held in the configuration file: /etc/yum.repos.d/myRepo.repo [myRepo] name=RHEL6.2 DVD baseurl= http://reposerver.mydomain.com/RHEL6_2/Server enabled=1 gpgcheck=0 NOTE: Replace reposerver.mydomain.com with your server's DNS name or IP address. NOTE: You can also place the configuration file on the server hosting the repository for all other servers such that it can also use the repository as a more permanent solution to what was done in step 2 . Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 8 Installing the Dell Oracle Preinstall RPM Once your nodes have attached to the appropriate yum repository, we will need to install the Dell- Oracle-RDBMS-Server-11gR2-Preinstall RPM package. The Dell-Oracle-RDBMS-Server-11gR2- Preinstall RPM package automates certain pieces of the installation process required for the installation of Oracle RAC or Oracle single instance. The process to install the Dell Oracle Preinstall RPM package is as follows: 1. Download the latest Dell Oracle Deployment tar file from en.community.dell.com/techcenter/enterprise-solutions/w/oracle_solutions/1429.current-release.aspx NOTE: The filename will follow the convention: Dell-Oracle-Deployment-OS-version-year- month.tar, for example: Dell-Oracle-Deployment-Lin-EL6-2012-08.tar 2. Copy the Dell Oracle Deployment tar file to a working directory of all your cluster nodes. 3. To go to your working directory, enter the following command: [ root@localhost ~]# cd 4. Untar the Dell-Oracle-Deployment release using the command: [ root@localhost ~]# tar zxvf Dell-Oracle-Deployment-o-v-y-m.tar.gz NOTE: Where, o is the operating system, v is the operating system version, y is the year, and m is the month of the tar release. 5. Change directory to Dell-Oracle-Deployment-o-v-y-m 6. Install the Dell-Oracle-RDBMS-Server-11gR2-Preinstall RPM package on all your cluster nodes using the following command: [ root@localhost ~]# yum localinstall dell-oracle-rdbms-server-11gR2- preinstall* --nogpgcheck Installing the Dell Oracle Utilities RPM The Dell Oracle utilities RPM is designed to do the following Dell and Oracle recommended settings: • Create Grid Infrastructure directories, set ownership, and permissions. • Create grid user. • Create Oracle Database (RDBMS) directories, set ownership, and permissions. • Create the Oracle base directories, set ownership, and permissions. • Set pam limits within /etc/pam.d/login • Setup /etc/profile. • Set SELinux to Disabled. • Install the Dell PowerEdge system component drivers if applicable. • Set kernel parameters. • Set nproc for grid user within /etc/security/limits.conf To install the RPM: Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 9 1. Download the latest Dell Oracle Deployment tar file from en.community.dell.com/techcenter/enterprise-solutions/w/oracle_solutions/1429.current-release.aspx NOTE: The filename will follow the convention: Dell-Oracle-Deployment-OS version-year- month.tar.gz, for example: Dell-Oracle-Deployment-Lin-2011-07.tar.gz 2. Copy the Dell Oracle Deployment tar file to a working directory of all your cluster nodes. 3. Go to your working directory via the command: [ root@localhost ~]# cd 4. Untar the Dell-Oracle-Deployment release using the command: [ root@localhost ~]# tar zxvf Dell-Oracle-Deployment-o-y-m.tar.gz NOTE: Where, o is the operating system version, y is the year, and m is the month of the tar release. 5. Change directory to Dell-Oracle-Deployment-o-y-m 6. Install the Dell oracle utilities RPM package on all your cluster nodes the by typing: [ root@localhost ~]# yum localinstall dell-oracle-utilities* --nogpgcheck 7. Once the rpm is installed, run the dodeploy script to setup the environment as follows: [ root@localhost ~]# dodeploy -g -r 11gR2 For more information about the Dell oracle utilities RPM and its options, check the man pages using the command: [ root@localhost ~]# man 8 dodeploy NOTE: The Dell-Oracle-Deplyoment tar contains the latest supported drivers provided from our Software Deliverable List (SDL). Consult the README file found within the Dell-Oracle-Deployment tar for installation instructions of the latest drivers. Oracle Software Binary Location The Oracle software binaries should be located on node 1 of your cluster. It is important to note that starting with Oracle 11g R2 (11.2.0.3), Oracle Database patch sets are full installation of the Oracle software. For more information on how this impacts future Oracle deployments, see My Oracle Support note: 1189783.1 Important Changes to Oracle Database Patch Sets Starting with 11.2.0.2. Setting up the Network Public Network NOTE: Ensure that the public IP address is a valid and routable IP address. NOTE: Ensure to disable NetworkManager via commands: [ root@localhost ~]# service NetworkManager stop [ root@localhost ~]# chkconfig NetworkManager off To configure the public network on each node: Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 10 1. Log in as root. 2. Edit the network device file /etc/sysconfig/network-scripts/ifcfg-em# where # is the number of the network device: NOTE: Ensure that the Gateway address is configured for the public network interface. If the Gateway address is not configured, the Oracle Grid installation may fail. DEVICE=em1 ONBOOT=yes NM_CONTROLLED=no IPADDR= NETMASK= BOOTPROTO=static HWADDR= SLAVE=no GATEWAY= 3. Edit the /etc/sysconfig/network file, and, if necessary, replace localhost.localdomain with the qualified public node name. For example, the command for node 1: hostname=node1.domain.com 4. Restart the network service. [ root@localhost ~]# service network restart 5. Type ifconfig to verify that the IP addresses are set correctly. 6. To check your network configuration, ping each public IP address from a client on the LAN that is not a part of the cluster. 7. Connect to each node via SSH to verify that the public network is functioning [ root@localhost ~]# ssh Private Network NOTE: Each of the two NIC ports for the private network must be on separate PCIe buses. The grid infrastructure of Oracle 11gR2 (11.2.0.3) supports IP failover natively using a new feature known as Redundant Interconnect. Oracle uses its ora.cluster_interconnect.haip resource to communicate with Oracle RAC, Oracle ASM, and other related services. The Highly Available Internet Protocol (HAIP) has the ability to activate a maximum of four private interconnect connections. These private network adapters can be configured during the initial install process of Oracle Grid or after the installation process using the oifcfg utility. Oracle Grid currently creates an alias IP (also known as virtual private IP) on your private network adapters using the 169.254.*.* subnet for the HAIP. If subnet range is already in use, Oracle Grid will not attempt to use it. The purpose of HAIP is to load balance across all active interconnect interfaces, and failover to other available interfaces if one of the existing private adapters becomes unresponsive. NOTE: Configure a different subnet for each of the private networks you want to configure as a part of HAIP. When adding additional HAIP addresses after the installation of Oracle Grid (maximum of four), restart your Oracle Grid environment to make these new HAIP addresses active. Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 11 The example below provides step-by-step instructions on enabling redundant interconnect using HAIP on a fresh Oracle 11gR2 (11.2.0.3) Grid Infrastructure installation. 1. Edit the file, /etc/sysconfig/network-scripts/ifcfg-emX, where X is the number of the em device. ifcfg-emX are the configuration files of the network adapters to be used for your private interconnect. The following example shows em2 using 192.168.1.* subnet and em3 using 192.168.2.* . For config file: /etc/sysconfig/network-scripts/ifcfg-em2 DEVICE=em2 BOOTPROTO=static HWADDR= ONBOOT=yes NM_CONTROLLED=no IPADDR=192.168.1.140 NETMASK=255.255.255.0 For config file: /etc/sysconfig/network-scripts/ifcfg-em3 DEVICE=em3 HWADDR= BOOTPROTO=static ONBOOT=yes NM_CONTROLLED=no IPADDR=192.168.2.140 NETMASK=255.255.255.0 2. Once you have saved both the configuration files, restart your network service using service network restart. The completion of the steps above has now prepared your system to enable HAIP using the Oracle Grid Infrastructure installer. When you have completed all the Oracle prerequisites and are ready to install Oracle, you will need to select em2 and em3 as 'private' interfaces at the 'Network Interface Usage' screen. This step enables redundant interconnectivity once your Oracle Grid Infrastructure has successfully completed and is running. To verify that your redundant interconnect using HAIP is running, you can test this feature using the ifconfig command. An example of the output is listed below. ifconfig em2 Link encap:Ethernet HWaddr 00:24:E8:6B:E9:5A inet addr:192.168.11.52 Bcast:192.168.11.255 Mask:255.255.255.0 inet6 addr: fe80::224:e8ff:fe6b:e95a/64 Scope:Link UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 RX packets:14140 errors:0 dropped:0 overruns:0 frame:0 TX packets:839 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:1000 RX bytes:2251664 (2.1 MiB) TX bytes:56379 (55.0 KiB) Interrupt:48 Memory:d8000000-d8012800 em2:1 Link encap:Ethernet HWaddr 00:24:E8:6B:E9:5A inet addr:169.254.167.163 Bcast:169.254.255.255 Mask:255.255.0.0 UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 em3 Link encap:Ethernet HWaddr 00:24:E8:6B:E9:5C Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 12 inet addr:192.168.11.53 Bcast:192.168.11.255 Mask:255.255.255.0 inet6 addr: fe80::224:e8ff:fe6b:e95c/64 Scope:Link UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 RX packets:12768 errors:0 dropped:0 overruns:0 frame:0 TX packets:2211 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:1000 RX bytes:2163856 (2.0 MiB) TX bytes:144187 (140.8 KiB) Interrupt:32 Memory:da000000-da012800 em3:1 Link encap:Ethernet HWaddr 00:24:E8:6B:E9:5C inet addr:169.254.167.163 Bcast:169.254.255.255 Mask:255.255.0.0 UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 lo Link encap:Local Loopback inet addr:127.0.0.1 Mask:255.0.0.0 inet6 addr: ::1/128 Scope:Host UP LOOPBACK RUNNING MTU:16436 Metric:1 RX packets:28662 errors:0 dropped:0 overruns:0 frame:0 TX packets:28662 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:0 RX bytes:1604049 (1.5 MiB) TX bytes:1604049 (1.5 MiB) For more information on Redundant Interconnect and ora.cluster_interconnect.haip, see metalink note: 1210883.1. IP Address and Name Resolution Requirements The steps below show how to setup your cluster nodes for using Domain Name System (DNS). For information on how to setup cluster nodes using GNS, see the wiki article en.community.dell.com/dell-groups/enterprise_solutions/w/oracle_solutions/1416.aspx For a Cluster Using DNS To set up an Oracle 11g R2 RAC using Oracle DNS (without GNS): 1. At least two interfaces must be configured on each node, one for the private IP address and one for the public IP address. 2. A SCAN NAME configured on the DNS for Round Robin resolution to three addresses (recommended) or at least one address. The SCAN addresses must be on the same subnet as virtual IP addresses and public IP addresses. NOTE: For high availability and scalability, it is recommended that you configure the SCAN to use Round Robin resolution to three IP addresses. The name for the SCAN cannot begin with a numeral. For installation to succeed, the SCAN must resolve to at least one address. The table below describes the different interfaces, IP address settings and the resolutions in a cluster. Cluster Requirements for DNS Table 3. Interface Type Resolution Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 13 Interface Type Resolution Public Static DNS Private Static Not Required Node virtual IP Static DNS SCAN virtual IP Static DNS Configuring a DNS Server To configure changes on a DNS server for an Oracle 11g R2 cluster using a DNS (without GNS): 1. Configure SCAN NAME resolution on DNS server. A SCAN NAME configured on the DNS server using the Round Robin policy should resolve to three public IP addresses (recommended), however the minimum requirement is one public IP address. For example: scancluster IN A 192.0.2.1 IN A 192.0.2.2 IN A 192.0.2.3 Where scancluster is the SCAN NAME provided during Oracle Grid installation. NOTE: The SCAN IP address must be routable and must be in public range. Configuring a DNS Client 1. Configure the resolv.conf on all the nodes in the cluster to contain name server entries that are resolvable to the appropriate DNS server. Provide an entry similar to the following: /etc/resolv.conf: search ns1.domainserver.com nameserver 192.0.2.100 Where, 192.0.2.100 is a valid DNS server address in your network and ns1.domainserver.com is the domain server in your network. 2. Verify the order configuration. /etc/nsswitch.conf controls the name service order. In some configurations, the NIS can cause issues with Oracle SCAN address resolution. It is recommended that you place the NIS entry at the end of the search list and place the dns entry first. For example, hosts: dns files nis 3. Once you have modified the /etc/nsswitch.conf , restart the nscd service by issuing the command: [ root@localhost ~]# /sbin/service nscd restart Preparing Shared Storage for Oracle RAC Installation Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 14 NOTE: In this section, the terms disk(s), volume(s), virtual disk(s), LUN(s) mean the same and are used interchangeably, unless specified otherwise. Similarly, the terms Stripe Element Size and Segment Size both can be used interchangeably. Oracle RAC requires shared LUNs for storing your Oracle Cluster Registry (OCR), voting disks, Oracle Home using ACFS, Oracle Database files, and Flash Recovery Area (FRA). To ensure high availability for Oracle RAC it is recommended that you have: • Three shared volumes/LUNs each of 1 GB in size for normal redundancy or five volumes/LUNs for high redundancy for the Oracle clusterware. • At least two shared disks to store your database. Each shared disk should be the same disk speed and size. • At least two shared volumes/LUNs to store your Automatic Storage Management Cluster File System (ACFS). Each shared disk must be at least 10 GB, for a total size of 20 GB. • At least two shared volumes/LUNs or volumes to store your FRA. Ideally, the FRA space should be large enough to copy all of your Oracle datafiles and incremental backups. For more information on optimally sizing your FRA, see My Oracle Support ID 305648.1 section "What should be the size of Flash Recovery Area?" NOTE: The use of device mapper multipath is recommended for optimal performance and persistent name binding across nodes within the cluster. NOTE: For more information on attaching shared LUNs/volumes, see the Wiki documentation found at: en.community.dell.com/dell-groups/enterprise_solutions/w/oracle_solutions/3- storage.aspx Setting up Device Mapper Multipath The purpose of Device Mapper Multipath is to enable multiple I/O paths to improve performance and provide consistent naming. Multipathing accomplishes this by combining your I/O paths into one device mapper path and properly load balancing the I/O. This section will provide the best practices on how to setup your device mapper multipathing within your Dell PowerEdge server. 1. Verify that your device-mapper and multipath driver are at least the version shown below or higher: [ root@localhost ~]# rpm -qa | grep device-mapper-multipath device-mapper-multipath-0.4.9-46.el6 2. Identify your local disks i.e. /dev/sda. Once your local disk is determined run the following command to obtain its scsi_id: [ root@localhost ~]# scsi_id --page=0x83 --whitelisted --device=/dev/sda 360026b900061855e000007a54ea53534 3. Open the /etc/multipath.conf file and locate and comment out the section below: #blacklist { # devnode “*” #} Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 15 4. Once the scsi_id of your local disk has been retrieved, you must blacklist this scsi_id from being used as a multipath device within /etc/multipath.conf file and locate,uncomment,and modify the section below: blacklist { wwid devnode "^(ram|raw|loop|fd|md|dm-|sr||scd|st)[0-9]*" devnode "^hd[a-z]" } 5. Uncomment your defaults section within your /etc/multipath.conf: defaults { udev_dir /dev polling_interval 10 selector "round-robin 0" path_grouping_policy multibus getuid_callout "/sbin/scsi_id -g -u -s/block/%n" prio_callout /bin/true path_checker readsector0 rr_min_io 100 max_fds 8192 rr_weight priorities failback immediate no_path_retry fail user_friendly_names yes } 6. Locate the multipaths section within your /etc/multipath.conf file. In this section you will provide the scsi_id of each LUN/volume and provide an alias in order to keep a consistent naming convention across all of your nodes. An example is shown below: multipaths { multipath { wwid alias alias_of_volume1 } multipath { wwid alias alias_of_volume2 } } 7. Restart your multipath daemon service using: [ root@localhost ~]# service multipathd restart 8. Verify that your multipath volumes alias are displayed properly: [ root@localhost ~]# multipath -ll 9. Make sure iSCSI service starts upon boot using the command: [ root@localhost ~]# chkconfig multipathd on Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 16 10. Repeat steps 1-9 for all nodes. Partitioning the Shared Disk This section describes how to use Linux’s native partition utility fdisk to create a single partition on a volume/virtual disk that spans the entire disk. To use the fdisk utility to create a partition: 1. At the command prompt, type one of the following: [ root@localhost ~]# fdisk –cu /dev/ [ root@localhost ~]# fdisk –cu /dev/mapper/ Where, is the name of the block device that you are creating a partition on. For example, if the block device is /dev/sdb, type: [ root@localhost ~]# fdisk –cu /dev/sdb Device contains neither a valid DOS partition table, nor Sun, SGI or OSF disklabel Building a new DOS disklabel with disk identifier 0x89058e03. Changes will remain in memory only, until you decide to write them. After that, of course, the previous content won't be recoverable. Warning: invalid flag 0x0000 of partition table 4 will be corrected by w(rite) The device presents a logical sector size that is smaller than the physical sector size. Aligning to a physical sector (or optimal I/O) size boundary is recommended, or performance may be impacted. 2. Enter the following commands: Command (m for help): n #”n” to create new partition Command action e extended p primary partition (1-4): p #”p” for primary parition Partition number (1-4): 1 #”1” for first partition First sector (4096-xxxxxx, default 4096): Using default value 1 Last cylinder or +sizeM or +sizeK (1-###, default ###): 3. Repeat steps 1-2 for all the disks. 4. Type the following to re-read the partition table to see the newly created partition(s) [ root@localhost ~]# partprobe or [ root@localhost ~]# service multipathd restart or [ root@localhost ~]# kpartx –a /dev/mapper/ 5. Reboot the system if your newly created partition is not displayed properly. Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 17 Identifying ASM Disks to Setup Udev Rules **Applies to RHEL 6 and Oracle Linux 6 (if not using ASMlib)** Red Hat Enterprise Linux 6/Oracle Linux 6 has the ability to use udev rules to ensure that the system properly manages permissions of device nodes. In this case, we are referring to properly setting permissions for our LUNs/volumes discovered by the OS. It is important to note that udev rules are executed in enumerated order. When creating udev rules for setting permissions, please include the prefix 20- and append .rules to the end of the filename. An example file name is 20- dell_oracle.rules In order to set udev rules, you need capture the WWIDs of each disk to be used within your ASM device using the scsi_id command: [ root@localhost ~]# scsi_id --page=0x83 --whitelisted --device=/dev/sdX where sdX is the name of your block device. If you need to run this command to capture multiple WWIDs you can use the following for loop via the shell: [ root@localhost ~]# for i in sdb sdc sdd sde; do \ printf "%s %s\n" "$i" "$(scsi_id --page=0x83 --whitelisted --device=/dev/$i)"; done Output: sdb 360026b900061855e000008a54ea5356a sdc 360026b9000618571000008b54ea5360b sdd 360026b900061855e000008a54ea5356a sde 360026b9000618571000008b54ea5360b Once the WWIDs have been captured, create a file within the /etc/udev/rules.d/ directory and name it 20-dell_oracle.rules. A separate KERNEL entry must exist for each storage device and will require adding the WWID to the "RESULT==" field. An example of what needs to be placed in the /etc/udev/rules.d/20-dell_oracle.rules file #------------------------ start udev rule contents ------------------# KERNEL=="dm-*", PROGRAM="scsi_id --page=0x83 --whitelisted -- device=/dev/%k",RESULT=="360026b9000618571000008b54ea5360b", OWNER:="grid", GROUP:="asmadmin" KERNEL=="dm-*", PROGRAM="scsi_id --page=0x83 --whitelisted -- device=/dev/%k",RESULT=="360026b900061855e000008a54ea5356a", OWNER:="grid", GROUP:="asmadmin" #-------------------------- end udev rule contents ------------------# As you can see from the above, the KERNEL command looks at all dm devices, executes the PROGRAM which captures the WWID, if the RESULT of the WWID is matched, appropriately assign the grid user as the OWNER and the asmadmin group as the GROUP. Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 18 Installing and Configuring ASMLib **Applies ONLY to Oracle Linux 6 and if not setting udev rules** 1. Use oracle.com/technetwork/server-storage/linux/downloads/rhel5-084877.html to download the following files: o oracleasm-support o oracleasmlib o oracleasm NOTE: If your current OS distribution is Oracle Linux, you can obtain the software from the Unbreakable Linux Network using ULN. NOTE: Download the latest versions of oracleasm-support and oracleasmlib but the version of oracleasm must match the current kernel used in your system. Check this information issuing the command uname -r. 2. Enter the following command as root: [ root@localhost ~]# rpm -Uvh oracleasm-support-* \ oracleasmlib-* \ oracleasm-$(uname -r)-* NOTE: Replace * by the correct version numbers of the packages or you can leave them in place of the command ensuring that there are no multiple versions of the packages in the shell's current working directory. Using ASMLib to Mark the Shared Disks as Candidate Disks **Applies ONLY to Oracle Linux 6 and if not setting udev rules** 1. To configure ASM use the init script that comes with the oracleasmsupport package. The recommended method is to run the following command as root: [ root@localhost ~]# /usr/sbin/oracleasm configure –i NOTE: Oracle recommends using the oracleasm command found under /usr/sbin. The /etc/init.d path has not been deprecated, but the oracleasm binary provided by Oracle in this path is used for internal purposes. Default user to own the driver interface []: grid Default group to own the driver interface []: asmadmin Start Oracle ASM library driver on boot (y/n) [ n ]: y Fix permissions of Oracle ASM disks on boot (y/n) [ y ]: y NOTE: In this setup the default user is set to grid and the default group is set to asmadmin. Ensure that the oracle user is part of the asmadmin group. You can do so by using the dell- validated and dell-oracle-utilities rpms. The boot time parameters of the Oracle ASM library are configured and a sequential text interface configuration method is displayed. 2. Set the ORACLEASM_SCANORDER parameter in /etc/sysconfig/oracleasm NOTE: When setting the ORACELASM_SCANORDER to a value, specify a common string associated with your device mapper pseudo device name. For example, if all the device mapper device had a Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 19 prefix string of the word "asm", (/dev/mapper/asm-ocr1,/dev/mapper/asm-ocr2), populate the ORACLEASM_SCANORDER parameter as:ORACLEASM_SCANORDER="dm". This would ensure that oracleasm will scan these disks first. 3. Set the ORACLEASM_SCANEXCLUDE parameter in /etc/sysconfig/oracleasm to exclude non- multipath devices. For example: ORACLEASM_SCANEXCLUDE= NOTE: If we wanted to ensure to exclude our single path disks within /dev/ such as sda and sdb, our ORACLEASM_SCANEXCLUDE string would look like: ORACLEASM_SCANEXCLUDE="sda sdb" 4. To create ASM disks that can be managed and used for Oracle database installation, run the following command as root: [ root@localhost ~]# /usr/sbin/oracleasm createdisk DISKNAME /dev/mapper/diskpartition NOTE: The fields DISKNAME and /dev/mapper/diskpartition should be substituted with the appropriate names for your environment respectively. NOTE: It is highly recommended to have all of your Oracle related disks to be within Oracle ASM. This includes your OCR disks, voting disks, database disks, and flashback recovery disks. 5. Verify the presence of the disks in the ASM library by running the following command as root: [ root@localhost ~]# /usr/sbin/oracleasm listdisks All the instances of DISKNAME from the previous command(s) are displayed. 6. To delete an ASM disk,run the following command: [ root@localhost ~]# /usr/sbin/oracleasm deletedisk DISKNAME 7. To discover the Oracle ASM disks on other nodes in the cluster, run the following command on the remaining cluster nodes: [ root@localhost ~]# /usr/sbin/oracleasm scandisks. Installing Oracle 11g R2 Grid Infrastructure This section gives you the installation instructions for Oracle 11g R2 in a cluster. Before You Begin Before you install the Oracle 11g R2 RAC software on your system: • Ensure that you have already configured your operating system, network, and storage based on the steps from the previous sections within this document. • Locate your Oracle 11g R2 media kit. Configure the System Clock Settings for All Nodes To prevent failures during the installation procedure, configure all the nodes with identical system clock settings. Synchronize your node system clock with the Cluster Time Synchronization Service (CTSS) which is built in Oracle 11g R2. To enable CTSS, disable the operating system network time protocol daemon (ntpd) service using the following commands in this order: Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 20 [ root@localhost ~]# service ntpd stop [ root@localhost ~]# chkconfig ntpd off [ root@localhost ~]# mv /etc/ntp.conf /etc/ntp.conf.orig [ root@localhost ~]# rm /var/run/ntpd.pid Configuring Node One The following steps are for node oneof your cluster environment, unless otherwise specified. 1. Log in as root. 2. If you are not in a graphical environment, start the X Window System by typing: startx 3. Open a terminal window and type: xhost + 4. Mount the Oracle Grid Infrastructure media. 5. Log in as grid user, for example: su - grid. 6. Type the following command to start the Oracle Universal Installer: /runInstaller 7. In the Download Software Updates window, enter your My Oracle Support credentials to download the latest patch updates. If you choose not to download the latest patches, select Skip software updates. 8. In the Select Installation Option window, select Install and Configure Grid Infrastructure for a Cluster and click Next. Select Installation Options Figure 1. 9. In the Select Installation Type window, select Advanced Installation option, and click Next. Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 21 Select Installation TypeFigure 2. 10. In the Select Product Languages window, select English, and click Next. 11. In the Grid Plug and Play Information window, enter the following information: a. Cluster Name—Enter a name for your cluster. b. SCAN Name—Enter the named registered in the DNS server which is unique for the entire cluster. For more details on setting up your SCAN name see, "IP Address and Name Resolution Requirements". For details on enabling GNS, see the wiki article: en.community.dell.com/dell-groups/enterprise_solutions/w/oracle_solutions/1416.aspx. c. SCAN Port—Retain the default port of 1521. d. Configure GNS—Uncheck this option. e. Click Next. Grid Plug and Play Information Figure 3. 12. In the Cluster Node Information window, click Add to add additional nodes that must be managed by the Oracle Grid Infrastructure. a. Enter the public Hostname information b. Enter the Virtual IP name c. Repeat step 12 for each node within your cluster Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 22 Cluster Node Information Figure 4. 13. Click SSH Connectivity and configure your passwordless SSH connectivity by entering the OS Password for the grid user and click Setup. NOTE: The default password set by the dell-validated and dell-oracle-utilities rpms is 'oracle' for both the grid user and oracle user. SSH Connectivity Figure 5. 14. Click Ok and then click Next to go to the next window. 15. In the Network Interface Usage window, make sure that the correct interface types are selected for the interface names. 16. From the Interface Type drop-down list, select the required interface type. The available options are Private, Public, and Do Not Use then click Next. Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 23 Specify Network Interface Usage Figure 6. 17. In the Storage Option Information window, select Automatic Storage Management (ASM) and click Next. Storage Option Information Figure 7. 18. In the Create ASM Disk Group window, enter the following information: a. ASM diskgroup— Enter a name, for example: OCR_VOTE b. Redundancy— For your OCR and voting disks, select High if five ASM disks are available, select Normal if three ASM disks are available, or select External if one ASM disk is available (not recommended). NOTE: If no candidate disks are displayed, click Change Discovery Path and enter ORCL:*or /dev/oracleasm/disks/*. Ensure that you have marked your Oracle ASM disks, for more informations see,"Using ASMLib to Mark the Shared Disks as Candidate Disks". Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 24 Create ASM Disk GroupFigure 8. 19. In the Specify ASM Password window, choose the relevant option under Specify the passwords for these accounts and enter the relevant values for the password. Click Next. Specify ASM Password Figure 9. 20. In the Failure Isolation Support window, select Do Not use Intelligent Platform Management Interface (IPMI). For information on enabling IPMI, see wiki article, en.community.dell.com/dell-groups/enterprise_solutions/w/oracle_solutions/1414.aspx Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 25 Failure Isolation Support Figure 10. 21. In the Privileged Operating Systems Groups window, select: a. asmdba for Oracle ASM DBA (OSDBA for ASM) Group b. asmoper for Oracle ASM Operator (OAOPER for ASM) Group c. asmadmin for Oracle ASM Administrator (OSASM) Group Privileged Operating System Groups 22. In the Installation Location window, specify the values of your Oracle Base and Software Location as configured within the Dell Oracle utilities RPM. NOTE: The default locations used within the Dell Oracle utilites RPM are: . Oracle Base -/u01/app/grid . Software Location - /u01/app/11.2.0/grid Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 26 Specify Installation Location Figure 11. 23. In the Create Inventory window, specify the location for your Inventory Directory. Click Next. Create Inventory Figure 12. NOTE:The default location based on the Dell Oracle utilites RPM for Inventory Directory is/u01/app/oraInventory 24. In the Perform Prerequisite Checks window, check the overall status of all the prerequisites. If any of the prerequisites fail and have the status as Fixable, click Fix & Check Again button and execute the runfixup.sh script provided by the Oracle Universal Installer (OUI). Perform Prerequisite Checks Figure 13. NOTE: If there are other prerequisites that contain the status Error, repeat step 23 else select Ignore All, if proper requirements have been met and Error status still persists after all changes have been fixed. 25. In the Summary window, select Install. Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 27 Summary Figure 14. 26. After the installation is complete, the Execute Configuration Scripts wizard is displayed. Complete the instructions in the wizard and click Ok. Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 28 Execute Configuration Scripts Figure 15. 27. In the Finish window, click Close. Configuring Shared Oracle Home using ACFS NOTE: Please review the following My Oracle Support document ACFS Supported On OS Platforms. [ID 1369107.1] as depending on your kernel version, a patch might be required for ACFS support. The following steps are for node 1 of your cluster environment unless otherwise specified: 1. Login as root and type: xhost + 2. Login as grid user and run the asmca utility by typing: [ root@localhost ~]# su grid [ grid@localhost ~]# $/bin/asmca Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 29 ASM Configuration Assistant Figure 16. 3. In the ASM Configuration Assistant window, select the Disk Groups tab, click Create, and perform the following steps: a. Enter a name of the disk group. For example, ORAHOME. b. Select the External Redundancy, and then select the ASM stamped disk that you want to use for the shared database home. NOTE: If no candidate disks are displayed, click Change Discovery Path and enterORCL:* or /dev/oracleasm/disks/* NOTE: Ensure that you have marked your Oracle ASM disks. For more information, see "Using ASMLib to Mark the Shared Disks as Candidate Disks". Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 30 Create Disk Group Figure 17. 4. Click Ok. 5. Right-click the disk group you have created for the shared Oracle home, and select Create ACFS for Database Home. 6. The Create ACFS Hosted Database Home Volume option is displayed. a. Enter name for the volume (for example, ORAHOME.) b. Enter name for the mount point for Database Home (for example: /u01/app/oracle/acfsorahome). c. Enter the Database Home Size (must equal at least 20 GB). d. Enter the name of the Database Home Owner. (for example: “oracle”). e. Enter the name of the Database Home Owner Group (for example: “oinstall”). f. Click Ok. Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 31 Create Database Home Figure 18. 7. As root, run the acfs_script.sh mentioned in the RUN ACFS Script window. This automounts the new ACFS Home on all nodes. Run ACFS Script Figure 19. 8. Click Close to exit ACFS script window. Installing Oracle 11g R2 Database (RDBMS) Software The following steps are for node 1 of your cluster environment, unless otherwise specified. 1. Log in as root and type: xhost +. 2. Mount the Oracle Database 11gR2 media. 3. Log out as root user and log in as Oracle user by typing: su - oracle Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 32 4. Run the installer script from your Oracle database media: /runInstaller 5. In the Configure Security Updates window, enter your My Oracle Support credentials to receive security updates, else click Next. Configure Security Updates Figure 20. 6. In the Download Software Updates window, enter your My Oracle Support credentials to download patch updates available after the initial release. If you choose not to update at this time, select Skip software updates and click Next. Download Software Updates Figure 21. 7. In the Select Installation Option window, select Install database software only. Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 33 Installation Options Figure 22. 8. In the Grid Installation Options window: a. Select Oracle Real Application Clusters database installation and select all the nodes by clicking the Select All button. b. Click SSH Connectivity and configure your passwordless SSH connectivity by entering the OS Password for the oracle user and selecting Setup. Click Ok and click Next to go the next window. NOTE: The default password set by the dell-validated and dell-oracle-utilities rpms is oracle for both the grid user and oracle user. Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 34 Grid Installation Options Figure 23. SSH Connectivity Figure 24. 9. In the Select Product Lanaguages window, select English as the Language Option and click Next. 10. In the Select Database Edition window, select Enterprise Edition and click Next. Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 35 Select Database Edition Figure 25. 11. In the Installation Location window, a. Specify the location of your Oracle Base configured within the Dell oracle utilities RPM. b. Enter the ACFS shared Oracle home address for Software Location. NOTE: The default locations used within the Dell Oracle utilites RPM are as follows: i. Oracle Base—/u01/app/oracle ii. Software Location—/u01/app/oracle/product/11.2.0/db_1 Specify Installation LocationFigure 26. 12. In the Privileged Operating System Groups window, select “dba” for Database Administrator (OSDBA) Group and “asmoper” for Database Operator (OSOPER) Group and click Next. Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 36 Privileged Operating System Groups Figure 27. 13. In the Perform Prerequisite Checks window, check the overall status of all the prerequisites. 14. If any prerequisites fail and have the status as Fixable, click the Fix & Check Again button. 15. Execute the runfixup.sh script provided by the Oracle OUI. NOTE: If there are other prerequisites that display status Error, repeat step 13, If the Error status still persists after all changes have been fixed, select Ignore All. Perform Prerequisite ChecksFigure 28. 16. In the Summary window, select Install. Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 37 Summary Window Figure 29. 17. On completion of the installation process, the Execute Configuration scripts wizard is displayed. Follow the instructions in the wizard and click Ok. Execute Configuration Scripts Figure 30. NOTE: Root.sh should be run on one node at a time. Creating Diskgroup Using ASM Configuration Assistant (ASMCA) This section contains procedures to create the ASM disk group for the database files and Flashback Recovery Area (FRA). Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 38 1. Log in as grid user. 2. Start the ASMCA utility by typing: [ grid@localhost ~]# $/bin/asmca 3. In the ASM Configuration Assistant window, select the Disk Groups tab. 4. Click Create. ASM Configuration Assistant: Configure ASM: Disk Groups Figure 31. 5. Enter the appropriate Disk Group Name, for example: DBDG. 6. Select External for Redundancy. 7. Select the appropriate member disks to be used to store your database files, for example: ORCL:DB1, ORCL:DB2 Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 39 Create Disk GroupFigure 32. NOTE: If no candidate disks are displayed, click Change Discovery Path and type: ORCL:* or/dev/oracleasm/disks/* NOTE: Please ensure you have marked your Oracle ASM disks. For more information, see "Using ASMLib to Mark the Shared Disks as Candidate Disks". 8. Click Ok to create and mount the disks. 9. Repeat step 4 to step 8 to create another disk group for your Flashback Recovery Area (FRA). NOTE: Make sure that you label your FRA disk group differently than your database disk group name. For labeling your Oracle ASM disks, see "Using ASMLib to Mark the Shared Disks as Candidate Disks". 10. Click Exit to exit the ASM Configuration Assistant. Creating Database Using DBCA The following steps are for node 1 of your cluster environment unless otherwise specified: 1. Login as oracle user. 2. From $, run the DBCA utility by typing: [ oracle@localhost ~]# $/bin/dbca & 3. In the Welcome window, select Oracle Real Application Cluster Database and click Next. Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 40 Database Configuration Assistant: Welcome Figure 33. 4. In the Operations window, select Create Database, and click Next. Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 41 Database Configuration Assistant: Operations Figure 34. 5. In the Database Templates window, select Custom Database, and click Next. Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 42 Database Configuration Assistant: Database Templates Figure 35. 6. In the Database Identification window: a. Select Admin-Managed for Configuration Type. b. Enter appropriate values for Global Database Name and SID Prefix. c. In the Node Selection list box, select All Nodes. d. Click Next. Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 43 Database Configuration Assistant: Database Identification Figure 36. NOTE: For more information on Policy-Managed configuration, see the wiki article en.community.dell.com/dell-groups/enterprise_solutions/w/oracle_solutions/1418.aspx. 7. In the Management Options window, select the default values and click Next. Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 44 Database Configuration Assistant: Management Options Figure 37. 8. In the Database Credentials window, enter the appropriate credentials for your database. Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 45 Database Configuration Assistant: Database Credentials Figure 38. 9. In the Database File Location window, select: a. Automatic Storage Management (ASM) for Storage Type. b. Use Oracle-Managed Files for Storage Location. c. Browse to select the ASM disk group that you created to store the database files (DBDG)for Database Area. Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 46 Database Configuration Assistant: Database File Locations Figure 39. 10. In the Recovery Configuration window: a. Select Specify Flash Recovery Area. b. Browse and select the ASM disk group that you created for Flash Recovery Area. c. Enter a value for Flash Recovery Area Size. d. Select Enable Archiving. e. Click Next. Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 47 Database Configuration Assistant: Recovery Configuration Figure 40. 11. In the Database Content window, click Next. 12. In the Initialization Parameters window: a. Select Custom. b. For the Memory Management section, select Automatic Shared Memory Management. c. Specify appropriate values for the SGA Size and PGA Size. d. Click Next. Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 48 Database Configuration Assistant: Initialization Parameters Figure 41. 13. In the Database Storage window, click Next. Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 49 Database Configuration Assistant: Database Storage Figure 42. 14. In the Creation Options window, click Finish. Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 50 Database Configuration Assistant: Creation Options Figure 43. 15. In the Summary window, click Ok to create database. Deploying Oracle RAC 11g R2 Database on RHEL 6.x / Oracle Linux 6.x 51 Create Database: SummaryFigure 44. NOTE: Database creation can take some time to complete. 16. Click Exit on the Database Configuration Assistant window after the database creation is complete. Acknowledgements The original version of this document was written by Roger Lopez and Adam Miller. Additional thanks to Naveen Iyengar, Phani MV, Balamurugan B, Kai Yu, and Wendy Chen for their contributions
还剩50页未读

继续阅读

下载pdf到电脑,查找使用更方便

pdf的实际排版效果,会与网站的显示效果略有不同!!

需要 6 金币 [ 分享pdf获得金币 ] 4 人已下载

下载pdf

pdf贡献者

metro

贡献于2013-06-21

下载需要 6 金币 [金币充值 ]
亲,您也可以通过 分享原创pdf 来获得金币奖励!
下载pdf