Sunday, March 24, 2013

Unrecognized option: -nojit Could not create the Java virtual machine.

Issue: Unrecognized option: -nojit
Could not create the Java virtual machine.

Cause: it is not gettin the java...while Calling $COMMON_TOP/util/j2sdk1.4.2_13/bin/jre

Soltuion:

export ADJVAPRG=$COMMON_TOP/util/j2sdk1.4.2_13/bin/java
export AFJVAPRG=$COMMON_TOP/util/j2sdk1.4.2_13/bin/java

it Worked for me.

Wednesday, March 20, 2013

how to check printer status in linux


[username@hostname ~]$ lpstat -p
printer apcheck is idle.  enabled since Sun 03 Mar 2013 04:02:09 AM EST
printer HP4 is idle.  enabled since Sun 03 Mar 2013 04:02:09 AM EST
printer HP5 is idle.  enabled since Sun 03 Mar 2013 04:02:09 AM EST


To check the printer status
[username@hostname ~]$lpstat -a




Script to execute a sql file in background.


script1.sh
======

sqlplus -s "uname/passwd" <<EOF > "logfile_name.txt"
Prompt Description of the script
Prompt
set head on;
commit;
@/ora/oracle/app1/filename.sql;
commit;
exit;
EOF
exit


bash-3.00$nohup sh script1.sh &

Saturday, March 16, 2013

Upgrading EBS R12 11gR1 RAC to 11gR2 RAC




This Document shows the step by step of upgrading E-Business Suites R 12.1.1 Database from 11gR1 RAC to 11gR2 RAC on 3-node database cluster. Below upgrade path was followed (in sequence) to achieve the final goal successfully.


(a)    Shutdown Applications on all the nodes gracefully.
(b)   Shutdown existing 11gR1 RAC database gracefully.
(c)    Shutdown the CRS on all the RAC nodes.
(d)   Take the Backup of entire system.


(e)   Install/Upgrade RPMs required for 11gR2 RAC Installation.
(f)     Add SCAN VIPs to the DNS.
(g)    Setup of Network Time Protocol.
(h)   Create 3 ASM Disks for 11gR2 OCR and Voting Disks.


(i)      Hide the Existing 11gR1 CRS installation as if there is no Clusterware existed on the RAC Nodes.
(j)     Install 11gR2 Grid Infrastructure under the new GRID_HOME.
(k)    Install 11gR2 RAC database software.
(l)      Install 11gR2 Example software.


(m) Move the existing 11gR1 ASM Diskgroups to the 11gR2 Grid Infrastructure ASM.


(n)   The Database should be up without any issue as nothing has been changed from Database standpoint at this stage except the ASM which is now running from 11gR2 HOME.


(o)   Apply the required patches to the existing 11gR1 HOME.


(p)   Apply the Required Patches.


(q)   Fresh install 11gR2 RAC (Software Only).
(r)     Apply 11gR2 RAC Example Software on the same 11gR2 RAC HOME.
(s)    Upgrade existing 11gR1 RAC database to 11gR2 RAC Manually.


(t)     Enable and run the Autoconfig on the DB Tier from the 11gR2 RAC HOME.
(u)   Run the AutoConfig on the APPS Tier to generate the new RAC enabled connect strings.
(v)    Take the Complete Backup of the Entire System.


(w)  listnener.ora, tnsnames.ora, sqlnet.ora and their ifiles from all the RAC nodes.
(x)    listener.ora and tnsnames.ora from both the application nodes.
(y)    CONTEXT_FILE from all the 5 nodes.
(z)    pfile of the RAC database.



Existing Node configuration:



Node Type
NAME
Version
Function
Application Node1
appsnode1.hingu.net
Oracle EBS 12.1.1
WEB Node
Application Node2
appsnode2.hingu.net
Oracle EBS 12.1.1
Concurrent Node (Batch Processing)
RAC DB Node1
node1.hingu.net
Oracle RDBMS 11.1.0.7
Clustered DB node1
RAC DB Node2
node2.hingu.net
Oracle RDBMS 11.1.0.7
Clustered DB node2
RAC DB Node3
node3.hingu.net
Oracle RDBMS 11.1.0.7
Clustered DB node3
DNS Server
lab-dns.hingu.net
OEL 5.5
DNS





               




            Before Upgrade
            After the Upgrade
E-Business Suite
Oracle EBS R12.1.1
Oracle EBS R12.1.1
BASE Directory
/u01/app/oracle/ebsR12
/u01/app/oracle/ebsR12
ORACLE_HOME
/u01/app/oracle/ebsR12/apps/tech_st/10.1.2
/u01/app/oracle/ebsR12/apps/tech_st/10.1.2
IAS_ORACLE_HOME
/u01/app/oracle/ebsR12/apps/tech_st/10.1.3
/u01/app/oracle/ebsR12/apps/tech_st/10.1.3
APPL_TOP
/u01/app/oracle/ebsR12/apps/apps_st/appl
/u01/app/oracle/ebsR12/apps/apps_st/appl
INST_TOP
/u01/app/oracle/ebsR12/inst/apps/<CONTEXT_NAME>
/u01/app/oracle/ebsR12/inst/apps/<CONTEXT_NAME>
Port Pool
1
1
Application OS Owner
applmgr:(oinstall, dba)
applmgr:(oinstall, dba)
OS Platform
Oracle Enterprise Linux 5.5 (32 bit)
Oracle Enterprise Linux 5.5 (32 bit)

DB Tier
DB Tier
RDBMS
Oracle 11gR1 RAC (11.1.0.7)
Oracle 11gR2 RAC (11.2.0.1)
ORACLE_BASE
/u01/app/oracle
/u01/app/oracle
ORACLE_HOME
/u01/app/oracle/EBSR12
/u01/app/oracle/db11201
DB_HOME Owner
oracle:(oinstall, dba)
oracle:(oinstall, dba)
DB LISTENER
LABDB_LISTENER
LABDB_LISTENER
DB Listener Host/port
node1-vip, node2-vip, node3-vip (TCP:1522)
node1-vip, node2-vip, node3-vip (TCP:1522)
DB Storage Type, File Management
ASM with OMFs
ASM with OMFs
ASM diskgroups for DB and FRA
DATA (60GB), FRA (10GB)
DATA (60GB), FRA (10GB)
OS Platform
Oracle Enterprise Linux 5.5 (32 bit)
Oracle Enterprise Linux 5.5 (32 bit)
Clusterware
Oracle 11gR1 Clusterware 11.1.0.7
Oracle 11gR2 Grid Infrastructure 11.2.0.1
ASM
Oracle 11gR1 RAC 11.1.0.7
Oracle 11gR2 Grid Infrastructure 11.2.0.1
ASM_HOME
/u01/app/oracle/asm
/u01/app/grid11201
Cluster Name
Lab
Lab
Database Cluster Nodes
node1, node2, node3
node1, node2, node3
Clusterware Home
/u01/app/oracle/crs (CRS_HOME)
/u01/app/grid11201 (GRID_HOME)
Clusterware Owner
oracle:(oinstall, dba)
oracle:(oinstall, dba)
VIPs
node1-vip, node2-vip, node3-vip
node1-vip, node2-vip, node3-vip
SCAN
N/A
lab-scan.hingu.net
SCAN VIPs
N/A
3 SCAN VIPs resolving lab-scan.hingu.net by DNS
SCAN_LISTENER Host/port
N/A
Scan VIPs Endpoint: (TCP:1525)
OCR and Voting Disks Storage Type
Raw Devices
ASM
OCR Disks/Voting Disk
/dev/raw/raw1-2 (OCR), /dev/raw/raw3-5 (Voting)
+GIS_FILES (Normal Redundancy)
ASM_HOME
/u01/app/oracle/asm
/u01/app/grid11201
ASM_HOME Owner
oracle:(oinstall, dba)
oracle:(oinstall, dba)
ASMLib user:group
oracle:oinstall
oracle:oinstall
ASM LISTENER
LISTENER (TCP:1521)
LISTENER (TCP:1521)


Software/Patches Required in this Upgrade:

Oracle 11g R2 Grid Infrastructure (11.2.0.1)
Oracle Database software 11g R2 (11.2.0.1)
Oracle Database Example software 11g R2 (11.2.0.1)
Patch 8919489 (12.1 TXK.B Delta 3)
Patch 9583541 (Post requirement for the patch 8919489)
Patch 9062910 (11g Release 2 interoperability patch for Release 12.1)
Patch 10163753 (Application Patch)
Patch 11071569 (Application Patch)
Patch 9738085 (Application Patch -- AutoConfig)
Patch 9852070 (Application Patch -- AutoConfig)
Patch 6400501 (Application Patch – 10.1.2 ORACLE_HOME)
Patch 7695070 (Database Patch (for 11gR1) -- Prereq Patch for DSTv16 12320006)
Patch 12320006 (Database Patch -- DSTv16 for 11gR1 and 11gR2)
Patch 12565410 (Database Patch -- OJVM DSTv16)
Patch 6880880 (Latest OPatch)
  
Documents Reviewed/Followed:

·         Interoperability Notes EBS R12 with Database 11gR2 [ID 1058763.1]
·         Oracle E-Business Suite Installation and Upgrade Notes Release 12 (12.1.1) for Linux x86 [ID 761564.1]
·         Using Oracle 11g Release 2 Real Application Clusters with Oracle E-Business Suite Release 12 [ID 823587.1]
·         Using AutoConfig to Manage System Configurations in Oracle E-Business Suite Release 12 [ID 387859.1]
·         Complying with Daylight Saving Time (DST) and Time Zone Rule Changes in E-Business Suite 12 [ID 563019.1]
·         Testing your Oracle JVM Time Zone Setup [ID 416860.1]
·         Applying the DSTv16 update for the Oracle Database [ID 1319204.1]
·         How To Download And Install The Latest OPatch Version [ID 274526.1]


Backup the System Before the Upgrade:

Perform the Application and Database Backup:

Shutdown Applications on all the Application nodes gracefully.

   $TNS_ADMIN/admin/scripts/adstpall.sh

Shutdown existing 11gR1 RAC database on all the RAC nodes gracefully.

    /u01/app/oracle/EBSR12/bin/srvctl stop database –d LABDB

Shutdown the CRS on all the RAC nodes.

/u01/app/oracle/crs/bin/crsctl stop crs  Ăź as root

appsnode1 and appsnode2

mkdir backup
cd backup
tar cvf appsnode1_R12_backup.tar /u01/app/oracle/ebsR12/*    Ăź on appsnode1
tar cvf appsnode2_R12_backup.tar /u01/app/oracle/ebsR12/*    Ăź on appsnode2

On RAC node1:

mkdir backup
cd backup

-- Backup the OCR and Voting Disk.

dd if=/dev/dev/raw1  of=ocr_disk_10gr2.bkp
dd if=/dev/dev/raw3  of=voting_disk_10gr2.bkp
      
-- Backup all the HOMEs and CRS related OS Files.

tar cvf node1_crs_10gr2.tar /u01/app/oracle/crs/*
tar cvf node1_asm_10gr2.tar /u01/app/oracle/asm/*
tar cvf node1_db_10gr2.tar /u01/app/oracle/db/*
tar cvf node1_etc_oracle /etc/oracle/*

cp /etc/inittab etc_inittab

mkdir etc_init_d
cd etc_init_d
cp /etc/init.d/init* .

On RAC node2:

mkdir backup
cd backup
      
tar cvf node2_crs_10gr2.tar /u01/app/oracle/crs/*
tar cvf node2_asm_10gr2.tar /u01/app/oracle/asm/*
tar cvf node2_db_10gr2.tar /u01/app/oracle/db/*
tar cvf node2_etc_oracle /etc/oracle/*

cp /etc/inittab etc_inittab
      
mkdir etc_init_d
cd etc_init_d
cp /etc/init.d/init* .
      
On RAC node3:

mkdir backup
cd backup

tar cvf node3_crs_10gr2.tar /u01/app/oracle/crs/*
tar cvf node3_asm_10gr2.tar /u01/app/oracle/asm/*
tar cvf node3_db_10gr2.tar /u01/app/oracle/db/*
tar cvf node3_etc_oracle /etc/oracle/*
      
cp /etc/inittab etc_inittab
      
mkdir etc_init_d
cd etc_init_d
cp /etc/init.d/init* .

RMAN Full database backup was taken.


·         Install/Upgrade RPMs required for 11gR2 RAC Installation.
·         Add SCAN VIPs to the DNS.
·         Setup of Network Time Protocol.
·         Create 3 ASM Disks to create a diskgroup (Normal Redundancy) to store 11gR2 OCR and Voting Disks.

Minimum Required RPMs for 11gR2 RAC on OEL 5.5 (All the 3 RAC Nodes):

Below command verifies whether the required rpms are installed or not. Any missing rpms can be installed from the OEL Media Pack

rpm -q binutils compat-libstdc++-33 elfutils-libelf elfutils-libelf-devel elfutils-libelf-devel-static \
              gcc gcc-c++ glibc glibc-common glibc-devel glibc-headers kernel-headers ksh libaio libaio-devel \
              libgcc libgomp libstdc++ libstdc++-devel make numactl-devel sysstat unixODBC unixODBC-devel

I had to install below RPM.

numactl-devel Ă  Located on the 3rd CD of OEL 5.5 Media pack.

[root@node1 ~]# rpm -ivh numactl-devel-0.9.8-11.el5.i386.rpm
warning: numactl-devel-0.9.8-11.el5.i386.rpm: Header V3 DSA signature: NOKEY, key ID 1e5e0159
Preparing...                ########################################### [100%]
   1:numactl-devel          ########################################### [100%]
[root@node1 ~]#

I had to upgrade the cvuqdisk RPM by removing and installing the same with higher version. This step is also taken care by rootupgrade.sh script.

cvuqdisk Ă  Available on Grid Infrastructure Media (under rpm folder)

rpm -e cvuqdisk
export CVUQDISK_GRP=oinstall
echo $CVUQDISK_GRP
rpm -ivh cvuqdisk-1.0.7-1.rpm


Add SCAN VIPs to DNS:

SCAN VIPS to configure in DNS which resolves to lab-scan.hingu.net:

192.168.2.151
192.168.2.152
192.168.2.153

In my existing DNS setup, the below two files were modified with the entry in RED to add these SCAN VIPs into the DNS.

/var/named/chroot/var/named/hingu.net.zone
/var/named/chroot/var/named/2.168.192.in-addr.arpa.zone

/var/named/chroot/var/named/hingu.net.zone

$TTL    1d
hingu.net.  IN    SOA   lab-dns.hingu.net. root.hingu.net. (
    100        ; se = serial number
    8h         ; ref = refresh
    5m         ; ret = update retry
    3w         ; ex = expiry
    3h         ; min = minimum
    )

    IN    NS    lab-dns.hingu.net.

; DNS server

lab-dns    IN    A    192.168.2.200

; Application Nodes Public name

appsnode1          IN    A    192.168.2.11
appsnode2          IN    A    192.168.2.22

; RAC Nodes Public name

node1          IN    A    192.168.2.1
node2          IN    A    192.168.2.2
node3          IN    A    192.168.2.3

; RAC Nodes Public VIPs

node1-vip      IN    A    192.168.2.51
node2-vip      IN    A    192.168.2.52
node3-vip      IN    A    192.168.2.53

; 3 SCAN VIPs

lab-scan       IN    A    192.168.2.151
lab-scan       IN    A    192.168.2.152
lab-scan       IN    A    192.168.2.153

; Storage Network

nas-server IN    A    192.168.1.101
node1-nas  IN    A    192.168.1.1
node2-nas  IN    A    192.168.1.2
node3-nas  IN    A    192.168.1.3


/var/named/chroot/var/named/2.168.192.in-addr.arpa.zone

$TTL    1d
@   IN    SOA   lab-dns.hingu.net. root.hingu.net. (
    100        ; se = serial number
    8h         ; ref = refresh
    5m         ; ret = update retry
    3w         ; ex = expiry
    3h         ; min = minimum
    )

    IN    NS    lab-dns.hingu.net.

; DNS machine name in reverse
200        IN    PTR lab-dns.hingu.net.

; Application Nodes Public Name in Reverse

11          IN    PTR      appsnode1.hingu.net.
22          IN    PTR      appsnode2.hingu.net.

; RAC Nodes Public Name in Reverse

1          IN    PTR node1.hingu.net.
2          IN    PTR node2.hingu.net.
3          IN    PTR node3.hingu.net.

; RAC Nodes Public VIPs in Reverse

51          IN    PTR      node1-vip.hingu.net.
52          IN    PTR      node2-vip.hingu.net.
53          IN    PTR      node3-vip.hingu.net.

; RAC Nodes SCAN VIPs in Reverse

151         IN    PTR      lab-scan.hingu.net.
152         IN    PTR      lab-scan.hingu.net.
153         IN    PTR      lab-scan.hingu.net.

Restarted the DNS Service (named):

service named restart

NOTE: nslookup for lab-scan should return names in random order every time.

Network Time Protocol Setting (On all the RAC Nodes):
      
Oracle Time Synchronization Service is chosen to be used over the Linux system provided ntpd. So, ntpd needs to be deactivated and deinstalled to avoid any possibility of it being conflicted with the Oracle’s Cluster Time Sync Service (ctss).

# /sbin/service ntpd stop
chkconfig ntpd off
mv /etc/ntp.conf /etc/ntp.conf.org

Also remove the following file:

/var/run/ntpd.pid


Network Service Cache Daemon (all the RAC nodes)

The Network Service Cache Daemon was started on all the RAC nodes.

Service nscd start

Create ASM Disks for diskgroup GIS_FILES to store 11gR2 OCR and Voting Disks


On the rest of the Nodes, it was simply scanned for the new disks.

oracleasm scandisks


·         Hide the Existing 11gR1 CRS installation as if there is no Clusterware existed on the RAC Nodes.
·         Install 11gR2 Grid Infrastructure under the new GRID_HOME (/u01/app/grid11201)
·         Install 11gR2 RAC database software (/u01/app/oracle/db11201).
·         Install 11gR2 Example Software (/u01/app/oracle/db11201).

Hide the Existing 11gR1 CRS Installation on RAC Nodes:

In order to install 11gR2 GI, it is required to hide the existing 11gR1 CRS so that 11gR2 GI is installed without any conflict with 11gR1 CRS.

Shutdown CRS on all the RAC nodes

crsctl stop crs

renamed these files/Directories

mv /etc/oracle /etc/oracle_bkp
mkdir /etc/init.d/bkp
mv /etc/init.d/init* /etc/init.d/bkp

removed below lines from the /etc/inittab (the inittab was already backed up in Pre-Upgrade tasks as well)

h1:35:respawn:/etc/init.d/init.evmd run >/dev/null 2>&1 </dev/null
h2:35:respawn:/etc/init.d/init.cssd fatal >/dev/null 2>&1 </dev/null
h3:35:respawn:/etc/init.d/init.crsd run >/dev/null 2>&1 </dev/null

removed the network socket files:

rm -rf /tmp/.oracle
rm -rf /var/tmp/.oracle

rebooted all the RAC nodes at this stage

reboot

NOTE: CRS stack did not come up at this stage because the existing 11gR1 CRS had made hidden.


Install 11gR2 Grid Infrastructure:

Grid Infrastructure installation process:

Installation Option:
 Install and Configure Grid Infrastructure for a Cluster
Installation Type:
Advanced Installation
Product Language:
English
Grid Plug and Play:
Cluster Name: lab
SCAN Name: lab-scan.hingu.net
SCAN Port: 1525
Configure GNS: Unchecked
                Cluster Node Information:
Entered the Hostname and VIP names of the Cluster Nodes.
Checked the SSH connectivity
Network Interface Usage:
OUI picked up all the interfaces correctly. I did not have to make any changes here.
Storage Option:
Automatic Storage Management (ASM)
Create ASM Disk Group:
Disk Group Name: GIS_FILES
Redundancy: Normal
Candidate Disks: ORCL:GISDSK01, ORCL:GISDSK02, ORCL:GISDSK03
ASM Password:
Use Same Password for these accounts. (Ignored password warnings.)
Failure Isolation:
Do not use Intelligent Platform Management Interface (IPMI)
Operating System Groups:
ASM Database Administrator (OSDBA) Group: dba
ASM Instance Administrator Operator (OSOPER) Group: oinstall
ASM Instance Administrator (OSASM) Group: oinstall
Installation Location:
Oracle Base: /u01/app/oracle
Software Location: /u01/app/grid11201
Create Inventory:
Inventory Directory: /u01/app/oraInventory
Prerequisite Checks:
No issue/Errors
Summary Screen:
Verified the information here and pressed “Finish” to start installation.

At the End of the installation, the root.sh script needed to be executed as root user.

/u01/app/grid11201/root.sh

After the successful completion of this script, the 11g R2 High Availability Service (CRS, CSS and EVMD) started up and running.
Verified that the status of the installation using below set of commands.

crsctl check cluster –all
crs_stat –t –v
crsctl check ctss

The GSD and OC4J resources are by default disabled. Enabled GSD them as below.

srvctl enable nodeapps –g
srvctl start nodeapps –n node1
srvctl start nodeapps –n node2
srvctl start nodeapps –n node3
srvctl enable oc4j
srvctl start oc4j

Next, netca was invoked from 11gR2 Grid Infrastructure Home to reconfigure the listener “LISTENER” from new 11gR2 HOME.

/u01/app/oracle/grid11201/bin/netca



Install 11gR2 RAC Software

Start the runInstaller from 11gR2 Real Application Cluster (RAC) Software Location:

/home/oracle/db11201/database/runInstaller

Real Application Cluster installation process:

Configure Security Updates:
Ignore the “Connection Failed” alert.
Installation Option:
Install database software only
Node Selection:
Selected All the Nodes (node1,node2 and node3)
Product Language:
English
Database Edition:
Enterprise Edition
Installation Location:
Oracle Base: /u01/app/oracle
Software Location: /u01/app/oracle/db11201
Operating System Groups:
Database Administrator (OSDBA) Group: dba
Database Operator (OSOPER) Group: oinstall
Summary Screen:
Verified the information here and pressed “Finish” to start installation.

At the End of the installation, the below scripts needs to be executed on all the nodes as root user.

/u01/app/oracle/db11201/root.sh

Install 11gR2 Example Software:

Start the runInstaller from 11gR2 Example Software Location:

/home/oracle/example/example/runInstaller

Specify Installation Location:
Oracle Base: /u01/app/oracle
Software Location: /u01/app/oracle/db11201
Summary Screen:
Verified the information here and pressed “Finish” to start installation.


·         Invoked the asmca from the 11gR2 Grid Infrastructure HOME (/u01/app/grid11201).
·         Mount the FRA and DATA diskgroup using ASMCA. This way, asmca moved the DATA and FRA diskgroups to the 11gR2 GI and same time also registered these diskgroups to the OCR.

/u01/app/grid11201/bin/asmca

03.JPG




At this Stage, I tried to start the existing 11gR1 RAC database to confirm that it comes up with no issue as it was then managed by 11gR2 Grid Infrastructure and running on 11gR2 Grid ASM. While Starting the existing 11gR1 RAC database using SQLPLUS, the below error was received. This is because, as per Oracle, 11gR2 has dynamic configuration of cluster and 11gR1 or older releases have static configuration.  So, in order to run older version databases on 11gR2 Grid Infrastructure, the cluster configuration needs to be made persistent by pinning the nodes.


ORA-01078: failure in processing system parameters
ORA-01565: error in identifying file '+DATA/labdb/spfilelabdb.ora'
ORA-17503: ksfdopn:2 Failed to open file +DATA/labdb/spfilelabdb.ora
ORA-15077: could not locate ASM instance serving a required diskgroup
ORA-29701: unable to connect to Cluster Manager


/u01/app/grid11201/bin/crsctl pin css –n node1 node2 node3


The LABDB database was then started using existing 11gR1 home through SQLPLUS after pinning the node.  Due to the Oracle bug, 11gR1 database was not getting started by srvctl after it was registered to 11gR2 Grid Infrastructure so I had to use SQLPLUS all the time to start/stop the 11gR1 RAC database until it was migrated to 11gR2 RAC.


·         Apply the DST RDBMS (DSTv16) Patch to the Existing 11gR1 ORACLE_HOME.
·         Apply the DST patch for OJVM component of RDBMS to the 11gR1 HOME.
·         Apply the DST RDBMS (DSTv16) Patch to the new 11gR2 ORACLE_HOME.
·         Apply the DST patch for OJVM component of RDBMS to the 11gR2 HOME.
·         Re-compile the Invalid Objects in the database.
·         Purge The RECYCLE BIN.
·         Create the required directories to the new 11gR2 RAC HOME.


DST Patch

The below Metalink Notes (563019.1 and 1319204.1) were followed to determine what DST patches were needed in this environment. The required patches were then downloaded and installed (in sequence) in the existing 11gR1 RAC HOME.

·         Complying with Daylight Saving Time (DST) and Time Zone Rule Changes in E-Business Suite 12 [ID 563019.1]
·         Testing your Oracle JVM Time Zone Setup [ID 416860.1]
·         Applying the DSTv16 update for the Oracle Database [ID 1319204.1]

Patch 6880880 (Latest OPatch)
Patch 7695070 (Database Patch -- Prereq Patch for DSTv16 12320006 for 11.1.0.7)
Patch 12320006 (Database Patch -- DSTv16 – for 11.1.0.7 and 11.2.0.1)
Patch 12565410 (Database Patch -- OJVM DSTv16 – for 11.1.0.7 and 11.2.0.1)

The Metalink note 1319204.1 was followed to apply the DSTv16 (database and JVM).


Shutdown the RAC Database LABDB and listener LABDB_LISTENER

srvctl stop listener –l LABDB_LISTENER_NODE1 –n node1
srvctl stop listener –l LABDB_LISTENER_NODE2 –n node2
srvctl stop listener –l LABDB_LISTENER_NODE3 –n node3
shutdown immediate

Patch 6880880 (Latest OPatch)

node1:

ssh node1 mv /u01/app/oracle/db11201/OPatch /u01/app/oracle/db11201/OPatch_orig
ssh node2 mv /u01/app/oracle/db11201/OPatch /u01/app/oracle/db11201/OPatch_orig
ssh node3 mv /u01/app/oracle/db11201/OPatch /u01/app/oracle/db11201/OPatch_orig
cd /u01/app/oracle/db11201
unzip /home/oracle/patches_for_R12_DB_upgrade_to_11gR2/p6880880_112000_LINUX.zip
scp -r OPatch node2:/u01/app/oracle/db11201
scp -r OPatch node3:/u01/app/oracle/db11201


Patch 7695070 (Database Patch -- Prereq Patch for DSTv16 12320006)

. ebslab.env
unzip p7695070_111070_Linux-x86.zip
cd 7695070
/u01/app/oracle/EBSR12/OPatch/opatch apply –local
relink all

The above process was repeated on all the remaining RAC Nodes to install this patch.

Patch 12320006 (DSTv16)

. ebslab.env
unzip p12320006_111070_Linux-x86.zip
cd 12320006
/u01/app/oracle/EBSR12/OPatch/opatch apply -local

The above process was repeated on all the remaining RAC Nodes to install this patch.

Patch 12565410 (DSTv16 OJVM)

. ebslab.env
unzip p12565410_111070_Generic.zip
cd 12565410
/u01/app/oracle/EBSR12/OPatch/opatch apply –local

The above process was repeated on all the remaining RAC Nodes to install this patch.
Then, the below two SQL scripts (fixTZa and fixTZb) were run in order as per the instructions in the same Metalink note 1319204.1.

SQL> @$ORACLE_HOME/javavm/admin/fixTZa
SQL> alter system set cluster_database=false scope=spfile;
SQL> shutdown immediate;
SQL> startup migrate
SQL> $ORACLE_HOME/javavm/admin/fixTZb
SQL> alter system set cluster_database=true scope=spfile sid ='*';
SQL> shutdown immediate;
SQL> startup;

Start the RAC Database LABDB and listener LABDB_LISTENER on all the RAC Nodes

srvctl start listener –l LABDB_LISTENER_NODE1 –n node1
srvctl start listener –l LABDB_LISTENER_NODE2 –n node2
srvctl start listener –l LABDB_LISTENER_NODE3 –n node3
startup


Here is the output of lsinventory after applying the DB patches.


Apply the DST Patches to the newly Installed 11gR2 RAC HOME:

In order to maintain the same DST patch level on the new 11gR2 RAC HOME, the same patches were downloaded for the 11.2.0.1 release and applied using the latest opatch driver.

Patch 6880880 (Latest OPatch)

node1:

ssh node1 mv /u01/app/oracle/db11201/OPatch /u01/app/oracle/db11201/OPatch_orig
ssh node2 mv /u01/app/oracle/db11201/OPatch /u01/app/oracle/db11201/OPatch_orig
ssh node3 mv /u01/app/oracle/db11201/OPatch /u01/app/oracle/db11201/OPatch_orig
cd /u01/app/oracle/db11201
unzip /home/oracle/patches_for_R12_DB_upgrade_to_11gR2/p6880880_112000_LINUX.zip
scp -r OPatch node2:/u01/app/oracle/db11201
scp -r OPatch node3:/u01/app/oracle/db11201

Patch 12320006 (DSTv16)

export ORACLE_HOME=/u01/app/oracle/db11201
unzip p12320006_111070_Linux-x86.zip
cd 12320006
/u01/app/oracle/db11201/OPatch/opatch apply –local

The above process was repeated on all the remaining RAC Nodes to install this patch.

Patch 12565410 (DSTv16 OJVM)

export ORACLE_HOME=/u01/app/oracle/db11201
unzip p12565410_112010_Generic.zip
cd 12565410
/u01/app/oracle/EBSR12/OPatch/opatch apply –local


The above process was repeated on all the remaining RAC Nodes to install this patch.

Create extra directories:

ssh node1 mkdir -p /u01/app/oracle/db11201/appsutil/outbound/LABDB1_node1
ssh node2 mkdir -p /u01/app/oracle/db11201/appsutil/outbound/LABDB2_node2
ssh node3 mkdir -p /u01/app/oracle/db11201/appsutil/outbound/LABDB3_node3
ssh node1 mkdir -p /u01/app/oracle/db11201/admin/LABDB1_node1
ssh node2 mkdir -p /u01/app/oracle/db11201/admin/LABDB2_node2
ssh node3 mkdir -p /u01/app/oracle/db11201/admin/LABDB3_node3



These patches needed to be applied in order to upgrade the 11gR1 EBS R12 Database to 11gR2 RAC. The application was brought down on both the application node gracefully and followed the standard process to apply these patches (i.e, adadmin, adpatch and README.txt).

·         Patch 8919489 (12.1 TXK Delta 3)
·         Patch 9583541 (Post requirement for the patch 8919489)
·         Patch 9062910 (11g Release 2 interoperability patch for Release 12.1)
·         Patch 10163753 (Application Patch)
·         Patch 11071569 (Application Patch)
·         Patch 9738085 (Application Patch -- AutoConfig)
·         Patch 9852070 (Application Patch -- AutoConfig)
·         Patch 6400501 (Application Patch – 10.1.2 ORACLE_HOME)

Apply Patch 8919289 (EBS R 12.1.3 Techstack Product Release Update Patch):

This patch needs to be applied to use the named listener on the DB tier if it has not been applied during the installation of R12.1.1 on 11gR1 RAC. In my case, this patch has already been applied at the time of installing the R12.1.1 on 11gR1 RAC as I wanted to use the named LISTENER (LABDB_LISTENER).

Here is how it was installed.

a)      Applied the patch 8919489 using adpatch by following its README on appsnode1
b)      Applied the patch 8919489 using adpatch by following its README on appsnode2
c)       Ran the AutoConfig again on Admin Node (appsnode1.hingu.net) after applying the patch on appsnode2.
d)      Tried to start the Application on appsnode1.hingu.net and it got warning message saying it could not deploy the latest EAR file.

The Forms were manually deployed by following the Metalink note “Deploying a New forms.ear File in Oracle Applications Release 12 [ID 397174.1]”. While running the script to deploy the forms, it got errored out with the below error message complaining it could not find the opmnctl executable.

cp $INST_TOP/ora/10.1.3/j2ee/forms/config/system-jazn-data.xml $INST_TOP/ora/10.1.3/j2ee/forms/config/system-jazn-data.xml.ori

*******FATAL ERROR*******
PROGRAM : /u01/app/oracle/ebsR12/apps/apps_st/appl/fnd/12.0.0/patch/115/bin/txkCfgOC4JApp.pl(/u01/app/oracle/ebsR12/apps/apps_st/appl/fnd/12.0.0/bin/txkrun.pl)
TIME    : Sat Oct 29 13:47:38 2011
FUNCTION: TXK::Process::run [ Level 3 ]
MESSAGES:
Command error: <rc> = 32512, <command> = /u01/app/oracle/ebsR12/apps/tech_st/10.1.3/opmn/bin/opmnctl start


STACK TRACE
       TXK::Error::abort('TXK::Error','HASH(0x923639c)') called at /u01/app/oracle/ebsR12/apps/apps_st/appl/au/12.0.0/perl/TXK/Common.pm line 299
       TXK::Common::doError('TXK::Process=HASH(0xb7cad114)','Command error: <rc> = 32512, <command> = /u01/app/oracle/ebsR...','undef') called at /u01/app/oracle/ebsR12/apps/apps_st/appl/au/12.0.0/perl/TXK/Common.pm line 314
       TXK::Common::setError('TXK::Process=HASH(0xb7cad114)','Command error: <rc> = 32512, <command> = /u01/app/oracle/ebsR...') called at /u01/app/oracle/ebsR12/apps/apps_st/appl/au/12.0.0/perl/TXK/Process.pm line 449
       TXK::Process::run('TXK::Process=HASH(0xb7cad114)','HASH(0x9756fe4)') called at /u01/app/oracle/ebsR12/apps/apps_st/appl/fnd/12.0.0/patch/115/bin/txkCfgOC4JApp.pl line 1587
       TXK::RunScript::execOPMNControl('HASH(0x9a3b224)') called at /u01/app/oracle/ebsR12/apps/apps_st/appl/fnd/12.0.0/patch/115/bin/txkCfgOC4JApp.pl line 599
       require /u01/app/oracle/ebsR12/apps/apps_st/appl/fnd/12.0.0/patch/115/bin/txkCfgOC4JApp.pl called at /u01/app/oracle/ebsR12/apps/apps_st/appl/au/12.0.0/perl/TXK/RunScript.pm line 105
       TXK::RunScript::require('TXK::RunScript','/u01/app/oracle/ebsR12/apps/apps_st/appl/fnd/12.0.0/patch/115...') called at /u01/app/oracle/ebsR12/apps/apps_st/appl/au/12.0.0/perl/TXK/Script.pm line 177
       eval {...} called at /u01/app/oracle/ebsR12/apps/apps_st/appl/au/12.0.0/perl/TXK/Script.pm line 177
      TXK::Script::run('TXK::Script=HASH(0x9b1c980)','/u01/app/oracle/ebsR12/inst/apps/LABDB_appsnode1/logs/appl/rg...','/u01/app/oracle/ebsR12/apps/apps_st/appl/fnd/12.0.0/patch/115...') called at /u01/app/oracle/ebsR12/apps/apps_st/appl/fnd/12.0.0/bin/txkrun.pl line 174


While further debugging, it was found out that the ORACLE_HOME and TNS_ADMIN variables in $IAS_ORACLE_HOME/opmn/bin/opmnctl were pointing to/nfs/bld/d26/PRDXBLD9/apps/tech_st/10.1.3 instead of /u01/app/oracle/ebsR12/apps/tech_st/10.1.3. These variables were correctly modified and then continued with the Note 397174.1 to deploy the forms.ear and after that, it got deployed successfully.

After that, the Application services were started on both the application nodes using $INST_TOP/admin/scripts/adstrtal.sh and tested logging in to the application and forms and the results were all successful.

The application services were brought down after that on both the application nodes to apply the remaining patches next.

Apply the Patches 9062910, 10163753, 11071569, 9738085:

Next, the above mentioned patches were applied using the standard adpatch utility with no issue. These patches have no Pre/post steps.

Apply the Patch 9852070:

After applying this patch using adpatch there are post steps mentioned in the README.txt which were followed to make the appsutil.zip. This file was then shipped to all the RAC nodes where it was unzipped using “unzip -o appsutil.zip” command under the $ORACLE_HOME.

Apply the Patch 6400501 to the ORACLE_HOME (10.1.2):

This patch was downloaded for the version 10.1.0.5 to apply to the ORACLE_HOME 10.1.2 on the application tier using opatch on both the application nodes.

opatch apply
cd $ORACLE_HOME/forms/lib
make -f ins_forms.mk install
cd $ORACLE_HOME/reports/lib
make -f ins_reports.mk install

At this stage, the Application was started on both the nodes to see if they comes up without any issue after applying these patches.




This phase includes the below tasks.

·         Create database listener LABDB_LISTENER to the new 11gR2 RAC HOME.
·         Upgrade of EBS R12 database from 11gR1 RAC to 11gR2 RAC manually.

Create the Listener “LABDB_LISTENER” in the 11gR2 RAC database Home:

RAC Node 1 (node1.hingu.net):

Point to note here is that, the original LABDB_LISTENER has been existed in the 11gR1 RAC HOME and has been registered in the OCR of 11gR1 Clusterware which is on raw devices. Creating the same listener from 11gR2 RAC HOME did not conflict with the 11gR1 listener because the OCR location of 11gR2 GI is on +GIS_FILES and the 11gR1 LABDB_LISTENER was brought down.

The netca was invoked from the 11gR2 RAC HOME to add the clustered listener LABDB_LISTENER.

export ORACLE_HOME=/u01/app/oracle/db11201
/u01/app/oracle/db11201/bin/netca

Select the “Cluster Configuration
Select “Listener Configuration
Select “Add
Listener Name: LABDB_LISTENER
Selected Protocol: TCP
Port: 1522
Click “Finish


[oracle@node1 admin]$ pwd
/u01/app/oracle/db11201/network/admin
[oracle@node1 admin]$ ssh node1 mkdir /u01/app/oracle/db11201/network/admin/LABDB1_node1
[oracle@node1 admin]$ ssh node2 mkdir /u01/app/oracle/db11201/network/admin/LABDB2_node2
[oracle@node1 admin]$ ssh node3 mkdir /u01/app/oracle/db11201/network/admin/LABDB3_node3
[oracle@node1 admin]$

Added the TNS_ADMIN entry in the /u01/app/oracle/db11201/bin/racgwrap (11gR2 RAC HOME) file right under the ORACLE_HOME entry.

node1:/u01/app/oracle/db11201/bin/racgwrap

ORACLE_HOME=/u01/app/oracle/db11201
export ORACLE_HOME
export TNS_ADMIN=/u01/app/oracle/dn11201/network/admin/LABDB1_node1

node2:/u01/app/oracle/db11201/bin/racgwrap

ORACLE_HOME=/u01/app/oracle/db11201
export ORACLE_HOME
export TNS_ADMIN=/u01/app/oracle/dn11201/network/admin/LABDB2_node2

node3:/u01/app/oracle/db11201/bin/racgwrap

ORACLE_HOME=/u01/app/oracle/db11201
export ORACLE_HOME
export TNS_ADMIN=/u01/app/oracle/dn11201/network/admin/LABDB3_node3




Copied the tnsnames.ora, listener.ora and their ifiles from the /u01/app/oracle/db11201/network/admin to the above created new TNS_ADMIN


From node1:

ssh node1 cp /u01/app/oracle/db11201/network/admin/tnsnames.ora /u01/app/oracle/db11201/network/admin/LABDB1_node1
ssh node2 cp /u01/app/oracle/db11201/network/admin/tnsnames.ora /u01/app/oracle/db11201/network/admin/LABDB2_node2
ssh node3 cp /u01/app/oracle/db11201/network/admin/tnsnames.ora /u01/app/oracle/db11201/network/admin/LABDB3_node3

ssh node1 cp /u01/app/oracle/db11201/network/admin/LABDB1_node1_ifile.ora /u01/app/oracle/db11201/network/admin/LABDB1_node1
ssh node2 cp /u01/app/oracle/db11201/network/admin/LABDB2_node2_ifile.ora /u01/app/oracle/db11201/network/admin/LABDB2_node2
ssh node3 cp /u01/app/oracle/db11201/network/admin/LABDB3_node3_ifile.ora /u01/app/oracle/db11201/network/admin/LABDB3_node3

ssh node1 cp /u01/app/oracle/db11201/network/admin/listener.ora /u01/app/oracle/db11201/network/admin/LABDB1_node1
ssh node2 cp /u01/app/oracle/db11201/network/admin/listener.ora /u01/app/oracle/db11201/network/admin/LABDB2_node2
ssh node3 cp /u01/app/oracle/db11201/network/admin/listener.ora /u01/app/oracle/db11201/network/admin/LABDB3_node3


Upgrade the database from 11gR1 RAC to 11gR2 RAC:

The below steps were executed in sequence to upgrade 11gR1 RAC EBS R12 database to 11gR2 RAC.

1.       Copied the instance specific init files and password files from 11gR1 HOME to 11gR2 RAC Home.

From node1:

ssh node1 cp /u01/app/oracle/EBSR12/dbs/initLABDB1.ora /u01/app/oracle/db11201/dbs
ssh node2 cp /u01/app/oracle/EBSR12/dbs/initLABDB2.ora /u01/app/oracle/db11201/dbs
ssh node3 cp /u01/app/oracle/EBSR12/dbs/initLABDB3.ora /u01/app/oracle/db11201/dbs

2.       Created the pfile from the spfile.
3.       Removed the obsolete parameters from the pfile (plsql_native_library_dir and plsql_native_library_subdir_count).
4.       Modified the below parameters in the pfile.

create pfile='/u01/app/oracle/db11201/dbs/pfile_before_upgrade.txt' from spfile='+DATA/LABDB/spfileLABDB.ora'

cluster_database=false
java_pool_size=67108864
LABDB1.diagnostic_dest='/u01/app/oracle/db11201/admin/LABDB1_node1'
LABDB2.diagnostic_dest='/u01/app/oracle/db11201/admin/LABDB2_node2'
LABDB3.diagnostic_dest='/u01/app/oracle/db11201/admin/LABDB3_node3'
LABDB1.utl_file_dir='/usr/tmp','/usr/tmp','/u01/app/oracle/db11201/appsutil/outbound/LABDB1_node1','/usr/tmp'
LABDB2.utl_file_dir='/usr/tmp','/usr/tmp','/u01/app/oracle/db11201/appsutil/outbound/LABDB2_node2','/usr/tmp'
LABDB3.utl_file_dir='/usr/tmp','/usr/tmp','/u01/app/oracle/db11201/appsutil/outbound/LABDB3_node3','/usr/tmp'

5.       Made sure that the SYSAUX tablespace had minimum of 538MB size.
6.       Purged the recycle bin using PURGE DBA_RECYCLEBIN
7.       Ran the /u01/app/oracle/db11201/rdbms/admin/utlu112i.sql

@/u01/app/oracle/db11201/rdbms/admin/utlu112i

01.JPG

8.       Ignored the warning regarding stale statistics.
9.       Ignored the warning regarding timezone file greater than version 11 because the new 11gR2 HOME has been patched with the same DST level in the section of “Prepare database for the upgrade (DB Tier)”.
10.   Updated the /etc/oratab file and modify the LABDB home to 11gR2 HOME (/u01/app/oracle/db11201).
11.   Stopped the database LABDB.
12.   Set the environment to the 11gR2 RAC HOME.
13.   Started the instance LABDB1 on node1 using UPGRADE option of start command from 11gR2 HOME using the pfile
14.   Ran the catupgrd.sql from 11gR2 HOME (/u01/app/oracle/db11201/catupgrd.sql).

export ORACLE_HOME=/u01/app/oracle/db11201
export ORACLE_SID=LABDB1
export PATH=$PATH:$ORACLE_HOME/bin:.
sqlplus / as sysdba
startup upgrade
@/u01/app/oracle/db11201/rdbms/admin/catupgrd

03.JPG

15.   Updated the cluster_database to true in pfile.
16.   Started the database instance on node1 after the upgrade.

Post-Upgrade Steps:

1.       Ran the utlu112s.sql to get the status of the upgrade of each component.
2.       Ran the catuppst.sql to perform upgrade actions that do not require database in UPGRADE mode.
3.       Ran the utlrp.sql to recompile any INVALID objects
4.       Created the spfile from the pfile

@/u01/app/oracle/db11201/rdbms/admin/utlu112s
@/u01/app/oracle/db11201/rdbms/admin/catuppst
@/u01/app/oracle/db11201/rdbms/admin/utlrp


I had to run the catupgrd.sql script 2 times because the first time it failed due to not having enough space in TEMP tablespace. Due to this, running the catupgrd.sql 2nd time was trying to insert the same records in this table failed with the unique constraint violation errors. Based on the Metalink note (ID 1240775.1), these errors are expected when catupgrd.sql runs more than once and can safely be ignored and has no impact on the upgrade process at all.

ORA-00001: unique constraint (SYS.I_STMT_AUDIT_OPTION_MAP) violated

5.       Stopped the database LABDB.
6.       Added the database LABDB in the OCR.
7.       Started the listener LABDB_LISTENER on all the RAC Nodes.
8.       Started the database LABDB all the RAC nodes.

sqlplus / as sysdba
shutdown immediate
srvctl add database -d LABDB -m hingu.net -o /u01/app/oracle/db11201 -p +DATA/LABDB/spfileLABDB.ora -y AUTOMATIC -a DATA,FRA
srvctl add instance -d LABDB -i LABDB1 -n node1
srvctl add instance -d LABDB -i LABDB2 -n node2
srvctl add instance -d LABDB -i LABDB3 -n node3
srvctl start listener –l LABDB_LISTENER
srvctl start database –d LABDB



9.       Re-create the Custom Database Link (if any).
10.   Ran the adgrants.sql
11.   Grant “create procedure” to the CTXSYS schema.
12.   Set CTXSYS parameter.
13.   Validate Workflow ruleset.


From appsnode1:

scp $APPL_TOP/admin/adgrants.sql oracle@node1:
scp $AD_TOP/patch/115/sql/adctxprv.sql oracle@node1:

from node1:

export ORACLE_HOME=/u01/app/oracle/db11201
export ORACLE_SID=LABDB1
export PATH=$PATH:$ORACLE_HOME/bin:.
export TNS_ADMIN=/u01/app/oracle/db11201/network/admin/LABDB1_node1
sqlplus / as sysdba
@adgrants APPS
@adctxprv manager CTXSYS
exec ctxsys.ctx_adm.set_parameter('file_access_role', 'public');

From appsnode1:

sqlplus apps/apps @$FND_TOP/patch/115/sql/wfaqupfix.sql APPLSYS APPS


At this stage, with the old (original) configuration, I started the application on appsnode1 and I was successfully able to login and open the forms with no issue. The database LOCAL_LISTENER and REMOTE_LISTENER was still the same as nothing was changed from connection standpoint except the listener was then running out from 11gR2 RAC HOME instead of 11gR1 RAC.  No SCAN was used at this stage.


·         Setup appsutil directory under /u01/app/oracle/db11201 HOME.
·         Verify the TNS connection to each of the Individual RAC Instances
·         Create nls/data/9idata directory
·         Delete the FND Node Configuration from the Database.
·         Build XML on each RAC node of the DB Tier.
·         Run AutoConfig on each RAC node of the DB Tier.

Setup appsutil Directory on All the RAC Nodes.

(1)    On the application admin node, appsnode1, the appsutil.zip was generated and copied over to all the DB Tier nodes (node1, node2 and node3) under the /u01/app/oracle/db11201.

Generate the appsutil.zip

[applmgr@appsnode1 ~]$ $AD_TOP/bin/admkappsutil.pl
Starting the generation of appsutil.zip
Log file located at /u01/app/oracle/ebsR12/inst/apps/LABDB_appsnode1/admin/log/MakeAppsUtil_11112117.log
output located at /u01/app/oracle/ebsR12/inst/apps/LABDB_appsnode1/admin/out/appsutil.zip
MakeAppsUtil completed successfully.

copy to all the RAC Nodes FROM appsnode1:

scp /u01/app/oracle/ebsR12/inst/apps/LABDB_appsnode1/admin/out/appsutil.zip oracle@node1:/u01/app/oracle/db11201
scp /u01/app/oracle/ebsR12/inst/apps/LABDB_appsnode1/admin/out/appsutil.zip oracle@node2:/u01/app/oracle/db11201
scp /u01/app/oracle/ebsR12/inst/apps/LABDB_appsnode1/admin/out/appsutil.zip oracle@node3:/u01/app/oracle/db11201

unzip the appsutil.zip on all the RAC Nodes:

cd /u01/app/oracle/db11201
unzip –o appsutil.zip


(2)    The jre directory was copied from the 11GR1 HOME (/u01/app/oracle/EBSR12/appsutil/jre) to the 11gR2 RAC HOME (/u01/app/oracle/db11201/appsutil) on all the nodes.

ssh node1 cp –r /u01/app/oracle/EBSR12/appsutil/jre /u01/app/oracle/db11201/appsutil
ssh node2 cp –r /u01/app/oracle/EBSR12/appsutil/jre /u01/app/oracle/db11201/appsutil
ssh node3 cp –r /u01/app/oracle/EBSR12/appsutil/jre /u01/app/oracle/db11201/appsutil

The Instance Connectivity from each of the RAC instances was verified using SQLPLUS from RAC nodes individually. I would not run adconfig.pl without having successful connection to each of the RAC instances.

RAC Instance Connections Verification:

The Instance Connectivity from each of the RAC instances was verified using SQLPLUS from RAC nodes individually. The connectivity was confirmed before running the adconfig.pl on each of the RAC instances.

node1:

export ORA_NLS10=/u01/app/oracle/db11201/nls/data/9idata
export ORACLE_HOME=/u01/app/oracle/db11201
export ORACLE_SID=LABDB1
export TNS_ADMIN=/u01/app/oracle/db11201/network/admin/LABDB1_node1
export LD_LIBRARY_PATH=/u01/app/oracle/db11201/lib:/u01/app/oracle/db11201/ctx/lib
export PATH=$ORACLE_HOME/bin:$PATH:.

sqlplus apps/apps@LABDB1

Repeated the above process for all the RAC Instances.

Create the nls/data/9idata Directory

On all the RAC Nodes, ran the /u01/app/oracle/db11201/nls/data/old/cr9idata.pl script to create the /u01/app/oracle/db11201/nls/data/9idata directory. Without this directory, the adconfig.sh on DB tier may fail on this.

[oracle@node1 db11201]$ cd /u01/app/oracle/db11201/nls/data/old/
[oracle@node1 old]$ perl cr9idata.pl
Creating directory /u01/app/oracle/db11201/nls/data/9idata ...
Copying files to /u01/app/oracle/db11201/nls/data/9idata...
Copy finished.
Please reset environment variable ORA_NLS10 to /u01/app/oracle/db11201/nls/data/9idata!

Clean the current Configuration in the Database

SQL> exec fnd_conc_clone.setup_clean;

Build XML (on All the RAC Nodes):

Ran the /u01/app/oracle/db11201/appsutil/bin/adbldxml.pl to build the XML on each of the RAC Nodes and provided the appropriate values of DB Host Name, Database Port, SID, and Service Name. This will generate the context file <SID>_hostname.xml under /u01/app/oracle/db11201/appsutil.

On node1:

[oracle@node1 bin]$ adbldxml.pl

Starting context file generation for db tier..
Using JVM from /u01/app/oracle/db11201/jdk/jre/bin/java to execute java programs..
APPS Password: apps

The log file for this adbldxml session is located at:
/u01/app/oracle/db11201/appsutil/log/adbldxml_11112333.log
AC-20010: Error: File - listener.ora could not be found at the location:
       /listener.ora
indicated by TNS_ADMIN. Context file can not be generated.

Could not Connect to the Database with the above parameters, Please answer the Questions below


Enter Hostname of Database server: node1

Enter Port of Database server: 1522

Enter SID of Database server: LABDB1

Enter Database Service Name: LABDB

Enter the value for Display Variable: 1

The context file has been created at:
/u01/app/oracle/db11201/appsutil/LABDB1_node1.xml

Once the context file got generated, the below variables were modified (in RED) to their appropriate values.


On node1’s xml file:

<!-- host information -->
   <oa_host>
      <host oa_var="s_hostname">node3</host>
      <host oa_var="s_virtual_hostname">node1-vip</host>
      <private_address oa_var="s_clusterInterConnects">node1-prv</private_address>
      <domain oa_var="s_domainname">hingu.net</domain>
...
...
    <DB_LISTENER oa_var="s_db_listener">LABDB_LISTENER</DB_LISTENER>


Repeated the same process on all the remaining nodes and modified the XML files as shown above with their node specific values.

Run autoconfig on all the RAC nodes one by one:

Ran the Autoconfig on all the RAC nodes separately using the script /u01/app/oracle/db11201/appsutil/bin/adconfig.sh.

On node1:

export ORA_NLS10=/u01/app/oracle/db11201/nls/data/9idata
export ORACLE_HOME=/u01/app/oracle/db11201
export ORACLE_SID=LABDB1
export TNS_ADMIN=/u01/app/oracle/db11201/network/admin/LABDB1_node1
export LD_LIBRARY_PATH=/u01/app/oracle/db11201/lib:/u01/app/oracle/db11201/ctx/lib
export PATH=$ORACLE_HOME/bin:$PATH:.

cd /u01/app/oracle/db11201/appsutil/bin
./adconfig.sh


Similarly, the AutoConfig was ran on the remaining RAC nodes after setting the node specific environment variable.

Then, the autoconfig was again ran on all the RAC Nodes in the reverse order so that all the nodes’ tnsnames.ora file contains the same information.

/u01/app/oracle/db11201/appsutil/scripts/LABDB3_node3/adautocfg.sh  Ăź on node3
/u01/app/oracle/db11201/appsutil/scripts/LABDB2_node2/adautocfg.sh  Ăź on node2
/u01/app/oracle/db11201/appsutil/scripts/LABDB1_node1/adautocfg.sh  Ăź on node1

At the end, the listener ifile was created on all the RAC Nodes by copying the listener.ora content to the ifile.

node1:

ssh node1 cp /u01/app/oracle/db11201/network/admin/LABDB1_node1/listener.ora /u01/app/oracle/db11201/network/admin/LABDB1_node1/listener_ifile.ora
ssh node2 cp /u01/app/oracle/db11201/network/admin/LABDB2_node2/listener.ora /u01/app/oracle/db11201/network/admin/LABDB2_node2/listener_ifile.ora
ssh node3 cp /u01/app/oracle/db11201/network/admin/LABDB3_node3/listener.ora /u01/app/oracle/db11201/network/admin/LABDB3_node3/listener_ifile.ora


Modify the $ORACLE_HOME/dbs/<SID>_APPS_BASE.ora on all the RAC Nodes:

The <SID>_APPS_BASE.ora file under the /u01/app/oracle/db11201/dbs contains the wrong values for the control_files parameter. So, this file was modified on all the RAC nodes with the correct value of control_files parameter (select name from v$controlfile)

control_files                   = '+DATA/labdb/controlfile/current.267.765911037','+FRA/labdb/controlfile/current.263.765911059'

At this stage, the FND Nodes contained the RAC nodes information in the application database.



After enabling and running the AutoConfig successfully on the DB tier RAC nodes, it is required to run the autoconfig on all the Application nodes so that they can generate the new RAC specific TNS and jdbc_url entry. I followed the below steps to accomplish this requirement.

·         Check the connection to the database before running the AutoConfig and the result should be successful as the LISTENER port on DB Tier was remained unchanged.
·         Run the AutoConfig on the application nodes.
·         Re-run the AutoConfig on the Admin Node (appsnode1) --- reverse order.
·         Start the Application services on both the application nodes.

Run the AutoConfig on appsnode1:

$AD_TOP/bin/adconfig.sh contextfile=/u01/app/oracle/ebsR12/inst/apps/LABDB_appsnode1/appl/admin/LABDB_appsnode1.xml

After successful completion of autoconfig and re-sourcing the environment, the $TNS_ADMIN/tnsnames.ora file was containing the tns aliases for all the RAC instances individually and also had a BALANCE entry containing all the DB VIPs as shown below:

LABDB_BALANCE=
        (DESCRIPTION=
            (ADDRESS_LIST=
                (LOAD_BALANCE=YES)
                (FAILOVER=YES)
                (ADDRESS=(PROTOCOL=tcp)(HOST=node2-vip.hingu.net)(PORT=1522))
                (ADDRESS=(PROTOCOL=tcp)(HOST=node1-vip.hingu.net)(PORT=1522))
                (ADDRESS=(PROTOCOL=tcp)(HOST=node3-vip.hingu.net)(PORT=1522))
            )
            (CONNECT_DATA=
                (SERVICE_NAME=LABDB)
            )
        )


The new jdbc_url in the CONTEXT_FILE on appsnode1 was reflecting all the 3 VIPs of database RAC nodes at this stage.

Run the Autoconfig on appsnode2:

Same way, I ran the autoconfig on the appsnode2 and started the application services on both the nodes.

$AD_TOP/bin/adconfig.sh contextfile=/u01/app/oracle/ebsR12/inst/apps/LABDB_appsnode2/appl/admin/LABDB_appsnode2.xml

Re-Run the AutoConfig on appsnode2 and appsnode1:

I re-ran the autoconfig on these nodes using the below command.

$INST_TOP/admin/scripts/adautocfg.sh

Tested the login page and verified the Nodes information in the OAM as well as database.

Start the application services on appsnode2 and appsnode1:

The application services were started on both the apps nodes at this stage.

$INST_TOP/admin/scripts/adstrtal.sh

The login page was successfully displayed and was able to login to the application and forms successfully.
                                                                                                                                                                                                                        
FND Nodes Configuration:

At this stage, the configuration has 5 nodes total, 3 are DB nodes and 2 are application nodes.




Here is the Screenshots of Running the AutoConfig on all RAC nodes as well as on Application Nodes after upgrading database to 11gR2 RAC.

Application Specific Database Post-Upgrade Steps.

The Metalink note 1058763.1 was followed to achieve the below tasks.

(a)    Gather statistics for the SYS schema.

From appsnode1:

scp $APPL_TOP/admin/adstats.sql oracle@node1:

from node1:

export ORACLE_HOME=/u01/app/oracle/db11201
export ORACLE_SID=LABDB1
export PATH=$PATH:$ORACLE_HOME/bin:.
export TNS_ADMIN=/u01/app/oracle/db11201/network/admin/LABDB1_node1
sqlplus / as sysdba
alter system enable restricted session;
@adstats
alter system disable restricted session;


(b)   Create Dementra Privileges.

In my case, the Dementra was not set up. But in case if there is, the Metalink Note 1091083.1 should be followed to create Dementra previliges.

(c)    Re-Create Custom Database Links.

Not applicable in my case as the listener port remained same from 11gR1 to 11gR2 after the upgrade. Otherwise, the DB links should be recreated with the proper Port information

(d)   Re-create grants and synonyms.

On the administration server node, as APPS, ran AD Administration and selected the "Recreate grants and synonyms for APPS schema" task from the Maintain Applications Database Entities menu.


(e)   Synchronize workflow.

Logged on to Oracle E-Business Suite with the "System Administrator" responsibility. Clicked Requests > Run > Single Request and the OK button. Entered the following parameters:

Request Name = Workflow Directory Services User/Role Validation
P_BatchSize = 10000
p_Check_Dangling = Yes
Add missing user/role assignments = Yes
Update WHO columns in WF tables = No
Click "OK" and "Submit".



Verified the OS processes on both the application nodes to confirm that the enabled services are started and running successfully.




Logged in to the Application using login page url.

http://appsnode1.hingu.net:8001/OA_HTML/AppsLogin



Verified that the Concurrent Managers are started and running on the Concurrent Node (appsnode2.hingu.net)




OAM Configuration:









(a)    listnener.ora, tnsnames.ora, sqlnet.ora and their ifiles from all the RAC nodes.
(b)   listnener.ora and tnsnames.ora from both the application nodes.
(c)    CONTEXT_FILE from all the 5 nodes.
(d)   Pfile of the RAC database.


Here is the Final look of the above Configuration files at the end of the successful upgrade of 2-node EBS R12 (12.1.1) Database from 3-node 11gR1 RAC to 3-node 11gR2 RAC.

 rss