SOA with Oracle 2 node RAC cluster

Hi All,
Just a simple doubt, I have successfully installed and configured SOA suite 11.1.1.3 & BAM in one wls Domain 10.3.3 in a linux box and could access all the application like BAM console, BPEL console etc .... also could see all my data-sources deployed in the Data Sources with a single node database.
1. Now I have to RE-configure this whole SOA suite with RAC (2 node database cluster) what changes or configuration needed to implement SOA suite with RAC database?
2. Do I need to create a "Multi Data Source" to configure RAC with SOA suite..?
Thanks
Sam

DB wrote:
This is regarding Oracle RAC..so if there is a specific category..please let me know..
I have installed OEL linux 5.6 as guest OS (using virtualbox) in two laptops.
I want to install 2 node oracle 10gR2 RAC with the OEL linux as OS and each laptop as one node.
Read docs and understood that there must be shared storage for oracle clusterware and oracle ASM for oracle RAC to work.
Please let me know the steps to create shared storage for oracle clusterware and oracle ASM (considering virtualbox OEL) and to configure public,private and virtual IPs.
I already have document to create 2node oracle RAC using virtualbox with two nodes on the same laptop.so please dont suggest that doc.
Thanks,
DBMay be my step by step RAC installation guide can help you somehow?
http://kamranagayev.wordpress.com/2011/04/05/step-by-step-installing-oracle-10g-rac-on-vmware/

Similar Messages

  • Automatic restart of services on a 1 node rac cluster with Clusterware

    How do we enable a service to automaticly start-up when the db starts up?
    Thanks,
    Dave

    srvctl enable service -d DBThanks for your reply M. Nauman. I researched that command and found we do have it enabled and that it only works if the database instance was previously taken down. Since the database does not go down on an Archiver Hung error as we are using FRA with an alt location, this never kicks in and brings up the service. What we are looking for something that will trigger off of when the archive logs error and switch from FRA(Flash Recovery Area) to our Alternate disk location. Or more presicely, when it goes back to a Valid status(on the FRA - after we've run an archive log backup to clear it).
    I found out from our 2 senior dba's that our other 2 node rac environment does not suffer from this problem, only the newly created 1 node rac cluster environment. The problem is we don't know what that is(a parameter on the db or cluster or what) and how do we set it?
    Anyone know?
    Thanks,
    Gib
    Message was edited by:
    Gib2008
    Message was edited by:
    Gib2008

  • All connections are connecting to 2nd node only in a 2 Node RAC Cluster

    Hello,
    I have a 10.2.0.3 database on a two node RAC Cluster with only one service configured. This service set to be preferred on both nodes.
    However, all the connections are falling on Node2 only. Any idea where to look.
    $> srvctl config service -d PSDB
    psdbsrv1 PREF: psdb1 psdb2 AVAIL:
    Thanks,
    MM

    Application is using the following connection string.
    jdbc:oracle:thin:@(DESCRIPTION =(ADDRESS = (PROTOCOL = TCP)(HOST = PQ2-PS-db-01-vip)(PORT = 1521))(ADDRESS = (PROTOCOL = TCP)(HOST = PQ2-PS-db-02-vip)(PORT = 1521)) (LOAD_BALANCE = yes) (CONNECT_DATA =(SERVER = DEDICATED)(SERVICE_NAME = PSDBSRV1)(FAILOVER_MODE =(TYPE = SELECT)(METHOD = BASIC)(RETRIES = 180)(DELAY = 5))))
    --MM                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                       

  • Multiple databases/instances on 4-node RAC Cluster including Physical Stand

    OS: Windows 2003 Server R2 X64
    DB: 10.2.0.4
    Virtualization: NONE
    Node Configuration: x64 architecture - 4-Socket Quad-Core (16 CPUs)
    Node Memory: 128GB RAM
    We are planning the following on the above-mentioned 4-node RAC cluster:
    Node 1: DB1 with instanceDB11 (Active-Active: Load-balancing & Failover)
    Node 2: DB1 with instanceDB12 (Active-Active: Load-balancing & Failover)
    Node 3: DB1 with instanceDB13 (Active-Passive: Failover only) + DB2 with instanceDB21 (Active-Active: Load-balancing & Failover) + DB3 with instanceDB31 (Active-Active: Load-balancing & Failover) + DB4 with instance41 (Active-Active: Load-balancing & Failover)
    Node 4: DB1 with instanceDB14 (Active-Passive: Failover only) + DB2 with instanceDB22 (Active-Active: Load-balancing & Failover) + DB3 with instanceDB32 (Active-Active: Load-balancing & Failover) + DB4 with instance42 (Active-Active: Load-balancing & Failover)
    Note: DB1 will be the physical primary PROD OLTP database and will be open in READ-WRITE mode 24x7x365.
    Note: DB2 will be a Physical Standby of DB1 and will be open in Read-Only mode for reporting purposes during the day-time, except for 3 hours at night when it will apply the logs.
    Note: DB3 will be a Physical Standby of a remote database DB4 (not part of this cluster) and will be mounted in Managed Recovery mode for automatic failover/switchover purposes.
    Note: DB4 will be the physical primary Data Warehouse DB.
    Note: Going to 11g is NOT an option.
    Note: Data Guard broker will be used across the board.
    Please answer/advise of the following:
    1. Is the above configuration supported and why so? If not, what are the alternatives?
    2. Is the above configuration recommended and why so? If not, what are the recommended alternatives?

    Hi,
    As far as i understand, there's nothing wrong in configuration except you need to consider below points while implementing final design.
    1. No of CPU on each servers
    2. Memory on each servers
    3. If you've RAC physical standby then apply(MRP0) will run on only one instance.
    4. Since you are configuring physical standby for on 3rd and 4th nodes of DB1 4 node cluster where DB13 and DB14 instances are used only for failver, if you've a disaster at data center or power failure in entire data center, you are losing both primary and secondary with an assumption that your primary and physical standby reside in same data center so it may not be highly available architecture. If you are going to use extended RAC for this configuration then it makes sense where Node 1 and Node 2 will reside in Datacenter A and Node 3 ,4 will reside in Datacenter B.
    Thanks,
    Keyur

  • R12 Apps with 4 node RAC and 4 apps nodes - Post Install check list

    Hi All,
    We have a new Oracle Apps R12 instalaltion with 4 node RAC and 4 Apps nodes with load balancing and external web tier being done by an outside firm. We are assigned the responsibility of checking that everything has been configured properly both on the RAC side and Apps side. I haven't worked on RAC earlier. Please let me know what all need to be checked before approving the install done by the outside firm.
    Thank you!

    Please check below metalink notes
    RAC Assurance Support Team: RAC and Oracle Clusterware Starter Kit and Best Practices (Generic) [ID 810394.1]
    Using Oracle 10g Release 2 Real Application Clusters and Automatic Storage Management with Oracle E-Business Suite Release 12 [ID 388577.1]
    Using Oracle 11g Release 2 Real Application Clusters with Oracle E-Business Suite Release 12 [ID 823587.1]
    Oracle E-Business Suite and Oracle Real Application Clusters Documentation Roadmap [ID 745759.1]
    745759.1 Oracle E-Business Suite and Oracle Real Application Clusters Documentation Roadmap
    384248.1 Sharing The Application Tier file system in Oracle E-Business Suite Release 12
    387859.1 Using AutoConfig to Manage System Configurations with Oracle E-Business Suite Release 12
    406982.1 Cloning Oracle Applications Release 12 with Rapid Clone
    240575.1 RAC on Linux Best Practices
    265633.1 Automatic Storage Management Technical Best Practices
    Loadbalancer
    note 380489.1 Using Load-Balancers with Oracle E-Business Suite Release 12
    note 727171.1 Implementing Load Balancing On Oracle E-Business Suite - Documentation For Specific Load Balancer Hardware
    note 601694.1 How To Check Session Persistence On BigIP F5 And Cisco Ace Load Balancer Appliances
    note 603325.1 Using Cisco ACE Series Application Control Engine with Oracle E-Business Suite Release 12
    Installation R12
    note 761564.1 Oracle Applications Installation and Upgrade Notes Release 12 (12.1.1) for Linux x86
    note 402310.1 Oracle Applications Installation and Upgrade notes Release 12 (12.0) for Linux (32-bit)
    note 559518.1 Cloning Oracle E-Business Suite Release 12 RAC-Enabled Systems with Rapid Clone
    note 735276.1 Interoperability notes E-Business Suite R12 with Oracle Database 11gR1
    Shared Applications Confiugurations:
    note 380483.1 Oracle E-Business Suite Release 12 Additional Configuration and Deployment Options
    note 384248.1 Sharing The Application Tier File System in Oracle E-Business Suite Release 12
    Edited by: Amigo on Dec 20, 2010 9:58 AM

  • Install mulutple RAC databases on 2-node RAC cluster

    I am installing 5 RAC databases on a 2-node RAC cluster. I have setup SCAN using 3 IP addresses.
    Do I have to use SCAN listener for all databases?
    When installing the 3 database, I get ORA-12537: TNS connection closed error.
    ENV: 11gR2 2-node RH5.x
    Thanks!

    I have setup SCAN using 3 IP addresses.
    Do I have to use SCAN listener for all databases?These 3 scan ip will work for youe all database running under this cluster setup.
    you may also use VIP to make connection like 10g.
    I get ORA-12537: TNS connection closed error.Appear some connectivity/configuration isue,please try MOS doc contain detail on this.
    How to Troubleshoot Connectivity Issue with 11gR2 SCAN Name [ID 975457.1]

  • Problem with Oracle 10g RAC VIP network setting at Solaris 9

    Dear All,
    I have tried to set up a Oracle 10g RAC Release 2.
    With OS solaris 9, and 2 nodes.
    The nodes setting as the following:
    nodes 1:
    Public address: 172.16.0.121
    Private address: 192.168.0.121, 192.168.1.121 (dual path for heart beat)
    nodes 2:
    Public address: 172.16.0.122
    Private address: 192.168.0.122, 192.168.1.122
    And i have assigned two IP adress 172.16.0.131, 172.16.0.132 as the VIP address for the
    RAC.
    And the following is the /etc/hosts file:
    root@shk01 # cat /etc/hosts
    # Internet host table
    127.0.0.1 localhost
    # public address
    172.16.0.121 shk01 loghost
    172.16.0.122 shk02
    # heart beat
    192.168.0.121 shk01-priv1
    192.168.1.121 shk01-priv2
    192.168.0.122 shk02-priv1
    192.168.1.122 shk02-priv2
    # VIP address
    172.16.1.131 vip-shk01
    172.16.1.132 vip-shk02
    But when i run the command "$ ./runcluvfy.sh comp nodecon -n shk01,shk02 -verbose"
    it shows the error:
    ERROR:
    Could not find a suitable set of interfaces for VIPs.
    Result: Node connectivity check failed.
    I did try to add the VIP address on bge0:1, as i using bge0 for the public address.
    Both nodes i did using the same interface name for it.
    Anyone have idea for me to check out the error?
    Also, I have other question about the raw device.
    As there is option for setting for ASM or raw device. If choosing raw device, does it mean that it just need to
    format the storage disk but without newfs it? And then the Oracle program will able to handle it?
    Thanks,
    Xentar

    You don't seem to state categorically that you are using Solaris Cluster, so I'll assume it since this is mainly a forum about Solaris Cluster (and IMHO, Solaris Cluster with Clusterware is better than Clusterware on its own).
    Clusterware has to see the same device names from all cluster nodes. This is why Solaris Cluster (SC) is a positive benefit over Clusterware because SC provides an automatically managed, consistent name space. Clusterware on its own forces you to manage either the symbolic links (or worse mknods) to create a consistent namespace!
    So, given the SC consistent namespace you simple add the raw devices into the ASM configuration, i.e. /dev/did/rdsk/dXsY. If you are using Solaris Volume Manager, you would use /dev/md/<setname>/rdsk/dXXX and if you were using CVM/VxVM you would use /dev/vx/rdsk/<dg_name>/<dev_name>.
    Of course, if you genuinely are using Clusterware on its own, then you have somewhat of a management issue! ... time to think about installing SC?
    Tim
    ---

  • Solaris 10 and Hitachi LUN mapping with Oracle 10g RAC and ASM?

    Hi all,
    I am working on an Oracle 10g RAC and ASM installation with Sun E6900 servers attached to a Hitachi SAN for shared storage with Sun Solaris 10 as the server OS. We are using Oracle 10g Release 2 (10.2.0.3) RAC clusterware
    for the clustering software and raw devices for shared storage and Veritas VxFs 4.1 filesystem.
    My question is this:
    How do I map the raw devices and LUNs on the Hitachi SAN to Solaris 10 OS and Oracle 10g RAC ASM?
    I am aware that with an Oracle 10g RAC and ASM instance, one needs to configure the ASM instance initialization parameter file to set the asm_diskstring setting to recognize the LUNs that are presented to the host.
    I know that Sun Solaris 10 uses /dev/rdsk/CwTxDySz naming convention at the OS level for disks. However, how would I map this to Oracle 10g ASM settings?
    I cannot find this critical piece of information ANYWHERE!!!!
    Thanks for your help!

    You don't seem to state categorically that you are using Solaris Cluster, so I'll assume it since this is mainly a forum about Solaris Cluster (and IMHO, Solaris Cluster with Clusterware is better than Clusterware on its own).
    Clusterware has to see the same device names from all cluster nodes. This is why Solaris Cluster (SC) is a positive benefit over Clusterware because SC provides an automatically managed, consistent name space. Clusterware on its own forces you to manage either the symbolic links (or worse mknods) to create a consistent namespace!
    So, given the SC consistent namespace you simple add the raw devices into the ASM configuration, i.e. /dev/did/rdsk/dXsY. If you are using Solaris Volume Manager, you would use /dev/md/<setname>/rdsk/dXXX and if you were using CVM/VxVM you would use /dev/vx/rdsk/<dg_name>/<dev_name>.
    Of course, if you genuinely are using Clusterware on its own, then you have somewhat of a management issue! ... time to think about installing SC?
    Tim
    ---

  • Routing all connections through a one node in a 2 node RAC cluster

    Hi everyone
    My client has the following requirement: an active/active RAC cluster (eg node1/node2), but with only one of the nodes being used (node1) and the other sitting there just in case.
    For things like services, i'm sure this is straightforward enough - just have them set to preferred on node1 and available on node 2.
    For connections, I imagine I would just have the vips in order in the tns file, but with LOAD_BALANCING=OFF, so they go through the tns entries in order (i.e node 1, then node 2), so this would still allow the vip to failover if node 1 is down.
    Does that sound about right? Have I missed anything?
    Many thanks
    Rup

    user573914 wrote:
    My client has the following requirement: an active/active RAC cluster (eg node1/node2), but with only one of the nodes being used (node1) and the other sitting there just in case.Why? What is the reason for a "+just in case+" node - and when and how is is "enabled" when that just-in-case situation occurs?
    This does not many any kind of sense from a high availability or redundancy view.
    For connections, I imagine I would just have the vips in order in the tns file, but with LOAD_BALANCING=OFF, so they go through the tns entries in order (i.e node 1, then node 2), so this would still allow the vip to failover if node 1 is down.
    Does that sound about right? Have I missed anything?Won't work on 10g - may not work on 11g. The Listener can and does handoff connections, depending on what the TNS connection string say. If you do not connect via a SID entry but via a SERVICE entry, and that service is available on multiple nodes, you may not (and often will not) be connected to instance on the single IP that you used in your TNS connection.
    Basic example:
    // note that this TEST-RAC alias refers to a single specific IP of a cluster, and use
    // SERVICE_NAME as the request
    /home/billy> tnsping test-rac
    TNS Ping Utility for Linux: Version 10.2.0.1.0 - Production on 18-JAN-2011 09:06:33
    Copyright (c) 1997, 2005, Oracle.  All rights reserved.
    Used parameter files:
    /usr/lib/oracle/xe/app/oracle/product/10.2.0/server/network/admin/sqlnet.ora
    Used TNSNAMES adapter to resolve the alias
    Attempting to contact (DESCRIPTION = (ADDRESS=(PROTOCOL=TCP)(HOST= 196.1.83.116)(PORT=1521)) (LOAD_BALANCE=no) (CONNECT_DATA=(SERVER=shared)(SERVICE_NAME=myservicename)))
    OK (50 msec)
    // now connecting to the cluster using this TEST-RAC TNS alias - and despite we listing a single
    // IP in our TNS connection, we are handed off to a different RAC node (as the service is available
    // on all nodes)
    // and this also happens despite our TNS connection explicitly requesting no load balancing
    /home/billy> sqlplus scott/tiger@test-rac
    SQL*Plus: Release 10.2.0.1.0 - Production on Tue Jan 18 09:06:38 2011
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to:
    Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bit Production
    With the Partitioning, Real Application Clusters, Data Mining and Real Application Testing options
    SQL> !lsof -n -p $PPID | grep TCP
    sqlplus 5432 billy    8u  IPv4 2199967      0t0     TCP 10.251.93.58:33220->196.1.83.127:37031 (ESTABLISHED)
    SQL> So we connected to RAC node 196.1.83.116 - and that listener handed us off to RAC node 196.1.83.127. The 11gr2 Listener seems to behave differently - it does not do a handoff (from a quick test I did on a 11.2.0.1 RAC) in the above scenario.
    This issue aside - how do you deal with just-in-case situation? How do you get clients to connect to node 2 when node 1 is down? Do you rely on the virtual IP of node 1 to be switched to node 2? Is this a 100% safe and guaranteed method?
    It can take some time (minutes, perhaps more) for a virtual IP address to fail over to another node. During that time, any client connection using that virtual IP will fail. Is this acceptable?
    I dunno - I dislike this concept of your client of treating the one RAC node as some kind of standby database for a just-in-case situation. I fail to see any logic in that approach.

  • Clean way to De-register/De-commission the database with EBS from RAC/Cluster.

    Hi,
    We have a database with EBS on RAC. We are looking for Clean way to decommissioning the database from the cluster. We tried earlier, but when we again installed a fresh database on that server some how we had issues with the user oracle in cluster configuration files.
    Early response is highly appreciated.
    Thanks,
    Rajashekar

    Hi Hussein,
    We have two node db and two node application server in a cluster environment.
    We want to Remove/Decommission JUST DATABASE, APPLICATION related all files from the cluster. Cluster files should not be effected. After removing them we want install a fresh Db and Application on the same cluster.
    Looking for any standard method for decommissioning.
    Thanks,
    Rajashekar

  • Theoretical question on having "stand by" instance on 4 node RAC cluster.

    We have been having discussions on how we can improve the overall stability and performance of our RAC cluster. Currently we are running multiple databases on a 2 node cluster. The idea is that we will add two more nodes to the cluster and then spread the instances across the 4 nodes but only have two active and 1 as a standby in the event of node failure.
    In other words:
    Node1, Node2, and Node3 would run DB1
    Node 4 Node3 and Node2 would run DB2
    Additionally DB1_Instance3 would be shutdown on Node3 and only started if Node1 or 2 failed.
    This would be the same for DB2 where DB2_Instance3 on Node2 would only be started if Node3 or Node4 failed.
    Underneath them all would be the same ASM database shared for all 4 nodes.
    Has anyone seen such a configuration? Is anyone aware of a white paper that discusses it.

    the parameter active_instance_count is a 9i parameter and should only be set if you want to run RAC with a 2 node cluster - one instance active (all users connected here) and the second instance up but no one actively using it. IF instance one goes away, all work is move to instance 2. This is hot failover .
    If you are adding a new instance to a rac database, you are recommended to use DBCA or EM to add the instance. If you do not want to have users connected to the instance all the time, you can either leave it running but do not enable any services on it until you need it or stop the instance with srvctl stop instance. The cluster will not try to restart the instance automatically.
    The services option, when you create a service, you define which instances you wish it to run on. You can dynamically change where it runs. For the end user, they just connect to the service, the listener directs the connection to the least loaded instance providing the service at connect time. See the workload management with RAC 10g white paper on otn (otn.oracle.com/rac) or Chapter 6 of admin & deploy guide

  • Diff hardwares for 2 node rac cluster

    Dear All,
    Can i implement RAC 2 node cluster on 2 nodes with diff hardwares like diff server models?
    Regards

    hungry_dba wrote:
    Dear All,
    Can i implement RAC 2 node cluster on 2 nodes with diff hardwares like diff server models?
    Hi,
    Also you can read tech note below:
    *RAC: Frequently Asked Questions [ID 220970.1]*
    Can I have different servers in my Oracle RAC? Can they be from different vendors? Can they be different sizes?
    Oracle RAC does support a cluster with nodes that have different hardware configurations. An example is a cluster with 3 nodes with 4 CPUs and another node with 6 CPUs. This can easily occur when adding a new node after the cluster has been in production for a while. For this type of configuration, customers must consider some additional features to get the optimal cluster performance. The servers used in the cluster can be from different vendors; this is fully supported as long as they run the same binaries. Since many customers implement Oracle RAC for high availability, you must make sure that your hardware vendor will support the configuration. If you have a failure, will you get support for the hardware configuration?
    Regards,
    Levi Pereira

  • ALBPM 5.7 BPM process admin issues with Oracle 11g RAC as directory service

    Hi,
    Setup ALBPM 5.7 enterprise for weblogic.
    BPM Engine deployed on weblogic cluster.
    BPM process admin (web console) deployed on tomcat.
    BPM process admin uses WEB-INF/directory.properties for directory configuration.
    things were fine with oracle SID,
    directory.default.url=oracle://<db_host>:1521/schema=diruser,sid=BPMDB { with this, login to process admin was working fine}
    But clueless, on how to configure RAC as the directory service in the above directory.properties file. Since this is a old version, no directory.xml in the installation.
    Connection string:
    (DESCRIPTION =
    (LOAD_BALANCE = on)
    (FAILOVER = on)
    (ADDRESS = (PROTOCOL = TCP)(HOST = DB_HOST1)(PORT=1521))
    (ADDRESS = (PROTOCOL = TCP)(HOST = DB_HOST2)(PORT=1521))
    (CONNECT_DATA =
    (SERVICE_NAME = BPMDB)
    (FAILOVER_MODE =
    (TYPE = SELECT)
    (METHOD = BASIC)
    Please help how to use service name (above connection string) in directory.properties, to be able to login to process admin successfully.
    Thanks in advance,
    Swamy

    Using admin center, able to get the string to be used in directory.properties.
    FYI, the string looks like below,
    directory.default.url=oracle://customURL:0/schema=bpmdirectory,customURL=jdbc:oracle:thin:@(DESCRIPTION = (LOAD_BALANCE = on)(FAILOVER = on)(ADDRESS = (PROTOCOL = TCP)(HOST = DB_HOST1)(PORT=1521))(ADDRESS = (PROTOCOL = TCP)(HOST = DB_HOST2)(PORT=1521)) (CONNECT_DATA = (SERVICE_NAME = BPMDB.DOMAIN.COM) (FAILOVER_MODE = (TYPE = SELECT) (METHOD = BASIC))))

  • Rolling upgrade in MAA database with 2 node RAC on each site

    All,
    I have 2 node rac on SITE A, and it has Physical Standby database on SITE B it also 2 node RAC. Currently it is on 10.2.0.4 on AIX systems. We have mission crictical application runing and we can't aford the downtime of 5 to 6 hr to upgrade our CRS,ASM,DB oracle homes to 11G or even 10.2.0.5. I read rolling upgrade document on metalink and also documnet as well, and come up with following sequence.
    SITE A==Primary Site, SITE B == Standby database
    1) Shutdown DB on SITE B
    2) Take the OS Level Backup of Binaries only on SITE B (Both node), I think this is just a Precautionary Measure
    3) Start the DB on SITE B Still Physical Standby database let it synch with primary Site SITE A
    4) Convert SITE B to Logical standby database
    5) let it run for some time and be the two site SYNCH
    6) now stop archivelog shipping to STANDBY SITE SITE B
    7) Shutdown SITE B and apply the patch/upgrade binaries (CRS/ASM/DB)
    8) Start and upgrade the DB on SITE B
    9) Start the logical standby apply (SQL APPLY) and let it catch up with SITE A
    10) Now Switchover to logical standby database on SITE B and ask application to connect to SITE B
    11) Make sure the archivelogs are moving from SITE B to SITE A and SQL Apply is runing.
    12) Now Stop archive log flow from SITE B to SITE A
    13) Stop the database/ASM/CRS on SITE A
    14) upgrade the SITE A binaries (CRS/ASM/DB)
    15) Start and upgrade the DB on SITE A
    16) Start the logical standby apply (SQL APPLY) and let it catch up with SITE B
    17) Now Switchover Back to SITE A and let the application connect to SITE A
    At this point by SITE B is a Logical Standby database
    I think these summaries the steps involved in rolling upgrade, but I can't understand how I can go back to my MAA with physical standby database mode?
    Does anyone has done this? Can anyone give me some tips or comments on my steps??
    Help me if you can... thanks in advance.

    Hi,
    Is it you want to know the OS/Kernel side setting? You must comply with the following kernel requirements to install the clusterware.
    http://docs.oracle.com/cd/B28359_01/install.111/b28259/pre_hpux.htm#BABJHCJI
    Thanks.

  • Configuring rman backup for oracle on rac cluster

    Hi GURUS
    We are buying installation oracle rac on a two node server. We do backup our db using rman and TSM .
    Does anyone have any guidance on using TSM scheduled TDP backups with
    Oracle RAC multi-node?
    More to that, is it possible to have multiple archive log destination?
    Any advice would be appreciated.
    David

    While you can have archives going to different destinations, on different nodes, it is not advisable because if these destinations are not reachable in case of a node/system failure, you have no access to these archive logs.
    Actually for easy accesibility it would be better to place the archive logs of all instnaces on shared storage. If you are using ASM and using a FRA disk group, you may want to have your archives go into this location.
    NFS mounting filesystems could be slower and may have certain perforamance implications.

Maybe you are looking for

  • Custom error message in user exit is giving Short dump

    Hi All, We have a scenario where in if any user try to create/Change/Delete particular types of contract in SAP system we have to issue error message saying ZIN/ZIR contract creation/Change/Deletion is blocked in SAP. This has to be done from the web

  • Unable to read data in Java

    Hi, I tried to use this sample problem to extract the data from a primary database. But I am not getting anything back. Could someone please help me to take a kind look at my java code see what is wrong? I am new to Java, any help will be greatly app

  • Schedule rman using cron job

    I have scheduled a cron job to take rman backup of database everyday at 4PM. But I see that the empty backup.log file is getting generated while the script is not running. If I run the backup.sh script manually then the backup is running fine. The fo

  • Website name: Lorem ipsum dolor sit amet

    Hello I have designed my website using iWeb in iLife '08 and I cannot get the name of my website to view properly. I can't get away from the above name and in search engines this latin name comes up and not my own. The page name and the site name are

  • Cannot see users with server 10.7.3 update

    I updated to 10.7.3 and when I launch the server app I see only the administrator user in the users pane (all other users do not show up). Further, the knobs for + and - are grayed (disabled). I still can log in with any network user, though. How can