Recover 2 nodes RAC to another server

Hi
I am trying to restore our 2 nodes database to another single instance server.
I successfully restore controlfile, datafile but when I try to recover the database,
rman ask for a archivelog seq from thread 2 even though the next system change # is in thread 1 archivelog.
Please refer to the v$archived_log below
RECID     THREAD#     SEQUENCE#     NEXT_CHANGE#     FIRST_CHANGE#
853     1     476     10058590635887     10058588406010
852     1     475     10058588406010     10058586932543
851     2     377     10058588406003     10058586932534
After rman recovered the log seq 475, it keep asking for the next seq 378 in thread 2 which hasn't been archived by thread 2 instance,
the next change actuall in thread# 1, seq 476.
Is that mean I will have to wait thread# 2 archive the seq 378 log before I can recover seq 476 in thread#1?
Pek

Hi,
Here is the result..
select status, checkpoint_change#, fuzzy,
to_char(checkpoint_time, 'DD-MON-YYYY HH24:MI:SS') as checkpoint_time,
count(*)
from v$datafile_header
group by status, checkpoint_change#, fuzzy, checkpoint_time
order by status, checkpoint_change#, fuzzy, checkpoint_time;
select distinct checkpoint_change# from v$datafile_header;
select hxfil file_id, fhscn scn, fhthr thread, fhrba_seq sequence, fhsta status from x$kcvfh;
select distinct fhscn from x$kcvfh;
STATUS CHECKPOINT_CHANGE# FUZ CHECKPOINT_TIME COUNT(*)
ONLINE 10058588406003 NO 25-NOV-2012 19:00:30 16
1 row selected.
SQL>
CHECKPOINT_CHANGE#
10058588406003
1 row selected.
SQL>
FILE_ID SCN THREAD SEQUENCE STATUS
1 10058588406003 2 378 8192
2 10058588406003 2 378 0
3 10058588406003 2 378 0
4 10058588406003 2 378 0
5 10058588406003 2 378 0
6 10058588406003 2 378 0
7 10058588406003 2 378 0
8 10058588406003 2 378 0
9 10058588406003 2 378 0
10 10058588406003 2 378 0
11 10058588406003 2 378 0
12 10058588406003 2 378 0
13 10058588406003 2 378 0
14 10058588406003 2 378 0
15 10058588406003 2 378 0
16 10058588406003 2 378 0
16 rows selected.
SQL>
FHSCN
10058588406003
SQL> select hxfil file_id, fhtnm tablespace_name from x$kcvfh;
FILE_ID TABLESPACE_NAME
1 SYSTEM
2 SYSAUX
3 UNDOTBS1
4 UNDOTBS2
5 USERS
6 DATA_SE
7 INDEX_SE
8 DATA_SE
9 INDEX_SE
10 DATA_SE
11 DATA_SE_LOJACK
12 INDEX_SE
13 CJHIT
14 FLOW_1
15 DATA_SE
16 INDEX_SE
16 rows selected.
Edited by: pj**** on 26-Nov-2012 16:18

Similar Messages

  • [wls8.1 sp.3] cannot remote start with node manager on another server

    In this domain run on 2 physical servers in windows 2003. In machine that has admin server can start with node manager but another machine without admin server cannot remote start . when i remote start , the node manager has facing error messages as below:
    # AT "node manager monitoring" in weblogic admin console
    [[NodeManager:300033]Could not execute command ping on the node manager. Reason: weblogic.nodemanager.NodeManagerException: [CommandInvoker: Failed to send command: 'ping to server 'null' to NodeManager at host: 'xxx.xxx.xxx.24:5555' with exception Connection reset by peer.. Please ensure that the NodeManager is active on the target machine].]
    # AT "nodemanger.log" has facing error messages as below:
    <Aug 20, 2008 7:34:07 PM ICT> <Info> <[email protected]:5555> <NodeManage r started, log messages being written into file C:\bea\WEBLOG~1\common\NODEMA~1\ NodeManagerLogs\NodeManagerInternal\nm_server02_08_20_2008-19_34_03.log>
    <Aug 20, 2008 7:59:58 PM ICT> <Warning> <Security> <BEA-090476> <Invalid/unknown SSL header was received from peer server01 - 10.208.50.23 during SSL handsha ke.> <Aug 20, 2008 8:00:11 PM ICT> <Warning> <Security> <BEA-090476> <Invalid/unknown SSL header was received from peer server01 - xxx.xxx.xxx.23 during SSL handsha key.>
    I would REALLY appreciate any suggestions as to what might be the cause or what i could try.
    Thanks ....

    Hi,
    It is up to you to decide whether you want to use nodemanager or not.
    It does not make any difference if the server is started via script or nodemanager.
    if i want everything (AS, MS, and NM) to come up at server boot i need to set each of them up as windows services (which is what i have at the moment) but this wont allow me to use the node manager ==> Yes this statement is correct.
    You can either set AS and MS as windows service (or) AS and nodemanager as service.
    Thanks,
    Sharmela

  • Setup 2-Node RAC 11.2.0.1.0 on Windows Server 2008

    I have been setting up RAC environments in VirtualBox and VMWare in my local machine and in our test server. I have done the following setup:
    1. RAC 10g on RHEL 5.4 (VMWare / VBox)
    2. RAC 11g on RHEL 5.4 (VBox)
    3. RAC 11g on Windows Server 2008 (VBox)
    Now, our management wants me to setup a 2-node RAC in real world. Cost is not an issue here as this will be financed by a big private group.
    I am excited to do the project as I am really enthusiastic in database clustering. Of course, there is a little nervous feeling since this is
    my first time doing it in real world (as the best RAC expert started from his first deployment :) ).
    I am going to build RAC 11.2.0.1.0 in Windows Server 2008
    I would like to seek advise on:
    - the best practices, what are the things that I have to consider?
    - any straight forward real world deployment guide or technical papers that can serve as my reference
    - any issues that I might encounter
    - any help and feedback
    I know I can be successful in this first time project if ill seek advise to the experts.
    I hope you can help me and I will be glad to read and review your contributions.
    Thanks a lot.....

    Hi,
    Few months ago I had a project to setup RAC on Windows 2008 R2. For six months now there are no problems and it's working fine, but I have to say that installation wasn't easy and took me a lot of time. Another thing is the troubleshooting, I feel completely helpless when (if) something screws up. It takes between 10 to 15 mins to have database running in case of reboot of the servers.
    So here are few things to consider:
    - Disable write cache on shared disks.
    - Disable User Access Control!
    - Disable firewall (really important)!
    - Use diskpart command to create extended and logical partition on all disks.
    - There is also nasty bug we hit, I've blogged about it:
    http://sve.to/2011/09/29/exhaust-of-windows-2008-heap-memory-with-oracle-database-11-2-0-2/
    - I had a terrible problems with user equivalence. Verify Privileges for copying files in the cluster:
    net use \\nodeX\c$
    - Once installation is completed apply latest Bundle Patch - currently this is BP6 (Patch 13965211):
    DB 11.2.0.3 Patch 6 includes all bugs fixed in 11.2.0.3 Patch 1 to Patch 5 and also includes CPU2012. It must be applied on top of the 11.2.0.3.
    Useful MOS notes:
    RAC and Oracle Clusterware Best Practices and Starter Kit (Windows) [ID 811271.1]
    Windows: CLUVFY Fails with TCP Check PRVF-7617 Due to Case of Node Names [ID 1286394.1]
    Finally if you have a choice then go with Linux, it's robust, easy to install and maintain. You have better control on the system and user processes, it's more flexible and easy to troubleshoot.
    Regards,
    Sve

  • 2 nodes RAC with Oracle 10g on Windows 2000 server

    Hi at all. I want to manage a cluster having 2 nodes on windows 2000 server each of them with Oracle 10g. I have VMWare 5.0. Someone can suggest me a detailed tutorial how to start and built it? Thank you very much for your answers.
    Saverio Sergio Mola

    the cookbooks here:
    http://www.oracle.com/technology/tech/linux/vmware/cookbook/index.html
    will explain how to install.. however the VMware Workstation product doesn't offer in a supportedway the sharing of two nodes of disks.. there are hacks out there, but they are not very stable. We therefor chose to install two instances on one node to showcase RAC, rather than make it complicated and unstable to use and showcase.
    there are also ready made VM's from above page, so u can quickly evaluate it for both Red HAt and SuSE.. when vmware can support such sharing in an easy supported way we might re-evaluate.. until then my personal suggestion is to stick to simple case.. 2 instances on one node.. its easier to copy, snapshot.. etc.
    Enjoy..
    Saar.

  • RAC and RAC One Node on the same server?

    Hello all,
    Anyone know if having RAC and RAC One Node on the same server is possible, supported, and recommended? Our architecture group is moving towards this configuration as a standard, and I need solid direction about feasability.
    Thanks for any help you can provide.
    Jack Silvey| CVS Caremark | Executive Advisor, IT Systems | CVS Caremark

    Yes you can have the RAC and RAC one on the same cluster. For support and recommendations raise a service request with oracle support  

  • Multiple copies of same database on a 2 node RAC server - How to merge ?

    I currently have multiple copies of the same database running on a 2 node Rac system. I am looking for a way to combine them into 1 large database but keeping the data separate.
    The databases are copies of production for testing, development and a yearly "historical" databases .
    All the databases are created from production, and generally have the same schema's , tables, procedures, etc however may be different versions and need to be.
    Is There a way to use one large database and logically split all the different versions of the same objects into their own space in one database ? The structure cannot change as the database is for a 3rd party's Forms application the relies on the objects not changing names etc.
    Ideally I am looking for a solution that will allow the forms application to connect to "test" and "historical" copies of our production database separately in the same database container.
    Thanks for any direction.

    I currently have multiple copies of the same database running on a 2 node Rac system. I am looking for a way to combine them into 1 large database but keeping the data separate.
    The databases are copies of production for testing, development and a yearly "historical" databases .
    All the databases are created from production, and generally have the same schema's , tables, procedures, etc however may be different versions and need to be.
    Is There a way to use one large database and logically split all the different versions of the same objects into their own space in one database ? The structure cannot change as the database is for a 3rd party's Forms application the relies on the objects not changing names etc.
    Ideally I am looking for a solution that will allow the forms application to connect to "test" and "historical" copies of our production database separately in the same database container.
    Thanks for any direction.

  • Mutliple serveres in a two node RAC

    hello friends m i newbie in RAC and i have 7 servers in my company 4 are for oracle database used for different purposes having same instance name on each of them and 3 are for application servers used for different purposes having same instance name and i want to use two node RAC for all of them now my question is
    1) can i use all the 7 servers (4 database and 3 application) together in a two node RAC?
    2) in RAC can i have different instances having same name?
    thanx in advance

    saugat chatterjee wrote:
    hello friends m i newbie in RAC and i have 7 servers in my company 4 are for oracle database used for different purposes having same instance name on each of them and 3 are for application servers used for different purposes having same instance name and i want to use two node RAC for all of them now my question isRAC is a single database running across multiple servers (db instance on each). That is how Oracle scales.
    Oracle does not scale the converse way - running multiple databases on a single server.
    1) can i use all the 7 servers (4 database and 3 application) together in a two node RAC?That depends entirely on whether the 2 severs have the capacity to deal with the processing load currently handled by the database servers.
    Also note that in Oracle the terms "+database+" and "+schema+" differ from what are used by some other RDBMS products.
    An Oracle database is a physical entity. An Oracle schema is a (amongst other things) a logical database. So a single physical Oracle database can contain 1000's of logical databases - and these logical databases can be physically separated from one another (via tablespaces), can have different rights, can have different resource profiles, etc.
    Each RAC node is part of a shared everything cluster - meaning that the entire physical database is available on each and every RAC node.
    On RAC, you can do what is called application partitioning. You can for example configure HR clients to make use of the HR logical database (schema) on RAC node 1. RAC node 3 is for example used for the Data Mart application and node 2 handles the customer billing clients.
    In other words, you partition your applications across RAC nodes.
    2) in RAC can i have different instances having same name? The instance name is a physical name and should very seldom be explicitly used. Your physical RAC database's SID may be PROD. Your 4 RAC instances, one on each server, will have SIDs PROD1 to PROD4.
    But you should instead create database services and have clients connect (via the RAC's SCAN/Single Client Access Name) to a specific service.
    Bottom line to keep in mind - RAC is Oracle's answer to scalability. And RAC is multiple servers and a single database. Doing the converse and running multiple database on a single server (RAC or no RAC), very seldom makes any technical sense. It almost always will reduce both performance and scalability.

  • How to apply archivelogs from 2 node rac to non-rac

    Is there any way to apply the archivelogs from 2 node RAC to sinlge node non-rac instance while migrating RAC to non-rac.
    Could anyone clarify it?
    -Shanlee

    There's no "different way".
    Each instance in RAC write's to it's own Redo Thread -- so each has it's own set of ArchiveLogs. In the log_archive_format specification %t is the thread identifier.
    Oracle needs the ArchiveLogs from both Threads of the 2 node RAC. When you issue a RECOVER DATABASE, it applies them in the correct (interleaved) sequence.
    What you have to do is to ensure that the ArchiveLogs of both threads are available on the new server.
    Hemant K Chitale
    Edited by: Hemant K Chitale on Jun 1, 2011 5:09 PM

  • 4 node RAC and services

    Hi All,
    We currently run 5 databases on a 5 node RAC.. 11.2.0.2 GI and a mix of 11.2 and 10.2 databases. Obviously, services are defined for use by the application based on load on each node..
    I am seeing something strange..
    CLV1P_COMMON is a service created to run on CLV1P2 instance but during my tests over the weekend, the service was coming up on different node each time.. Why is this behavior ? Is this expected ? My understanding was that the service should run only on the node it services.. and in the case, where you specifiy PREFFERRED to more than one instance, it uses them randomly.. ( load balancing )
    Can anyone guide me please ?
    Thanks
    /oracle/dba/output/STR1P/rman
    camgmolpdb01 [CLV1P1]-> srvctl status service -d CLV1P
    Service CLV1P1.aaaaaaaaaaaa.com is running on instance(s) CLV1P1
    Service CLV1P2.aaaaaaaaaaaa.com is running on instance(s) CLV1P2
    Service CLV1P3.aaaaaaaaaaaa.com is running on instance(s) CLV1P3
    Service CLV1P4.aaaaaaaaaaaa.com is running on instance(s) CLV1P4
    Service CLV1P_COMMON.aaaaaaaaaaaa.com is running on instance(s) CLV1P2
    /oracle/dba/output/STR1P/rman
    camgmolpdb01 [CLV1P1]-> srvctl config service -d CLV1P -s CLV1P_COMMON.aaaaaaaaaaaa.com
    Service name: CLV1P_COMMON.aaaaaaaaaaaa.com
    Service is enabled
    Server pool: CLV1P_CLV1P_COMMON.aaaaaaaaaaaa.com
    Cardinality: 1
    Disconnect: false
    Service role: PRIMARY
    Management policy: AUTOMATIC
    DTP transaction: false
    AQ HA notifications: false
    Failover type: NONE
    Failover method: NONE
    TAF failover retries: 0
    TAF failover delay: 0
    Connection Load Balancing Goal: LONG
    Runtime Load Balancing Goal: NONE
    TAF policy specification: NONE
    Edition:
    Preferred instances: CLV1P2
    Available instances: CLV1P1,CLV1P3,CLV1P4
    This is the command I used to create the service
    srvctl add service -d CLV1P -s CLV1P_COMMON.aaaaaaaaaaaaaa.com -r "CLV1P2" -a "CLV1P1,CLV1P3,CLV1P4" -P BASIC -l primary -y AUTOMATIC -e SELECT -m BASICThanks in advance
    Edited by: USER101 on Oct 15, 2012 1:59 PM
    Edited by: USER101 on Oct 15, 2012 2:01 PM

    Hi,
    preferred and available is normally only evaluated during startup of the service.
    So if you started you server in the order 1 then 2,3,4 as a result the service will get started on node 1 and not on the preferred node 2. Reason: Service is defined that it should start at least one instance. At the moment of startup there was only server 1 active, hence it is starting there. If node 2 (preferred) started after that, it is not evaluated, since Oracle does not want to move a service simply because of startup.
    If then node 1 is rebooted, the service will get failed over. to the pereferred instance. If you then stop instance on node 2 again it will be started on a different node (not fix).
    Important: The service does not fail to preferred instance, only if it is up.
    As a workaround (if you need the service always to be started on node 2 if available and only run on another instance, when node 2 is not up) you can do this with a FAN Callout script.
    Regards
    Sebastian

  • Restore procedures of 2 node RAC

    I have 2 node RAC running 11gr2 (11.2.0.2), linux RHEL 5.x in my test server using Oracle VM. This set up has 5 diskgroups in ASM (ocrvote1, ocrvote2, ocrvote3), data, fra. Both data and fra are in externally redundancy and ocrvote1-3 in normal redundancy. As i'm getting ready to test the recovery procedures of the whole setup, I wanted to check here to see, if i have covered the necessary bits and not missing anything obvious.
    I will be taking full backup of database and md_backup of disk groups prior to testing these procedures. Below is the list of procedures i had come up, for getting the system restored in shortest time frame. If you think i can cut-down the time even further, would appreciate your inputs.
    -- Install OS via jump start on all nodes.
    -- Partition disks same set of disk and give similar names as before.
    -- Stamp disks with asmlib and present them to all nodes
    -- Install GI and RDBMS homes
    -- Bring up ASM with OCRVOTE diskgroup on all nodes. OCR and voting disks will be created as part of ASM setup.
    -- Restore data and fra dg from backup using md_restore
    -- Restore latest backup from tapes or disk in to FRA
    -- Restore and Recover database via Rman.
    Thanks
    Steve

    Very good point Sebastian. I guess i need to replace OCR created from new GI cluster install with OCR from backup right, unless you say i can restore OCR from backup while building the new GI cluster. If so, can you point to me to link or documentation in doing so.
    I was also thinking about using clones of GI and RDBMS home instead of making fresh install as part of the restore procedures. Is that an option?
    If i have all the storage partitions intact from previous install and if have to restore the system without any data loss, anything different i should be doing here.
    Steve

  • RMAN, RAC, NFS, and server lock ups

    Good day. My environment is:
    --a 2-node RAC
    --Enterprise Edition 11.2.0.3
    --RHEL 5.1
    The goal is to use RMAN to push backups to a shared NFS mount (on a different server). Both nodes will have access to this location (in the event one node goes down, the other can still run backups). Easy, right?
    Wrong.
    I've tried every NFS mount option in the book. Most work just fine, some don't. When I use the recommended NFS mount options:
    rw,bg,hard,nointr,rsize=32768,wsize=32768,tcp, vers=3,timeo=600, actimeo=0
    or
    rw,bg,hard,nointr,rsize=32768,wsize=32768,proto=tcp,noac,forcedirectio, vers=3,suid
    The mount works normally. I can "ls" and "mkdir" and "touch" and "vi" and "cp" files back and forth from the NFS backup location to the RAC node all day long. No problems. However, when I try to do almost anything in RMAN which requires writing to the NFS backup location such as the command "backup archive all delete input;" (or even things as simple as a Crosscheck or RMAN configuration change which writes any changes back to the autobackup ControlFile) the node locks up. There are no errors (or if there are, I don't know where to find them), even when I use RMAN log.
    Just to recap: I run a Crosscheck (or any RMAN process that writes to the NFS backup location), the node will lock up, and I can let it sit for a day, inaccessible, with CRSCTL on the other node saying it's offline, and the node will never come out of a "frozen" state. It cannot be pinged or connected to.
    I think I can safely rule out NFS mount options at this point.
    I understand (after extensive reading of MOS docs and testing) that RAC RMAN can and does suffer from inefficient I/O when writing to an NFS mount. I don't think that's the culprit either. The autobackup ControlFile is not that big and I cannot see how running a simple Crosscheck would lock an entire node.
    I am hoping someone has encountered this in the past and hopefully it's just a simple misconfiguration somewhere.

    My NFS line in /etc/fstab is (these options are for supporting 11.2.0.3, 11.1.0.7, and 10.2.0.4/5 simultaneously): server.domain:/NFS_Export /backup nfs rw,bg,hard,nointr,rsize=32768,wsize=32768,tcp,actimeo=0,vers=3,timeo=600 0 0
    Before you installed GI, did you by chance do a yum update? I've encountered a similar issue which ended up being due to mkinitrd creating a corrupted kernel; mkinitrd is invoked during the GI installation when the ADVM drivers are added and in my case mkinitrd created a new kernel prior to the new kernel being installed. Second to that, make sure you have the matching kernel headers to your kernel version. If they are different then you could probably get away with just creating a new kernel with mkinitrd and relinking GI/RDBMS homes, but be prepared to wipe GI and reinstall.

  • Multiple databases/instances on 4-node RAC Cluster including Physical Stand

    OS: Windows 2003 Server R2 X64
    DB: 10.2.0.4
    Virtualization: NONE
    Node Configuration: x64 architecture - 4-Socket Quad-Core (16 CPUs)
    Node Memory: 128GB RAM
    We are planning the following on the above-mentioned 4-node RAC cluster:
    Node 1: DB1 with instanceDB11 (Active-Active: Load-balancing & Failover)
    Node 2: DB1 with instanceDB12 (Active-Active: Load-balancing & Failover)
    Node 3: DB1 with instanceDB13 (Active-Passive: Failover only) + DB2 with instanceDB21 (Active-Active: Load-balancing & Failover) + DB3 with instanceDB31 (Active-Active: Load-balancing & Failover) + DB4 with instance41 (Active-Active: Load-balancing & Failover)
    Node 4: DB1 with instanceDB14 (Active-Passive: Failover only) + DB2 with instanceDB22 (Active-Active: Load-balancing & Failover) + DB3 with instanceDB32 (Active-Active: Load-balancing & Failover) + DB4 with instance42 (Active-Active: Load-balancing & Failover)
    Note: DB1 will be the physical primary PROD OLTP database and will be open in READ-WRITE mode 24x7x365.
    Note: DB2 will be a Physical Standby of DB1 and will be open in Read-Only mode for reporting purposes during the day-time, except for 3 hours at night when it will apply the logs.
    Note: DB3 will be a Physical Standby of a remote database DB4 (not part of this cluster) and will be mounted in Managed Recovery mode for automatic failover/switchover purposes.
    Note: DB4 will be the physical primary Data Warehouse DB.
    Note: Going to 11g is NOT an option.
    Note: Data Guard broker will be used across the board.
    Please answer/advise of the following:
    1. Is the above configuration supported and why so? If not, what are the alternatives?
    2. Is the above configuration recommended and why so? If not, what are the recommended alternatives?

    Hi,
    As far as i understand, there's nothing wrong in configuration except you need to consider below points while implementing final design.
    1. No of CPU on each servers
    2. Memory on each servers
    3. If you've RAC physical standby then apply(MRP0) will run on only one instance.
    4. Since you are configuring physical standby for on 3rd and 4th nodes of DB1 4 node cluster where DB13 and DB14 instances are used only for failver, if you've a disaster at data center or power failure in entire data center, you are losing both primary and secondary with an assumption that your primary and physical standby reside in same data center so it may not be highly available architecture. If you are going to use extended RAC for this configuration then it makes sense where Node 1 and Node 2 will reside in Datacenter A and Node 3 ,4 will reside in Datacenter B.
    Thanks,
    Keyur

  • How to create 11.2.0.2 physical standby database from 2 node RAC (11.2.0.2)

    Hi,
    Can any one please help me How to manually create 11.2.0.2 standalone physical standby database from 2 node RAC (11.2.0.2) database which is running in RHEL5 and ASM plugged in.
    DB : 11.2.0.2
    OS : RHEL5
    RMAN duplicate is causing problem with network and we decided to go for manual creation of the same.
    Thanks in Advance..

    Hi;
    Can any one please help me How to manually create 11.2.0.2 standalone physical standby database from 2 node RAC (11.2.0.2) database which is running in RHEL5 and ASM plugged in.
    DB : 11.2.0.2
    OS : RHEL5I had similar issue, what i did
    1. Used source oracle_home on standby server
    2. Created new asm instance and use same naming
    3. I took RMAN full backup on source and move it to target
    4. I edit initora file remove RAC setting and restore db(also edited listener file)
    Regard
    Helios

  • 11gR2- webutil upload file to AS on two nodes RAC?

    Hellow experts plz help with the following issue,
    we are using 11gR2 forms on two node rac,webutil configured on both nodes. upload/download files to AS folder(UP_FILES) using our forms. node 2 is actually replica of node 1(forms,reports,UP_FILES). we are now facing a problem that the form which upload file to AS, only upload that to the node from where it is running say if it is running from node1 then it upload files to folder UP_FILES on that node( as the entry in webutil.cfg file) but we want that UP_FILES folder should be synced with each other on both nodes and a form running whether from node1 or node2 upload the file to both nodes at time.
    How this will be accomplished?
    --------webutil.cfg entry
    transfer.appsrv.read.3=D:\UP_FILES
    transfer.appsrv.write.3=D:\UP_FILES
    -----FORM UPLOAD CODE
          IF :CONTROL.FILE_LOC IS NOT NULL THEN
            acyr3 := :CONTROL.TXTVOUCHERNO||'-'||acyr2 ;
           FILE_RESULT := WEBUTIL_FILE_TRANSFER.CLIENT_TO_AS_WITH_PROGRESS(CLIENTFILE=>:CONTROL.FILE_LOC,
           SERVERFILE =>'D:\UP_FILES\'||acyr3||'.PDF',
          PROGRESSTITLE=>'UPLOAD TO DATABASE IN PROGRESS',
          PROGRESSSUBTITLE=>'PLEASE WAIT' );
          END IF;
    --FORM DOWNLOAD CODE
            FILE_RESULT := WEBUTIL_FILE_TRANSFER.AS_TO_CLIENT_WITH_PROGRESS
                 CLIENTFILE=>'D:\UP_FILES\'||acyr3||'.PDF',
            SERVERFILE =>'D:\UP_FILES\'||acyr3||'.PDF',
          PROGRESSTITLE=>'DOWNLOAD FROM DATABASE IN PROGRESS',
          PROGRESSSUBTITLE=>'PLEASE WAIT'
                 CLIENT_HOST('rundll32.exe url.dll,FileProtocolHandler D:\UP_FILES\'||acyr3||'.PDF');
                 IF FILE_RESULT THEN
      message('File downloaded successfully from the Application Server');
      END IF;

    Well, you're uploading the file to one node, so this is no surprise. What you can do is to store your file on a shared folder (which might be a bit tricky on windows) or synchronize a folder between your nodes. Unfortunately Forms can't access ASM directly (assuming you are using ASM), so you can't store your files directly in ASM where it would be accessible from both nodes.
    cheers

  • How to add a second database along with existing two node RAC environment

    Hi,
    I was wondering if anyone can help me with this.
    My Environment:
    1. Two node RAC Cluster database (11.2.0.2) with ASM running perfectly (Oracle Sid = test-1)
    2. I have installed a second single instance db on node 2 (Oracle Sid = test-2) with NTFS file system for datafiles
    3. Database is up and running, but I am not able to connect it from any client.
    4. I am getting ORA-12514:  TNS:listener does not currently know of service requested in connect descriptor
    5. Database (test2) is registered with grid LISTENER
    5. TNSPING from client machine response is ok
    Am I missing something here, happy to provide more info if requested.
    Thanks,
    PS

    C:\Users\root.test_prod>lsnrctl
    LSNRCTL for 64-bit Windows: Version 11.2.0.2.0 - Production on 23-FEB-2012 11:58:05
    Copyright (c) 1991, 2010, Oracle. All rights reserved.
    Welcome to LSNRCTL, type "help" for information.
    LSNRCTL> status
    Connecting to (DESCRIPTION=(ADDRESS=(PROTOCOL=IPC)(KEY=LISTENER)))
    STATUS of the LISTENER
    Alias LISTENER
    Version TNSLSNR for 64-bit Windows: Version 11.2.0.2.0 - Production
    Start Date 18-FEB-2012 19:51:47
    Uptime 4 days 16 hr. 6 min. 24 sec
    Trace Level off
    Security ON: Local OS Authentication
    SNMP OFF
    Listener Parameter File C:\app\11.2.0\grid\network\admin\listener.ora
    Listener Log File C:\app\11.2.0\grid\log\diag\tnslsnr\IRIS11G-DB-2\listener\alert\log.xml
    Listening Endpoints Summary...
    (DESCRIPTION=(ADDRESS=(PROTOCOL=ipc)(PIPENAME=\\.\pipe\LISTENERipc)))
    (DESCRIPTION=(ADDRESS=(PROTOCOL=tcp)(HOST=10.30.0.202)(PORT=1520)))
    (DESCRIPTION=(ADDRESS=(PROTOCOL=tcp)(HOST=10.30.0.215)(PORT=1520)))
    Services Summary...
    Service "+ASM" has 1 instance(s).
    Instance "+asm2", status READY, has 1 handler(s) for this service...
    Service "test-11gRAC.test_prod.internal" has 1 instance(s).
    Instance "test-11gr2", status READY, has 1 handler(s) for this service...
    Service "test-11gRXDB.test_prod.internal" has 1 instance(s).
    Instance "test-11gr2", status READY, has 1 handler(s) for this service...
    Service "irisapps.test_prod.internal" has 1 instance(s).
    Instance "test-11gr2", status READY, has 1 handler(s) for this service...
    Service "test-2" has 1 instance(s).
    Instance "test-2", status READY, has 1 handler(s) for this service...
    Service "test-2XDB" has 1 instance(s).
    Instance "test-2", status READY, has 1 handler(s) for this service...
    The command completed successfully
    LSNRCTL> service
    Connecting to (DESCRIPTION=(ADDRESS=(PROTOCOL=IPC)(KEY=LISTENER)))
    Services Summary...
    Service "+ASM" has 1 instance(s).
    Instance "+asm2", status READY, has 1 handler(s) for this service...
    Handler(s):
    "DEDICATED" established:1322 refused:0 state:ready
    LOCAL SERVER
    Service "test-11gRAC.test_prod.internal" has 1 instance(s).
    Instance "test-11gr2", status READY, has 1 handler(s) for this service...
    Handler(s):
    "DEDICATED" established:1829 refused:0 state:ready
    LOCAL SERVER
    Service "test-11gRXDB.test_prod.internal" has 1 instance(s).
    Instance "test-11gr2", status READY, has 1 handler(s) for this service...
    Handler(s):
    "D000" established:0 refused:0 current:0 max:1022 state:ready
    DISPATCHER <machine: IRIS11G-DB-2, pid: 4496>
    (ADDRESS=(PROTOCOL=tcp)(HOST=IRIS11G-DB-2.test_prod.internal)(PORT=57695))
    Service "irisapps.test_prod.internal" has 1 instance(s).
    Instance "test-11gr2", status READY, has 1 handler(s) for this service...
    Handler(s):
    "DEDICATED" established:1829 refused:0 state:ready
    LOCAL SERVER
    Service "test-2" has 1 instance(s).
    Instance "test-2", status READY, has 1 handler(s) for this service...
    Handler(s):
    "DEDICATED" established:0 refused:0 state:ready
    LOCAL SERVER
    Service "test-2XDB" has 1 instance(s).
    Instance "test-2", status READY, has 1 handler(s) for this service...
    Handler(s):
    "D000" established:0 refused:0 current:0 max:1022 state:ready
    DISPATCHER <machine: IRIS11G-DB-2, pid: 9340>
    (ADDRESS=(PROTOCOL=tcp)(HOST=IRIS11G-DB-2.test_prod.internal)(PORT=58653))
    The command completed successfully
    LSNRCTL>

Maybe you are looking for

  • Hard drive prevents waking from sleep

    I have a mac mini 2011 2.5Ghz core i5 that has developed an unusual problem. I just got a new Lacie D2 quadra external hard drive and if I leave it connected as the backup I cannot get the computer to wake up. When I move the mouse or tap the keyboar

  • How to use apple remote with macbook air

    How to set-up apple remote to macbook air

  • How do i Update 5800 Software

    i am using Nokia 5800 rm-356 and lives in India but i am not able to figure out how to update it's software to v60. currently i am suing v52 and surely like to upate it Firmware either by Nokia Suite or by FOTA.

  • Odd error message that started after update to Mountain Lion

    Ever since updating to Mountain Lion, I get this error message every now and then: Nothing in particular seems to cause it, and there isn't a server to which I'm trying to connect. My computer is a mid 2010 MacBook Pro Can someone tell me what this m

  • Can sound levels be altered later?

    I have recorded narration in a project, mainly hooking the sound recordings to different objects and text. I now want to change the general sound setting, as the Captivate-decided Pre-amplifier value of 0.4 is really too low. Is there any way to chan