Fatal error while transferring data from one bb to another

I was trying to transfer data from one bb to a different bb using the bb desktop manager tool.  It collected all the data from the first bb fine and then while transferring data to the new one it said a fatal error occured to try again.  That bb will not even turn on now.  No white screen or anything.  Its like there is no operating system.  I don't know what to do.  I tried to connect it and just update the software but that isnt working either.  Says it is not responding during initialization! I think I killed it.... any ideas what to do???

You can try to reload the OS in one of the following ways
http://www.blackberry.com/btsc/KB03485
http://blackberryfaq.net/index.php/How_do_I_wipe_the_BlackBerry_using_Jl_Cmder%3F
If someone has been helpful please consider giving them kudos by clicking the star to the left of their post.
Remember to resolve your thread by clicking Accepted Solution.

Similar Messages

  • Performance issue while transferring data from one itab to another itab

    hi experts,
    i have stored all the general material details in one internal table and description, valuation details of the meterial is stored in another internal table which is of standard table. now i need to tranfer all the data from these two internal tables into one final internal table but it is taking lot of time as it has to transfer lacs of data.
    i have declared the output table as shown below
    DATA:
      t_output TYPE standard TABLE
               OF type_output
               INITIAL SIZE 0
               WITH HEADER LINE.
    (according to the standard i have to declare like this and the two internal tables are declared similar to the above one)
    could somebody suggest me how shall I procced through this...
    thanks in advance....
    ragerds,
    Deepu

    Have a look at the following article which you may find useful:
      <a href="https://www.sdn.sap.comhttp://www.sdn.sap.comhttp://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/40729734-3668-2910-deba-fa0e95e2c541">Improving performance in nested loops</a>
    good luck, damian

  • Getting error , while passing parameters from one page to another page

    Hello friends,
    i am getting error, while passing parameters from one page to another page, below code i wrote.
    public void processFormRequest(OAPageContext pageContext, OAWebBean webBean)
    super.processFormRequest(pageContext, webBean);
    ArrayList arl=new ArrayList();
    EresFrameworkAMImpl am=(EresFrameworkAMImpl)pageContext.getApplicationModule(webBean);
    ERecordImpl ERecordObj=new ERecordImpl();
    HashMap hMap = new HashMap();
    hMap.put("1",ERecordObj.getTransactionName());
    hMap.put("2",ERecordObj.getTransactionKey());
    hMap.put("3",ERecordObj.getDeferredMode());
    hMap.put("4",ERecordObj.getUserKeyLabel());
    hMap.put("5",ERecordObj.getUserKeyValue());
    hMap.put("6",ERecordObj.getTransactionAuditId());
    hMap.put("7",ERecordObj.getRequester());
    hMap.put("8",ERecordObj.getSourceApplication());
    hMap.put("9",ERecordObj.getPostOpAPI());
    hMap.put("10",ERecordObj.getPayload());
    // hMap.put(EresConstants.ERES_PROCESS_ID,
    if(pageContext.getParameter("item1")!=null)
    pageContext.forwardImmediately(EresConstants.EINITIALS_FUNCTION,
    OAWebBeanConstants.KEEP_MENU_CONTEXT,
    null,
    hMap,
    true,
    OAWebBeanConstants.ADD_BREAD_CRUMB_YES
    Error(71,2): method forwardImmediately(java.lang.String, byte, null, java.util.HashMap, boolean, java.lang.String) not found in interface oracle.apps.fnd.framework.webui.OAPageContext
    Thanks
    krishna.

    Hi,
    You have imported the wrong class for HashMap.
    Import
    com.sun.java.util.collections.HashMap; instead of java.util.HashMap
    Thanks,
    Gaurav

  • Error while loading data from one ODS to anther ODS

    Hi all,
    while loading the data from one ODS to another ODS, we are getting error message - Error: No match found in target. No Space on ODS.
    I have done a runstat on these tables of the ODS.. but the issue repeated.
    Can anyone tell how do we solve this issue.

    Check if you get any other Dump message in ST22 to see if it is due to some space availability in the DSO or underlying tables.
    Or Post in detail if any other errore message is available.

  • Fatal error while fetching data from bi

    hi,
    i am getting following error while fetching data from bi using select statement
    i have written code in this way
    SELECT  [Measures].[D2GFTNHIOMI7KWV99SD7GPLTU] ON COLUMNS, NON EMPTY { [DEM_STATE].MEMBERS} ON ROWS FROM DEM_CUBE/TEST_F_8
    error description when i click on test
    Fatal Error
    com.lighthammer.webservice.SoapException: The XML for Analysis provider encountered an error

    thanks for answering .but when i tried writing the statement in transaction 'MDXTEST' and clicked on check i am getting following error
    Error occurred when starting the parser: timeout during allocate / CPIC-CALL: 'ThSAPCMRCV'
    Message no. BRAINOLAPAPI011
    Diagnosis
    Failed to start the MDX parser.
    System Response
    timeout during allocate / CPIC-CALL: 'ThSAPCMRCV'
    Procedure
    Check the Sys Log in Transaction SM21 and test the TCP-IP connection MDX_PARSER in Transaction SM59.
    SO I WENT IN SM 59 TO CHECK THE CONNECTION.
    CAN U TELL ME WHAT CONFIGERATION I NEED TO DO FOR MAKING SELECT STATEMENTS WORK?

  • BADI for transfering data from one modal to another modal within single appset

    Hallo Experts,
    My Business Requirement is Transfer of data from one modal to another in same environment and did this taking reference from below document.(How to custom badi for replicating destination app)
    http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/e04b5d24-085f-2c10-d5a2-c1153a9f9346?QuickLink=index&…
    Document contains TR for implementing BADI, but this document supports bpc version 7.0 and we are using is bpc 10.0.
    So i make all compatible changes in BADI implementation and activate it. Now i am testing using transaction UJKT using following script.
    and the result is records successfully written back to application MIS, but when i check data is not moved to target application MIS.
    I am facing stuck situation in my project work. Please Suggest. Hope for positive reply.
    Script:
    *XDIM_MEMBERSET WM_ACCOUNT = WM_041,
    *XDIM_MEMBERSET WM_UOM_02 = UOM_004
    *XDIM_MEMBERSET WM_UD_2 = WM_07
    *XDIM_MEMBERSET WD_EXT_MAT_GRP =CHALK-PH-I,CHALK-PH-II,CHAVN-PH-I,CHAVN-PH-II,NASHIK-WM,RAJASTHAN-WM,TAMILNADU-WM
    *XDIM_MEMBERSET CATEGORY= Plan
    *XDIM_MEMBERSET AUDITTRAIL=Input
    *XDIM_MEMBERSET P_ENTITY = SIL
    *XDIM_MEMBERSET RPTCURRENCY = LC
    *START_BADI DAPP
       DESTINATION_APP ="MIS"
       RENAME_DIM ="WD_EXT_MAT_GRP= PRODUCT"
       ADD_DIM ="PLANT=NO_PLANT","MIS_ACCOUNTS=CAIN0058040008","COST_CENTER=NO_COST_CENTER","FLOW=Opening","UOM=AMT","CUSTOMER_SALES_2=NO_CUSTOMER","CATEGORY=Plan","AUDITTRAIL=Input"             
       DEBUG = ON
       WRITE = OFF
       QUERY = ON
    *END_BADI
    Please find attached result.
    Regards,
    Dipesh Mudras.

    Hello,
    Here is the manual to copy data between apps (it works with BPC NW75):
    http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/b0480970-894f-2d10-f9a5-d4b1160be203?quicklink=index&overridelayout=true
    It works me, but now I need to modify the Script Logic to make that a "property" from the origin dimension has to be copied to the destination dimension as "id", like follow:
    //*XDIM_MEMBERSET CATEGORY = %CATEGORY_SET%
    //*XDIM_MEMBERSET TIME = %TIME_SET%
    //*XDIM_MEMBERSET ENTITY = %ENTITY_SET%
    //*XDIM_MEMBERSET RPTCURRENCY =  %RPTCURRENCY_SET%
    *START_BADI FiltroPD
         WRITE = OFF
         APPL = $APPLICATION$
         ADD_DIM = "ORIGEN = APPVENTAS"
         ADD_DIM ="O_COSTE = no_input"
         ADD_DIM="CECO = no_input"
         RENAME_DIM="P_ACCT = RATIOS.P_ACCT "
    *END_BADI

  • Data load failed while loading data from one DSO to another DSO..

    Hi,
    On SID generation data load failed while loading data  from Source DSO to Target DSO.
    Following are the error which is occuuring--
    Value "External Ref # 2421-0625511EXP  " (HEX 450078007400650072006E0061006C0020005200650066
    Error when assigning SID: Action VAL_SID_CONVERT, InfoObject 0BBP
    So, i'm  not getting  WHY in one DSO i.e Source  it got successful but in another DSO i.e. Target its got failed??
    While analyzing all i check that SIDs Generation upon Activation is ckecked in source DSO but not in Target DSO..so it is reason its got failed??
    Please explain..
    Thanks,
    Sneha

    Hi,
    I hope your data flow has been designed in such a way where the 1st DSO as a staging Device and all transformation rules and routine are maintained in between 1st to 2nd dso and sid generation upon activation maintained in 2nd DSO.  By doing so you will be getting your data 1st DSO same as your source system data since you are not doing any transformation rules and routine etc.. which helps to avoid data load failure.  
    Please analyze the following
    Have you loaded masterdata before transaction data ... if no please do it first
    go to the property of first dso and check whether there maintained sid generation up on activation (it may not be maintained I guess)
    Goto the property of 2nd Dso and check whether there maintained sid generation up on activation (It may be maintained I hope)
    this may be the reason.
    Also check whether there is any special char involvement in your transaction data (even lower case letter)
    Regards
    BVR

  • Help transferring data from one account to another

    I bought my computer in August 2008 and had previously had an iBook. When transferring my data from my old computer to the new one, I guess I did it wrong so instead of it transferring all to the same place it created two accounts.
    Is there way to transfer my music/pics etc to my main account?
    Also, my iphone is backed up the other account because thats where all my music is, is there a way to also keep all this data once I am using only the main account?

    When you use Migraine Assistant, it transfers the account of the older machine to the newer because the users name is involved in everything, like iTunes playlists etc.,
    Since to run Migraine Assistant you need to have gone through setup on the Mac and created the main Admin account, when you Migrate, you transfer the other account so you wind up with two accounts on the new machine.
    All you have to do is log into the Admin account of the first user you created, then change the acount status of the Migrained account to Admin and reboot, check it out well, then delete the first Admin account at your own leaisure. However it's best to leave two accounts on the machine, just in case your locked out for some reason.
    If you rather transfer your files from one account to another use the Shared Drop Box
    If you have three accounts on the machine (original admin and two migrained migrateda accounts), you'll just have to monkey around and clean up the mess.

  • Fatal Error while retrieving data from Visual Fox Pro using OLE-DB UDS

    Hello all,
    in a customer project we connected a Visual Fox Pro Database via OLE-DB UDS.
    Whenever data is retrieved, a fatal error shows up:
    Fatal Error: Error occurred while processing data stream, null
    Although the data is retrieved correctly!
    MII used: 11.5
    UDS-Server: xMII OLE DB UDS, version 4.0.1.10
    This is the log from the UDS-Server:
    [02/29/2008-11:22:40.195] oleDB [P:4404, T:4792, Framework,   ERROR] 0x0      {extra params}: {5}     [LHDSChannelHandler.cpp @ 488, CLHDSChannelHandler::ExecuteIOCP]
    [02/29/2008-11:22:40.195] oleDB [P:4404, T:4792, Framework,   ERROR] 0x17009     Exception occured: pid = 5, original = 5, data = <ExecutionContext><param name="MODE" value="QUERY"/><param name="QUERY" value="SELECT artikelnr FROM artikel "/><param name="ENDDATE" value="1204280560164"/><param name="ROWCOUNT" value="100"/><param name="STARTDATE" value="1204276960164"/></ExecutionContext>.     [LHDSChannelHandler.cpp @ 493, CLHDSChannelHandler::ExecuteIOCP]
    ********** Start of Exception Stack Trace **********
         1)     [02/29/2008-11:22:40.210] oleDB [P:4404, T:4792,       UDS, ] 0x0     HRESULT: 0x80004005 [Unspecified error]     [BaseOleDbMode.h @ 1133, BaseOleDbMode<class COleDbQuery>::ProcessDynamicResults]
         2)     [02/29/2008-11:22:40.210] oleDB [P:4404, T:4792,       UDS, RETHROW] 0x0     Rethrowing exception.     [BaseOleDbMode.h @ 1211, BaseOleDbMode<class COleDbQuery>::ProcessDynamicResults]
         3)     [02/29/2008-11:22:40.210] oleDB [P:4404, T:4792,       UDS, RETHROW] 0x0     Rethrowing exception.     [Query.cpp @ 107, COleDbQuery::ExecuteOleDbMode]
         4)     [02/29/2008-11:22:40.210] oleDB [P:4404, T:4792,       UDS, RETHROW] 0x0     Rethrowing exception.     [BaseOleDbMode.h @ 381, BaseOleDbMode<class COleDbQuery>::ExecuteMode]
         5)     [02/29/2008-11:22:40.210] oleDB [P:4404, T:4792,       UDS, RETHROW] 0x0     Rethrowing exception over COM.     [LHDSUtil.h @ 1086, lhds::LHDSModeImpl<class COleDbQuery,struct ILHDSMode>::Execute]
         6)     [02/29/2008-11:22:40.210] oleDB [P:4404, T:4792,       UDS, RETHROW] 0x0     Rethrowing exception.     [LHDSUtil.h @ 1124, lhds::LHDSModeImpl<class COleDbQuery,struct ILHDSMode>::Execute]
         7)     [02/29/2008-11:22:40.210] oleDB [P:4404, T:4792, Framework, RETHROW] 0x0     Rethrowing exception over COM.     [LHDSDispatcher.cpp @ 135, CLHDSDispatcher::CallMode]
         8)     [02/29/2008-11:22:40.210] oleDB [P:4404, T:4792,       UDS, RETHROW] 0x0     Rethrowing exception over COM.     [LHDSUtil.h @ 893, lhds::LHDataServerImpl<class CLHOleDbDataServer>::Execute]
         9)     [02/29/2008-11:22:40.210] oleDB [P:4404, T:4792, Framework, ] 0x0     Caught exception over a COM boundry.     [LHDSChannelHandler.cpp @ 566, CLHDSChannelHandler::OnExecute]
         10)     [02/29/2008-11:22:40.210] oleDB [P:4404, T:4792, Framework, RETHROW] 0x0     Rethrowing exception.     [LHDSChannelHandler.cpp @ 574, CLHDSChannelHandler::OnExecute]
         11)     [02/29/2008-11:22:40.210] oleDB [P:4404, T:4792, Framework, RETHROW] 0x0     Rethrowing exception.     [LHDSChannelHandler.cpp @ 502, CLHDSChannelHandler::ExecuteIOCP]
         12)     [02/29/2008-11:22:40.210] oleDB [P:4404, T:4792, Framework, RETHROW] 0x0     Rethrowing exception over COM.     [LHDSChannelHandler.cpp @ 189, CLHDSChannelHandler::Execute]
         13)     [02/29/2008-11:22:40.210] oleDB [P:4404, T:4792,      Host, HANDLED] 0x0     Handled exception.     [ThreadPool.cpp @ 228, ThreadPool::ThreadProc]
    ************** End of Exception Trace **************
    Any help is very much appreciated.
    Regards,
    Matthias
    Edited by: Matthias Wald on Feb 29, 2008 12:12 PM

    Hi Salvatore,
    tried your suggestian but no luck, same error.
    I checked the Version. It shows 4.0.1.10.
    Is there a later version available?
    UDS Settings are (apologize any misspellings):
    Collect Requests                              false
    Log Level                                         Info
    Maximum Concurrent Connections    1000
    Pool Size                                         5
    Port                                                 8087
    Runtime Mode                                  Service
    Service Dependencies                       Eventlog
    Service Passowrd                              *******
    Service Startup                                 Manual
    Service User                                     .\LocalSystem
    Shutdown Timeout                            10
    Stack Size                                       0
    Threading Model                               MTA
    Trusted Requesters                         
    Use IOCP                                        true
    Connection String                            Provider=VFPOLEDB.1;Data Source=D:\Daten;Mode=Share Deny None;Extended Properties="";User ID="";Password="";Mask Password=False;Cache Authentication=False;Encrypt Password=False;Collating Sequence=MACHINE;DSN="";DELETED=True;CODEPAGE=1252;MVCOUNT=16384;ENGINEBEHAVIOR=90;TABLEVALIDATE=3;REFRESH=5;VARCHARMAPPING=False;ANSI=True;REPROCESS=5
    Persistent Connectikon                     true
    Show Affected Rows                         Minimal
    Variant Representation                      String
    I have changed MTA to STA and also tried other admin user for service without luck.
    Any other hints?

  • Help transferring data from one computer to another

    My aunt uses Palm Desktop as her calendar but does not have a palm pda or phone.
    Her computer at home recently died but she was able to get the palm installation folder from the old hard drive.
    She has Palm Desktop installed at work but the information in it is totally different than what she had at home.  She wants to import the data from her home installation into the installation she has at work.
    It used to be really easy to transfer settings from one Palm installation to another.  Side note: My wife has a Palm Treo 650 and I recently formatted her computer and tried to copy her old installation folder from a backup to the new install directory.  Well, it didn't work like it used to.  I don't know if the software has changed or what.  But, she didn't get everything.  It's been a while, I can't remember if she didn't get ANY of it, or just some.  I remember at one point it only showed her contacts for A-M...N-Z were just plain gone.  Really strange.
    But, back on topic, I tried creating a new user in Palm Desktop and then copy the files under her old username folder to the new username folder in the Palm install directory.  But, Palm Desktop doesn't show anything for the new username.
    How should I go about doing this for her?
    Thanks!
    Post relates to: Treo 650 (Alltel)

    The user on her home install was also Trisha.  So, I created a new user on the work computer, "Trisha2".  Then I closed Palm and copied the contents of her backed-up Trisha user folder from home into the new Trisha2 folder on the work install.  Then I started Palm and flipped to the Trisha2 user.  Nothing showed up.  So, I've pretty much done the exact process from the link already.  Which is exactly the way I thought it should work and have had it work in the past.
    I'll try it again without using numbers in the username and see what happens.
    Thanks!
    Post relates to: Treo 650 (Alltel)

  • Transferring data from one table to another table using a Keycolumn using SSIS row by row dynamically

    Hi All,
    I have a Store Procedure(SP) which has a output variable name "AcivityID" which is the key column. In this SP, transformation  of data is done from one table to insert data into other table. I have to execute the SP and insert row by row data
    using the output variable "ActivityID"  whose value will keep on changing. How can I do it?
    Thanks,
    Kallu

    Value changing on a row by row basis? Not quite sure what you mean, but it seems that you want to use the results of an insert into one table as input for another. If so then SSIS is not needed, inside the stored proc use the SQL that will do that and for
    all records as
    INSERT A INTO dbo.table1
    OUTPUT INSERTED.A INTO MyTable;
    Arthur My Blog

  • Error while duplicate database from one server to another server

    Hi there,
    I am practicing duplicate database from one single database server to another server. Hardware and driver path on 2 machines is the same. Here are my steps:
    --------------------- Clone site:
    1. modify listener.ora with content:
    LISTENER =
    (DESCRIPTION_LIST =
    (DESCRIPTION =
    (ADDRESS = (PROTOCOL = IPC)(KEY = EXTPROC1521))
    (ADDRESS = (PROTOCOL = TCP)(HOST = localhost.localdomain)(PORT = 1521))
    ADR_BASE_LISTENER = /u01/app/oracle
    SID_LIST_LISTENER =
    (SID_LIST =
    (SID_DESC =
    (GLOBAL_DBNAME = orcl)
    (ORACLE_HOME = /u01/app/oracle/product/11.2.0/db_1)
    (SID_NAME= orcl)
    ENABLE_GLOBAL_DYNAMIC_ENDPOINT_LISTENER = ON
    2. stop and start listener again, message show like :
    Service “orcl” has 1 instance(s).
    Instance “orcl”, status UNKNOWN, has 1 handler(s) for this service…
    The command completed successfully
    3. create pfile from spfile.
    4. shutdown and startup nomount with pfile is new pfile create from step 3.
    --------------------------------------- Source site
    1. modify tnsnames.ora with content like:
    VVCOREDB =
    (DESCRIPTION =
    (ADDRESS = (PROTOCOL = TCP)(HOST = localhost.localdomain)(PORT = 1521))
    (CONNECT_DATA =
    (SERVER = DEDICATED)
    (SERVICE_NAME = vvcoredb)
    ORCL =
    (DESCRIPTION =
    (ADDRESS = (PROTOCOL = TCP) (HOST = 10.254.1.235) (PORT=1521))
    (CONNECT_DATA =
         (SERVER = DEDICATED)
         (SERVICE_NAME = orcl)
    2. run command from rman like these:
    rman> connect target sys@vvcoredb
    rman> target database Password:
    rman> connect auxiliary sys@orcl
    rman> auxiliary database Password:
    rman> DUPLICATE TARGET DATABASE to orcl from active database spfile nofilenamecheck;
    Then the result show below:
    Starting Duplicate Db at 24-SEP-12
    using target database control file instead of recovery catalog
    allocated channel: ORA_AUX_DISK_1
    channel ORA_AUX_DISK_1: SID=63 device type=DISK
    contents of Memory Script:
    backup as copy reuse
    targetfile '/u01/app/oracle/product/11.2.0/db_1/dbs/spfilevvcoredb.ora' auxiliary format
    '/u01/app/oracle/product/11.2.0/db_1/dbs/spfileorcl.ora' ;
    sql clone "alter system set spfile= ''/u01/app/oracle/product/11.2.0/db_1/dbs/spfileorcl.ora''";
    executing Memory Script
    Starting backup at 24-SEP-12
    allocated channel: ORA_DISK_1
    channel ORA_DISK_1: SID=74 device type=DISK
    Finished backup at 24-SEP-12
    sql statement: alter system set spfile= ''/u01/app/oracle/product/11.2.0/db_1/dbs/spfileorcl.ora''
    contents of Memory Script:
    sql clone "alter system set db_name =
    ''ORCL'' comment=
    ''duplicate'' scope=spfile";
    shutdown clone immediate;
    startup clone nomount;
    executing Memory Script
    sql statement: alter system set db_name = ''ORCL'' comment= ''duplicate'' scope=spfile
    Oracle instance shut down
    connected to auxiliary database (not started)
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03002: failure of Duplicate Db command at 09/24/2012 17:44:12
    RMAN-03015: error occurred in stored script Memory Script
    RMAN-04014: startup failed: ORA-01261: Parameter db_recovery_file_dest destination string cannot be translated
    ORA-01262: Stat failed on a file destination directory
    Linux-x86_64 Error: 2: No such file or directory
    I don't know what wrong with my pfile? I specify correct my pfile on clone site like this:
    orcl.__db_cache_size=1476395008
    orcl.__java_pool_size=16777216
    orcl.__large_pool_size=16777216
    orcl.__oracle_base='/u01/app/oracle'#ORACLE_BASE set from environment
    orcl.__pga_aggregate_target=1342177280
    orcl.__sga_target=1979711488
    orcl.__shared_io_pool_size=0
    orcl.__shared_pool_size=436207616
    orcl.__streams_pool_size=0
    *.audit_file_dest='/u01/app/oracle/admin/orcl/adump'
    *.audit_trail='db'
    *.compatible='11.2.0.0.0'
    *.control_files='/u01/app/oracle/oradata/orcl/control01.ctl','/u01/app/oracle/flash_recovery_area/orcl/control02.ctl'
    *.db_block_size=8192
    *.db_domain=''
    *.db_name='orcl'
    *.db_recovery_file_dest='/u01/app/oracle/flash_recovery_area/vvcoredb'
    *.db_recovery_file_dest_size=4070572032
    *.diagnostic_dest='/u01/app/oracle'
    *.dispatchers='(PROTOCOL=TCP) (SERVICE=orclXDB)'
    *.memory_target=3313500160
    *.open_cursors=300
    *.processes=150
    *.remote_login_passwordfile='EXCLUSIVE'
    *.undo_tablespace='UNDOTBS1'
    Please suggest me do i miss some steps when try to duplicate database?
    Any suggestions are welcome.
    Thanks in advance!
    P/s: Sorry for my bad English.

    Oh god, it worked. Thank you. But i got another problem like this:
    duplicate target database to orcl from active database spfile set "db_recovery_file_dest" = "/u01/app/oracle/flash_recovery_area/vvcoredb" nofilenamecheck;
    Starting Duplicate Db at 24-SEP-12
    using target database control file instead of recovery catalog
    allocated channel: ORA_AUX_DISK_1
    channel ORA_AUX_DISK_1: SID=63 device type=DISK
    contents of Memory Script:
    backup as copy reuse
    targetfile '/u01/app/oracle/product/11.2.0/db_1/dbs/spfilevvcoredb.ora' auxiliary format
    '/u01/app/oracle/product/11.2.0/db_1/dbs/spfileorcl.ora' ;
    sql clone "alter system set spfile= ''/u01/app/oracle/product/11.2.0/db_1/dbs/spfileorcl.ora''";
    executing Memory Script
    Starting backup at 24-SEP-12
    allocated channel: ORA_DISK_1
    channel ORA_DISK_1: SID=74 device type=DISK
    Finished backup at 24-SEP-12
    sql statement: alter system set spfile= ''/u01/app/oracle/product/11.2.0/db_1/dbs/spfileorcl.ora''
    contents of Memory Script:
    sql clone "alter system set db_name =
    ''ORCL'' comment=
    ''duplicate'' scope=spfile";
    sql clone "alter system set db_recovery_file_dest =
    ''/u01/app/oracle/flash_recovery_area/vvcoredb'' comment=
    '''' scope=spfile";
    shutdown clone immediate;
    startup clone nomount;
    executing Memory Script
    sql statement: alter system set db_name = ''ORCL'' comment= ''duplicate'' scope=spfile
    sql statement: alter system set db_recovery_file_dest = ''/u01/app/oracle/flash_recovery_area/vvcoredb'' comment= '''' scope=spfile
    Oracle instance shut down
    connected to auxiliary database (not started)
    Oracle instance started
    Total System Global Area 3307048960 bytes
    Fixed Size 2217872 bytes
    Variable Size 1811941488 bytes
    Database Buffers 1476395008 bytes
    Redo Buffers 16494592 bytes
    contents of Memory Script:
    sql clone "alter system set db_name =
    ''VVCOREDB'' comment=
    ''Modified by RMAN duplicate'' scope=spfile";
    sql clone "alter system set db_unique_name =
    ''ORCL'' comment=
    ''Modified by RMAN duplicate'' scope=spfile";
    shutdown clone immediate;
    startup clone force nomount
    backup as copy current controlfile auxiliary format '/u01/app/oracle/oradata/vvcoredb/control01.ctl';
    restore clone controlfile to '/u01/app/oracle/recovery_area/vvcoredb/control02.ctl' from
    '/u01/app/oracle/oradata/vvcoredb/control01.ctl';
    alter clone database mount;
    executing Memory Script
    sql statement: alter system set db_name = ''VVCOREDB'' comment= ''Modified by RMAN duplicate'' scope=spfile
    sql statement: alter system set db_unique_name = ''ORCL'' comment= ''Modified by RMAN duplicate'' scope=spfile
    Oracle instance shut down
    Oracle instance started
    Total System Global Area 3307048960 bytes
    Fixed Size 2217872 bytes
    Variable Size 1811941488 bytes
    Database Buffers 1476395008 bytes
    Redo Buffers 16494592 bytes
    Starting backup at 24-SEP-12
    using channel ORA_DISK_1
    channel ORA_DISK_1: starting datafile copy
    copying current control file
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03002: failure of Duplicate Db command at 09/24/2012 18:42:09
    RMAN-03015: error occurred in stored script Memory Script
    RMAN-03009: failure of backup command on ORA_DISK_1 channel at 09/24/2012 18:42:09
    ORA-17628: Oracle error 19505 returned by remote Oracle server
    I have a comment that source site doesn't have flash_recovery_area, instead it's name is recovery_area, is it problem?
    Edited by: 788420 on Sep 24, 2012 4:44 AM

  • Dtp error when loading data from one cube to another cube?

    hi,experts
      it seems strange that when I tick the some characteristics as navigational attributes in the cube , the error occurs during the execution of the DTP:error while updating to target Z* (Cube name ).
    once i turn the flag off , no error appear. could anyone give me a clue ?
    thanks in advance!

    Hi,
    When u make changes in the cube u need to make necessary changes in Transformation and activate the DTP.The checking of Navigational attributes in cube will appear in u r transformation as new fields where u need to create a mapping for them and then activate, also activate your DTP and then load the data. This should resolve your issue.
    Regards

  • Best method for transferring data from one database to another?

    There is an 8i database I have to deal with on a regular basis. Besides being completely outdated and unsupported, in my opinion the database has not been well-maintained--for a table with 25 million records, should ‘select count(*) from table;’ take 5 minutes to run? I don’t really know, but that seems long to me. Many complex queries (most including only tables with less than a million records) take a ridiculously long time to run, to the point that I can’t even run some of them.
    I am not the DBA; I don't have the authority to fiddle with the database (nor would I feel comfortable doing that), and the powers that be will not put effort into improving functionality of this database due to an alleged plan to update/replace it within the next year. However, in the mean time, I still have to get data out of this database on a regular basis.
    I have XE 10g installed on my local machine, and I have set up a database link in it to the 8i database. I have found that I can pull in basic data (simple queries) from the 8i database into tables in my XE database (e.g. create table tbl1 as select data from tbl1@8idb) and then query those tables to get the information I need much, much faster (including creation of the tables). While this option does not allow me to create queries/reports that other people can run, it makes work I’m doing only for myself much faster.
    What I’m wondering is, what is the best way to bring the information I need over to my database? I usually don’t need entire tables, and I can probably identify a number of key tables (or parts of tables) I need. What I’ve been doing up until now is writing CREATE TABLE statements on the fly, but then I end up forgetting what all I’ve done, and each time I want up-to-date data, I have to drop the tables and re-create them. It seems to me that there should be an easier way to do this than to copy and paste from a text document into SQL*Plus.
    Does anyone have any suggestions for me on how best to do this?

    Sorry, I guess I posted this in the wrong forum. I re-posted in the database-general forum.

  • Transferring data from one DB to another DB

    Hi Experts,
    In production database 30 tables are there in different schemas.
    In some tables data is existed for 5 years
    in some tables data is existed 8 years
    in some tables data is existed for 6 years.
    And all these tables are having billions and millions of records.
    Let us assume the tables in the production database are
    Table1 -- 8 years data -- No. Of Records 3538969000
    Table2 -- 6 years data -- No. Of Records  592844435
    Table3 -- 3 years data -- No. Of Records   33224993
    Table4 -- 4 years data -- No. Of Records   52361756
    Table5 -- 5 years data -- No. Of Records    8948567
    Table15 -- 6 years data -- No. Of Records 308476987Now I want to trnasfer these 15 tables data to test database.
    Based on the following conditions.
    For Table1 I want to transfer 6 years data to test database, keeep 2 years data in production
    and delete all 6 years data from production databse.
    For Table2 I want to transfer 4 years data to test database, keeep 2 years data in production
    and delete all 4 years data from production databse.
    For Table3 I want to transfer 2 years data to test database keeep 1 years data in production
    and delete all 2 years data from production databse.This will be done periodically.
    I.e. suppose if we run the script now and transfer the requested data to test database,
    again after one year the data got increased in production then again we have to run the
    script and transfer data to test database based on some conditions,so it should work for
    long time.
    Please help me what is the good and fast process to transfer the data.
    Thanks in advance.

    Thanks for your information.
    Hi ,
    To make the process dynamic why we want to make it dynamic in future any tables needs to e archived
    only one script should execute.
    I have created the below control table to pass ID, TABLE_NAME, ARCHIVE_TABLE_NAME, WHERE_CONDITION dynamically.
    ID  TABLE_NAME ARCHIVE_TABLE_NAME  WHERE_CONDITION
    1   Table1     archive_table1       created_date< MONTHS_BETWEEN(SYSDATE,-24)
    1   Table2     archive_table2    Accepted_date< MONTHS_BETWEEN(SYSDATE,-24)
    1   Table3     archive_table3    last_proposal_date< MONTHS_BETWEEN(SYSDATE,-24)
    1   Table4     archive_table4    lat_modified_date< MONTHS_BETWEEN(SYSDATE,-24)
    2   Table5     archive_table5       changed_date< MONTHS_BETWEEN(SYSDATE,-24)
    2   Table6     archive_table6    received_date< MONTHS_BETWEEN(SYSDATE,-24)
    3   Table7     archive_table7       send_date< MONTHS_BETWEEN(SYSDATE,-24)
    3   Table8     archive_table8       exact_date< MONTHS_BETWEEN(SYSDATE,-24)I want to pass ID from the JOB then the procedure should execute for all the tables with respective ID.
    this is my job
    DECLARE
      X NUMBER;
    BEGIN
      SYS.DBMS_JOB.SUBMIT
      ( job       => X
       ,what      => 'APPS.myprocedure(1);'
       ,next_date => to_date('24/12/2012 00:00:00','dd/mm/yyyy hh24:mi:ss')
       ,interval  => 'TRUNC(SYSDATE+1)'
       ,no_parse  => FALSE
      SYS.DBMS_OUTPUT.PUT_LINE('Job Number is: ' || to_char(x));
    COMMIT;
    END;
    For example if I pass the ID as '1' through job,then the procedure should execute for
    1   Table1     archive_table1       created_date< MONTHS_BETWEEN(SYSDATE,-24)
    1   Table2     archive_table2    Accepted_date< MONTHS_BETWEEN(SYSDATE,-24)
    1   Table3     archive_table3    last_proposal_date< MONTHS_BETWEEN(SYSDATE,-24)
    INSERT INTO archive_table1
    (SELECT * FROM Table1 WHERE created_date< MONTHS_BETWEEN(SYSDATE,-24));
    DELETE FROM Table1 WHERE created_date< MONTHS_BETWEEN(SYSDATE,-24));
    INSERT INTO archive_table2
    (SELECT * FROM Table2 WHERE created_date< MONTHS_BETWEEN(SYSDATE,-24));
    DELETE FROM Table2 WHERE created_date< MONTHS_BETWEEN(SYSDATE,-24));
    INSERT INTO archive_table3
    (SELECT * FROM Table3 WHERE created_date< MONTHS_BETWEEN(SYSDATE,-24));
    DELETE FROM Table3 WHERE created_date< MONTHS_BETWEEN(SYSDATE,-24));If I pass ID as 2 through job,then the procedure should execute for respective table names and where conditions.
    I have written the following code but it's not giving expected result.
    CREATE OR REPLACE PROCEDURE myprocedure (P_ID IN NUMBER)
    IS
       CURSOR C IS
          SELECT id, table_name,archive_table_name,where_condition FROM apps_global.control_ram WHERE id=p_id
          ORDER BY id,table_name;
    BEGIN
       FOR I IN C
       LOOP
    EXECUTE IMMEDIATE 'INSERT INTO '|| I.ARCHIVE_TABLE_NAME || '
    (SELECT * FROM '|| I.TABLE_NAME ||' WHERE '||I.WHERE_CONDITION||)';
    EXECUTE IMMEDIATE 'DELETE FROM ' ||I.TABLE_NAME|| ' WHERE '||I.WHERE_CONDITION||' ;
       END LOOP;
    END;
    END myprocedure;Please help me.
    Thanks in advance.

Maybe you are looking for