Database Refresh Status log (Info)

Dear All,
Is there thing that can help to find out completion status for database refresh.
Application log like hyplive.log located in the following location
Oracle\Middleware\user_projects\epmsystem1\diagnostics\logs\essbase\essbase_0\app\HypLive
the log file says this
"[Wed Apr 04 14:53:50 2012]Local/HypLive/Plan1/admin@Native Directory/Info(1013091)
Received Command [Restructure] from user [admin@Native Directory]
[Wed Apr 04 14:53:50 2012]Local/HypLive/Plan1/admin@Native Directory/Info(1019017)
Reading Parameters For Database [Drxxxxxx]
[Wed Apr 04 14:53:50 2012]Local/HypLive/Plan1/admin@Native Directory/Info(1019012)
Reading Outline For Database [Drxxxxxx]
[Wed Apr 04 14:53:50 2012]Local/HypLive/Plan1/admin@Native Directory/Info(1019024)
Reading Outline Transaction For Database [Drxxxxxx]
[Wed Apr 04 14:53:50 2012]Local/HypLive/Plan1/admin@Native Directory/Info(1007043)
Declared Dimension Sizes = [1408 4 82 2 2 3 10 14 16 17 33 66 183 193 265 852 922 920 2 1 4 3 3 3 6 ]
[Wed Apr 04 14:53:50 2012]Local/HypLive/Plan1/admin@Native Directory/Info(1007042)
Actual Dimension Sizes = [1395 4 81 1 1 3 10 12 16 8 26 64 145 163 245 848 898 914 0 0 0 0 0 0 5 ]
[Wed Apr 04 14:53:50 2012]Local/HypLive/Plan1/admin@Native Directory/Info(1007125)
The number of Dynamic Calc Non-Store Members = [191 0 54 0 0 0 0 2 0 9 6 2 34 20 19 4 0 5 ]
[Wed Apr 04 14:53:50 2012]Local/HypLive/Plan1/admin@Native Directory/Info(1007126)
The number of Dynamic Calc Store Members = [0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 ]
[Wed Apr 04 14:53:50 2012]Local/HypLive/Plan1/admin@Native Directory/Info(1007127)
The logical block size is [112995]
[Wed Apr 04 14:53:50 2012]Local/HypLive/Plan1/admin@Native Directory/Info(1010008)
Maximum Declared Blocks is [26928283342271066000000000] with data block size of [115456]
[Wed Apr 04 14:53:50 2012]Local/HypLive/Plan1/admin@Native Directory/Info(1010007)
Maximum Actual Possible Blocks is [5868290301232498000000000] with data block size of [32508]
[Wed Apr 04 14:53:50 2012]Local/HypLive/Plan1/admin@Native Directory/Info(1200481)
Formula for member [2050103] will be executed in [CELL] mode
[Wed Apr 04 14:53:51 2012]Local/HypLive/Plan1/admin@Native Directory/Info(1007067)
Total Restructure Elapsed Time : [0.235] seconds
[Wed Apr 04 14:53:52 2012]Local/HypLive/Plan2/admin@Native Directory/Info(1007067)
Total Restructure Elapsed Time : [0.266] seconds
but i am looking for the status of database refresh failed or succeeded?
Any suggestions.
Regards
Ace

If it failed then you would of got an error generated in the essbase application log and the planning log.
Cheers
John
http://john-goodwin.blogspot.com/

Similar Messages

  • Database Refresh Failed

    Hi All,
    I am working on a database refresh and it failed with the following errors
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03002: failure of Duplicate Db command at 04/15/2011 16:27:41
    RMAN-03015: error occurred in stored script Memory Script
    ORA-00283: recovery session canceled due to errors
    RMAN-11003: failure during parse/execution of SQL statement: alter database recover logfile '/srv/database/arch2/1_65508_689972797.dbf'
    ORA-00283: recovery session canceled due to errors
    ORA-00368: checksum error in redo log block
    ORA-00353: log corruption near block 256 change 43879018791 time 04/12/2011 04:30:54
    ORA-00334: archived log: '/srv/database/arch2/1_65508_689972797.dbf'
    Do you have any idea of how to solve this error.
    Please help.

    Hi;
    What is your OS? If its solaris than please see:
    Duplicate Fails with Corruption Errors on Solaris QFS Filesystem [ID 764113.1]
    Regard
    Helios

  • Physical standby database standby redo log problem

    Hello
    We have a physical standby database , I've created some standby redo log files but my problem is that they aren't used,
    their status in v$stanby_log view is UNASSIGNED
    and I see this message (ORA-16086: standby database does not contain available standby log files) in primary database alert_log file
    while when I run "alter system switch logfile" in the primary database it transfer redo logs to the physsical standby database
    and archive log file will be created in standby database
    I've even recreated the standby redo log files and I added new ones to them but the problem wasn't solved
    Do you know what is problem ?
    elect group#,THREAD#,BYTES,STATUS from V$STANDBY_LOG;
    group#     THREAD#      BYTES       STATUS
    1                   0                   524288000                   UNASSIGNED                  
    2                   0                   524288000                   UNASSIGNED                  
    3                   0                   524288000                   UNASSIGNED                  
    8                   0                   524288000                   UNASSIGNED                  
    9                   0                   524288000                   UNASSIGNED                  
    10                   0                   524288000                   UNASSIGNED                  
    select group#,THREAD#,BYTES,MEMBERS,STATUS from v$log;
    group#                    THREAD#                    BYTES                    MEMBERS                    STATUS
    4                   1                   524288000                   2                   CLEARING                  
    7                   1                   524288000                   2                   CLEARING_CURRENT                  
    6                   1                   524288000                   2                   CLEARING                  
    5                   1                   524288000                   2                   CLEARING                  
    thanks

    Hello Anurag
    Thank you for your reply
    I have found some issue in the standby database alert_log too , in the standby database alert_log it has been written:
    RFS[782]: Assigned to RFS process 3919
    RFS[782]: Identified database type as 'physical standby'
    Primary database is in MAXIMUM AVAILABILITY mode
    Standby controlfile consistent with primary
    Primary database is in MAXIMUM AVAILABILITY mode
    Standby controlfile consistent with primary
    RFS[782]: No standby redo logfiles selected (reason:6)
    Sun Jan 31 13:59:43 2010
    Errors in file /u01/app/oracle/admin/tehrep/udump/tehrep_rfs_3919.trc:
    ORA-16086: standby database does not contain available standby log files
    Sun Jan 31 13:59:48 2010
    RFS[781]: Archived Log: '/disks/sda/tehrep/archivelogs/1_6516_670414641.dbf'
    Sun Jan 31 13:59:50 2010
    and the context "/u01/app/oracle/admin/tehrep/udump/tehrep_rfs_3919.trc"  is below :
    +/u01/app/oracle/admin/tehrep/udump/tehrep_rfs_3919.trc+
    Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    ORACLE_HOME = /u01/app/oracle/product/10.2.0/db_1
    System name:    Linux
    Node name:      linserver2.com
    Release:        2.6.9-42.ELsmp
    Version:        #1 SMP Wed Jul 12 23:27:17 EDT 2006
    Machine:        i686
    Instance name: tehrep
    Redo thread mounted by this instance: 1
    Oracle process number: 58
    Unix process pid: 3919, image: [email protected]
    *** SERVICE NAME:() 2010-01-31 13:59:43.865
    *** SESSION ID:(109.1225) 2010-01-31 13:59:43.865
    KCRRFLAS
    KCRRSNPS
    No space in recovery area for active standby redo logs
    The primary database is operating in MAXIMUM PROTECTION
    or MAXIMUM AVAILABILITY mode, and the standby database
    does not contain adequate disk space in the recovery area
    to safely archive the contents of the standby redo logfiles.
    ORA-16086: standby database does not contain available standby log files
    when I saw this line "No space in recovery area for active standby redo logs" I thought that STANDBY_ARCHIVE_DEST parameter points where that there is no enough space , but when I consider I found out that points a directory on disk a "sda" that has enough space , I don't know what that means
    by the way, at below I've written a section of the primary database alert_log context and "lgwr" trace file around Sun Jan 31 13:30:34 2010
    alert_log :
    ORA-16086: standby database does not contain available standby log files
    Sun Jan 31 13:30:34 2010
    LGWR: Failed to archive log 7 thread 1 sequence 6512 (16086)
    Thread 1 advanced to log sequence 6512
    Current log# 7 seq# 6512 mem# 0: /disks/sdb/tehrep/redo71.log
    Current log# 7 seq# 6512 mem# 1: /disks/sdd/tehrep/redo72.log
    LNSc started with pid=53, OS id=11451
    Sun Jan 31 13:36:34 2010
    Errors in file /u01/app/oracle/admin/tehrep/bdump/tehrep_lgwr_3692.trc:
    ORA-16086: standby database does not contain available standby log files
    Sun Jan 31 13:36:34 2010
    LGWR: Failed to archive log 5 thread 1 sequence 6513 (16086)
    Thread 1 advanced to log sequence 6513
    Current log# 5 seq# 6513 mem# 0: /disks/sdb/tehrep/redo51.log
    Current log# 5 seq# 6513 mem# 1: /disks/sdd/tehrep/redo52.log
    */u01/app/oracle/admin/tehrep/bdump/tehrep_lgwr_3692.trc file :*
    Error 16086 creating standby archive log file at host '(DESCRIPTION=(ADDRESS_LIST=(ADDRESS=(PROTOCOL=tcp)(HOST=linserver2.com
    +)(PORT=1521)))(CONNECT_DATA=(SERVICE_NAME=tehrep_XPT.com)(INSTANCE_NAME=tehrep)(SERVER=dedicated)))'+
    *** 2010-01-31 13:30:34.712 60679 kcrr.c
    LGWR: Attempting destination LOG_ARCHIVE_DEST_3 network reconnect (16086)
    *** 2010-01-31 13:30:34.712 60679 kcrr.c
    LGWR: Destination LOG_ARCHIVE_DEST_3 network reconnect abandoned
    ORA-16086: standby database does not contain available standby log files
    *** 2010-01-31 13:30:34.712 60679 kcrr.c
    LGWR: Error 16086 creating archivelog file '(DESCRIPTION=(ADDRESS_LIST=(ADDRESS=(PROTOCOL=tcp)(HOST=linserver2.com)(PORT=1521
    +)))(CONNECT_DATA=(SERVICE_NAME=tehrep_XPT.com)(INSTANCE_NAME=tehrep)(SERVER=dedicated)))'+
    *** 2010-01-31 13:30:34.712 58941 kcrr.c
    kcrrfail: dest:3 err:16086 force:0 blast:1
    Receiving message from LNSc
    *** 2010-01-31 13:30:34.718 55444 kcrr.c
    Making upidhs request to LNSc (ocis 0x0xb648db48). Begin time is <01/31/2010 13:30:30> and NET_TIMEOUT <180> seconds
    NetServer pid:11196
    *** 2010-01-31 13:30:38.718 55616 kcrr.c
    upidhs done status 0
    *** 2010-01-31 13:36:31.062
    LGWR: Archivelog for thread 1 sequence 6513 will NOT be compressed
    *** 2010-01-31 13:36:31.062 53681 kcrr.c
    +Initializing NetServer[LNSc] for dest=(DESCRIPTION=(ADDRESS_LIST=(ADDRESS=(PROTOCOL=tcp)(HOST=linserver2.com)(PORT=1521)))(CO+
    NNECT_DATA=(SERVICE_NAME=tehrep_XPT.com)(INSTANCE_NAME=tehrep)(SERVER=dedicated))) mode SYNC
    LNSc is not running anymore.
    New SYNC LNSc needs to be started
    Waiting for subscriber count on LGWR-LNSc channel to go to zero
    Subscriber count went to zero - time now is <01/31/2010 13:36:31>
    Starting LNSc ...
    Waiting for LNSc to initialize itself
    *** 2010-01-31 13:36:34.116 53972 kcrr.c
    +Netserver LNSc [pid 11451] for mode SYNC has been initialized+
    Performing a channel reset to ignore previous responses
    +Successfully started LNSc [pid 11451] for dest (DESCRIPTION=(ADDRESS_LIST=(ADDRESS=(PROTOCOL=tcp)(HOST=linserver2.com)(PORT=1+
    +521)))(CONNECT_DATA=(SERVICE_NAME=tehrep_XPT.com)(INSTANCE_NAME=tehrep)(SERVER=dedicated))) mode SYNC ocis=0x0xb648db48+
    *** 2010-01-31 13:36:34.116 54475 kcrr.c
    +Making upiahm request to LNSc [pid 11451]: Begin Time is <01/31/2010 13:36:31>. NET_TIMEOUT = <180> seconds+
    Waiting for LNSc to respond to upiahm
    *** 2010-01-31 13:36:34.266 54639 kcrr.c
    upiahm connect done status is 0
    Receiving message from LNSc
    Receiving message from LNSc
    Destination LOG_ARCHIVE_DEST_3 is in STANDBY RESYNCHRONIZATION mode
    Receiving message from LNSc

  • Database refresh from standby database

    All,
    OS : Sun version : 11.1.0.7
    I'm doing database refresh using stand by database datafiles in different server.
    1)Disabled Data guard broker in standby database .
    2)Copied all datafiles in mount stage to target server (different server)
    3)created controlfile and started the database in mount stage in different server
    4)Applied all archive log files .
    5)But still the database not recovered .
    SQL> recover database using backup controlfile until cancel;
    ORA-00279: change 10401147296308 generated at 01/13/2012 09:03:43 needed for
    thread 1
    ORA-00289: suggestion :
    /u01/app/oracle/admin/qecgdeva/arch/qecgdeva_727091849_1_57879.arc
    ORA-00280: change 10401147296308 for thread 1 is in sequence #57879
    Specify log: {<RET>=suggested | filename | AUTO | CANCEL}
    /u01/app/oracle/admin/qecgdeva/arch/qecgdeva_727091849_1_57879.arc
    Specify log: {<RET>=suggested | filename | AUTO | CANCEL}
    /u01/app/oracle/admin/qecgdeva/arch/qecgdeva_727091849_1_57885.arc
    ORA-00279: change 10401156453328 generated at 01/13/2012 10:48:01 needed for
    thread 1
    ORA-00289: suggestion :
    /u01/app/oracle/admin/qecgdeva/arch/qecgdeva_727091849_1_57886.arc
    ORA-00280: change 10401156453328 for thread 1 is in sequence #57886
    ORA-00278: log file
    '/u01/app/oracle/admin/qecgdeva/arch/qecgdeva_727091849_1_57885.arc' no longer
    needed for this recovery
    Specify log: {<RET>=suggested | filename | AUTO | CANCEL}
    CANCEL
    ORA-01547: warning: RECOVER succeeded but OPEN RESETLOGS would get error below
    ORA-01196: file 1 is inconsistent due to a failed media recovery session
    ORA-01110: data file 1:
    '/mounts/qecgdeva_data/oradata/qecgdeva/dbfiles/qecgprod_system.dbf'
    ORA-01112: media recovery not started
    SQL> alter database open resetlogs;
    alter database open resetlogs
    ERROR at line 1:
    ORA-01196: file 1 is inconsistent due to a failed media recovery session
    ORA-01110: data file 1:
    '/mounts/qecgdeva_data/oradata/qecgdeva/dbfiles/qecgprod_system.dbf'Am I missing any steps here ? Please advice me on this .
    Thanks.

    Hi,
    When you copy files of standby server to another target server ? - Yesterday.
    Is standby is SYNC with primary? or MRP is still running? - ---------yes, standby db is running fine .
    Even you can open database with resetlogs after recreating controlfile, no need of recovery too -- >I was unable to open . So went to apply arch files.
    Do you have archive of thread 1 is in sequence #57887, -- No , Because sequence 57887 yet generate. I have applied all recent available archive log files.
    ORA-00289: suggestion :
    /u01/app/oracle/admin/qecgdeva/arch/qecgdeva_727091849_1_57886.arc
    ORA-00280: change 10401156453328 for thread 1 is in sequence #57886
    ORA-00278: log file
    '/u01/app/oracle/admin/qecgdeva/arch/qecgdeva_727091849_1_57885.arc' no longer
    needed for this recovery
    Specify log: {<RET>=suggested | filename | AUTO | CANCEL}
    CANCEL
    ORA-01547: warning: RECOVER succeeded but OPEN RESETLOGS would get error below
    ORA-01196: file 1 is inconsistent due to a failed media recovery session
    ORA-01110: data file 1:
    '/mounts/qecgdeva_data/oradata/qecgdeva/dbfiles/qecgprod_system.dbf'
    ORA-01112: media recovery not started
    SQL> alter database open resetlogs;
    alter database open resetlogs
    ERROR at line 1:
    ORA-01196: file 1 is inconsistent due to a failed media recovery session
    ORA-01110: data file 1:
    '/mounts/qecgdeva_data/oradata/qecgdeva/dbfiles/qecgprod_system.dbf'Edited by: 805877 on Jan 12, 2012 10:43 PM

  • Database refresh from cold backup and hotbackup.

    How can we perform Database refresh from cold backup and hotbackup?

    OracleM wrote:
    How can we perform Database refresh from cold backup and hotbackup?I understand that you have Cold/hot backup and you need recover(refresh) this backup?If yes then restore cold/hot backup and
    sqlplus "/as sysdba"
    startup mount;
    recover database using backup controlfile unil cancel;
    /*then apply all available archive logs*/
    alter database open resetlogs;

  • Database Refresh error

    Hi John,
    I have multiple queries now
    1. I pushed the account hierarchy using ODI. Capex and Wrkforce have been initialised on my application before pushing the hierarchy. In the log it show that member 'buildings' already exist. When i checked the hierarchy there are other members with 'buildings' as part of them like AFE-Buildings.
    2. Also, when the database refresh is run the capex default account members are not visible in the hiearchy. They are seen in the planning applcn but not in the EAS console
    3. Can we de-initialise workforce and capex?
    Edited by: Sravan Ganti on May 18, 2009 8:23 PM

    Hi,
    I am not sure what you mean in question 1, if there is a member in any of the dimension called "buildings" then it will fail as it is a duplicate.
    If you see the members in planning and refresh from planning with no errors then it is more than likely they in exist in one of the cubes, make sure you have the plan types correctly set up in planning and are checking the cube that relates to the plan type.
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Test ASM Database refresh

    Hello,
    I am trying to refresh the test db from prod and both are of type ASM. Can someone provide me the detailed steps?
    The test and prod nodes are part of a cluster.
    Thanks.

    803045 wrote:
    I am trying to refresh the test db from prod and both are of type ASM. Can someone provide me the detailed steps?
    The test and prod nodes are part of a cluster.You need to be more precise than this.
    Are you referring to 2 ASM instanaces? On which case, exactly what do you want to refresh?
    Or are you referring to to 2 database instances that both are using ASM? Also, exactly what is to be refreshed?
    Cluster - are you referring to RAC? How are these test and prod nodes configured? Are these separate RAC db instances in the same RAC? Are these stand-alone db instances running on RAC nodes? Something else? And for the ASM part - do these ASM instances use the same shared storage server, or are local disks used instead?
    Are the databases in archive log mode? Are rman used for backup and recovery?
    And as already mentioned, the Oracle and o/s versions are very applicable too. Oracle version must include all 4 digits (Oracle 11gr2 is for example a product identifier and not a version number).

  • Oracle eBS clone : only database refresh

    Hi,
    Someone last week around asked me regardings steps I will follow wrt database refresh in eBS.
    My answer was ( considering a cold backup)
    1) Restore the datafiles,redo logs to the target env adn run adcfgclone dbtier contextfile=< context file location for RDBMS>
    This will work if the mount point naming convention is exactly the same as on target.
    2) Run autoconfig.
    My queries here are:
    1) Do I need to do something regarding Apps side, ( running autoconfig , or adcfgclone etc )?
    2)Was there any need for running autoconfig specially if I have given the contextfile location in step 1?
    Thx,
    Romit

    2)Was there any need for running autoconfig specially if I have given the contextfile location in step 1?Yes, you need to run AutoConfig to populate FND_NODES with the application nodes information. You may also need to clear the FND_NODES table as described in the following note:
    Note: 260887.1 - Steps to Clean Nonexistent Nodes or IP Addresses from FND_NODES
    https://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=260887.1

  • Modify database schema and logging expression

    I want to add more information to my TestStand database. For example I want to add the UUT Part Number to the database. Presently I collect the UUT part number in a station globals variable. I assume that I need to move this part number to the database local expression "logging.UUT.PartNumber" so that I can log this info, but I can't figure out how to add this variable. How do I create this variable?
    Thanks in advance for any suggestions.

    In the process model (i.e. SequentialModel.seq) you can modify the custom data type "UUT". Go to the "Sequence File Types" window and select "Custom Data Types" and you will see the UUT container. Add a variable called "PartNumber". Then you will need to add a statement in your process model to assign the value of the part number. Go to the "Single Pass" sequence (and/or Test UUTs) and add a statement similar to "Locals.UUT.PartNumber = StationGlobal.PartNumber". The statement needs to be after the MainSequence call, assuming that is where the part number gets assigned.
    The database logging utility automatically adds the UUT container to the Logging container, so you will now be able to access it as Logging.UUT.PartNumber when you configure your databas
    e statement.

  • How can I determine what sites are being referenced within Central Admin Upgrade and Migration Manage Databases Upgrade Status?

    When I go to Central Admin > Upgrade and Migration  > Manage Databases Upgrade Status, I have 2 content databases which have the status:
    Database is up to date, but some sites are not completely upgraded.
    How can I determine which sites are not completely upgraded?

    Manage Databases Upgrade Status will provide you all active and offline DB details, you can get same result
    using below PowerShell cmdlet.
    Get-SPDatabase and Get-SPContentDatabase will provide all active database/Content DB in Farm which include Service application db, central admin DB.
    Get-SPDatabase | Format-Table Name, ID
    Coming back to your question, if you find that there are some site are not completely upgraded then run below command and understand the cause if issue on specific DB.
    Test-SPContentDatabase WSS_ContentDB_Name
    If you find any missing file issue in DB then resolve these issue to upgrade content database.
    (verify all customizations referenced within the content database are also installed in the web application. This cmdlet can be issued
    against a content database currently attached to the farm, or a content database that is not connected to the farm )
    Use the Upgrade-SPContentDatabase cmdlet
    to resume a failed database upgrade or begin a build-to-build database upgrade against a SharePoint content database
    Upgrade-SPContentDatabase WSS_Content
    reference:
    http://technet.microsoft.com/en-us/library/ff607813(v=office.15).aspx
    http://technet.microsoft.com/en-us/library/ff607941(v=office.15).aspx
    If my contribution helps you, please click Mark As Answer on that post and
    Vote as Helpful
    Thanks, ShankarSingh(MCP)

  • How to recover a database with archive log

    how to recover database with archivle log

    Hi,
    With in information no one can tell the answer.
    Kindly post your qusetion in details information, you want to recover the database in archive log mode, what type of error you get, bcoz depending up on the errors you recover your database,
    please mention all about your database
    cheers
    Senthil Kumar

  • Problem with Whole Database Online+Redo log Backup

    Dear Marcus Sir,
    I am facing while taking "Whole Database Online+Redo log Backup" through DB13 T-Code.
    Below is the "Problem part of the Backup log", however if you need I will send you full log.
    Hope you will find out problem soon
    #FILE..... /oracle/ANP/sapdata2/sr3_8/sr3.data8
    #SAVED.... sr3.data8  ANPB260810/12
    BR0280I BRBACKUP time stamp: 2010-08-27 11.03.22
    BR0063I 9 of 51 files processed - 19400.070 MB of 135986.469 MB done
    BR0204I Percentage done: 14.27%, estimated end time: 11:47
    BR0001I *******___________________________________________
    BR0202I Saving /oracle/ANP/sapdata2/sr3_9/sr3.data9
    BR0203I to /dev/rmt0.1 ...
    BR0278E Command output of 'LANG=C dd obs=1024K bs=1024K if=/oracle/ANP/sapdata2/sr3_9/sr3.data9 of=/dev/rmt0.1':
    dd read error: I/O error
    462+0 records in
    462+0 records out
    BR0280I BRBACKUP time stamp: 2010-08-27 11.03.36
    BR0279E Return code from 'LANG=C dd obs=1024K bs=1024K if=/oracle/ANP/sapdata2/sr3_9/sr3.data9 of=/dev/rmt0.1': 2
    BR0222E Copying /oracle/ANP/sapdata2/sr3_9/sr3.data9 to/from /dev/rmt0.1 failed due to previous errors
    BR0280I BRBACKUP time stamp: 2010-08-27 11.03.41
    BR0317I 'Alter tablespace PSAPSR3 end backup' successful
    BR0056I End of database backup: bedziydp.ant 2010-08-27 11.03.36
    BR0280I BRBACKUP time stamp: 2010-08-27 11.03.41
    BR0054I BRBACKUP terminated with errors
    Warm Regards
    Ahsan

    Hi,
    since you are getting a read error, it might as well be, that your datafile is defective.
    Try the same dd-command to /dev/null, to see if it is possible to read the entire file.
    First make sure that your null-device is existing, otherwise you might face a root-fs full problem.
    dd obs=1024K bs=1024K if=/oracle/ANP/sapdata2/sr3_9/sr3.data9 of=/dev/null
    or try a dbverify on it, which would also read the entire file and do a checksum test.
    brbackup -c -u / -m /oracle/ANP/sapdata2/sr3_9/sr3.data9 -t online -w only_dbv
    Good luck
    Volker

  • Risks associated with database refresh.

    Hi experts,
    I need to clarify a  doubt .I have a cloned SAP BW system (from another system i.e SAn to SAN copy)and then after few months after the first clone I need to refresh the database(Oracle 10 G).The reason being I will have some changes in data at the source system from where it was copied earlier.I shall be applying the application changes by tracking the transport requests(from the original system).
    I need to clarify if we don't refresh the whole system again  by SAN to SAN copy and then do  a database refresh will there be any problem .Is there any risk associated with it.
    Will the requests be available to me on the monitor with the new data or there will be some issues.
    A quick response shall be very useful.
    Shailja.

    Hello,
    Orders for Production , Maintenance, Deliveries, Transfer Order  etc. are created according to the factory calendar .
    You must be sure that all operations can be carried out in the seven days of the week (e.g. deliveries), because SAP can be scheduling aTransfer Order for a Sunday and somebody can be calling you for fixing the issue!
    The normal case is that some operations can take place in the weekend (Production, Maintenance) but other Logistic Operations ( Deliveries) are done only from Mo to Fr. If this is your case, I would create an additional seven days calendar and assign it to the work centers in Production and maintenance, but the Plant would remain linked to the old five days calendar
    Hope this can solve your issue.
    Dario

  • Database refresh from Prod to UAT

    Please check whether i have missed out any step in database refresh in 11.5.10.2
    Step 1 preclone executed on source Instance
    Step 2 copied all datafiles in backup mode to UAT instance
    Step 3 controlfile create and opened db in resetlogs.
    Step 4 adcfgclone.pl dbconfig <target context file>
    Step 5 Autoconfig executed on both db aswell apps
    Is it required to execute adcfgclone.pl dbTechstack and adcfgclone.pl appsTier
    environment is apps 11.5.10.2
    db 10g
    os AIX 5.3

    Hello,
    You will need to run:
    perl adcfgclone.pl dbTier [Database Target Context File]
    perl adcfgclone.pl appsTier [Appltop Target Context File]
    and other tasks as per Section 4: Advanced Cloning Options / 1. Refreshing a Target System
    of Note 230672.1 : "Cloning Oracle Applications Release 11i with Rapid Clone"
    Regards, Carlo.

  • Database Refresh From ASM Filesystem to Local Filesystem

    Hi ALL,
    I am performing a database refresh from production server to a demo server. Our Production database is 11.2.0.1 and it is using ASM filesystem to keep the data, redo and other files in ASM disks.
    On the other hand demo server is not having ASM, all the database files are stored in a local filesytem.
    I have taken a fresh backup of our production database, but I am not sure to perform the restore part as the demo server is not having ASM.
    Can anyone suggest me how to perform this, I mean datafile restore from ASM to local filesystem.
    Any usefull links will be helpful.
    Regards,
    Arijit

    Hello,
    You can restore the backup of your Production database which is using ASM to your demo server (using file system).
    Make sure that the control_files parameter in the pfile/spfile is pointing to the file system location where you want to have contol files located on the demo server.
    Next, before you use the restore command to restore the database, provide the location of the datafile where you need to restore using the "set newname" clause.
    run
    set newname for datafile 1 to '<file-system-location-on-demo-server>';
    set newname for datafile 2 to '<file-system-location-on-demo-server>';
    restore database;
    switch datafile all;
    recover database;
    }

Maybe you are looking for

  • Can I use the camera on my iPod Touch with FaceTime?

      I have been trying to use the camera on my iPod Touch with FaceTime.  Will it work?  How do I sign in?  When I click on Preferences the preferences window does not appear, and preferences in the drop down menu becomes greyed out.  I am using an old

  • How Journal Import groups records from GL_INTERFACE into Journal Headers

    Hello, Does anyone know how or using what criteria the Journal Import program groups records from the GL_INTERFACE into Journal Entries? The R12 User Guide says that : REFERENCE4 (Journal entry name): Enter a journal entry name for your journal entry

  • Help: IMAQdx scale image size

    HI all, i'm new in LebVIEW and have a question. Im using the functions of IMAQdx for a camera (XCD-U100 UXGA black&white). Is it possible to take images with smaller size? I want to increase the frame-rate. I can scale the image, after it was taken i

  • New block size

    <p>hi</p><p>can anyone tell me</p><p>if I add a dynamic calc member to Dense dim member(stored), what will happen to my block size ?</p><p> </p><p>thanks and regards</p><p>balu</p>

  • Return formatted text from LOV

    I have text stored in a table as a varchar2 with carriage returns in the text using utl_tcp.CRLF. This works fine when I send an email using this text thru a procedure. I am trying to load this text to an unbound item (defined as char) thru an LOV cr