Using old file system backup for Cloning

I have taken an off-line backup of Oracle 11i (11.5.10.2) 15 days ago. Before taking backup of file system , I verified that all the latest Rapid Clone Patches are applied. No changes or patch work in APPL_TOP or DB has been done since that backup. Now I need to do cloning of this instance, how I can use this backup for Cloning.
Rapid Clone scripts create and generate some files/directories so I am not sure whether my Old backup of file system will work or not. What is the best way to use old backup for cloning , also what are the file and directories in addition to the old backup of file system which I need to copy to Target System.
Thanks for reviewing and suggestions.
Samar

Samar,
If you have run preclone before backing it up, your backup should be valid for cloning.
2.1 in the cloning doc has to be in the backup.
These doc's should clear out yours doubts on cloning.
Cloning Oracle Applications Release 11i with Rapid Clone
https://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=230672.1
FAQ: Cloning Oracle Applications Release 11i
https://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=216664.1

Similar Messages

  • How to recover DB2 using offline file system backup.

    Gurus,
    I had a problem with my BI PRD's database. 1 of my BI consultant accidently deleted some important 2010's data. We've try recover online database backup of BWP into BWQ and retrieve the missing data. However the backup retention period was expired and we dont have any available online backup now.
    Only hope is my offline file system backup which backup the whole file system of BWP ONLY.
    My question:-
    Can i restore the file system (offline backup) into BWQ and rebuild my DB2 in BWQ. Please enlighten me.
    Thanks,
    Devan.

    Hi
    You can do it .The file system backup image can be mirroed into production system and by making using of "db2inidb" utility you can restore the whole database.
    http://publib.boulder.ibm.com/infocenter/db2luw/v8/index.jsp?topic=/com.ibm.db2.udb.doc/core/r0004473.htm
    Thanks
    Romansh

  • How to Create a OLAP Cube in DEV using SSAS from Raw file system backup from Production?

    How to Create a OLAP Cube in DEV using SSAS from Raw file system backup from Production? I dont have a .abf file available. Two paritions in production are missing data. We were able to get back file system backup which contains the files for these two paritions.
    How do I create a cube in Dev using this file system backup.
    we are on SQL Server 2008R2.
    Thanks,

    How to Create a OLAP Cube in DEV using SSAS from Raw file system backup from Production? I dont have a .abf file available. Two paritions in production are missing data. We were able to get back file system backup which contains the files for these two paritions.
    How do I create a cube in Dev using this file system backup.
    we are on SQL Server 2008R2.
    Thanks,

  • Warming up File System Cache for BDB Performance

    Hi,
    We are using BDB DPL - JE package for our application.
    With our current machine configuration, we have
    1) 64 GB RAM
    2) 40-50 GB -- Berkley DB Data Size
    To warm up File System Cache, we cat the .jdb files to /dev/null (To minimize the disk access)
    e.g
         // Read all jdb files in the directory
         p = Runtime.getRuntime().exec("cat " + dirPath + "*.jdb >/dev/null 2>&1");
    Our application checks if new data is available every 15 minutes, If new Data is available then it clears all old reference and loads new data along with Cat *.jdb > /dev/null
    I would like to know that if something like this can be done to improve the BDB Read performance, if not is there any better method to Warm Up File System Cache ?
    Thanks,

    We've done a lot of performance testing with how to best utilize memory to maximize BDB performance.
    You'll get the best and most predictable performance by having everything in the DB cache. If the on-disk size of 40-50GB that you mention includes the default 50% utilization, then it should be able to fit. I probably wouldn't use a JVM larger than 56GB and a database cache percentage larger than 80%. But this depends a lot on the size of the keys and values in the database. The larger the keys and values, the closer the DB cache size will be to the on disk size. The preload option that Charles points out can pull everything into the cache to get to peak performance as soon as possible, but depending on your disk subsystem this still might take 30+ minutes.
    If everything does not fit in the DB cache, then your best bet is to devote as much memory as possible to the file system cache. You'll still need a large enough database cache to store the internal nodes of the btree databases. For our application and a dataset of this size, this would mean a JVM of about 5GB and a database cache percentage around 50%.
    I would also experiment with using CacheMode.EVICT_LN or even CacheMode.EVICT_BIN to reduce the presure on the garbage collector. If you have something in the file system cache, you'll get reasonably fast access to it (maybe 25-50% as fast as if it's in the database cache whereas pulling it from disk is 1-5% as fast), so unless you have very high locality between requests you might not want to put it into the database cache. What we found was that data was pulled in from disk, put into the DB cache, stayed there long enough to be promoted during GC to the old generation, and then it was evicted from the DB cache. This long-lived garbage put a lot of strain on the garbage collector, and led to very high stop-the-world GC times. If your application doesn't have latency requirements, then this might not matter as much to you. By setting the cache mode for a database to CacheMode.EVICT_LN, you effectively tell BDB to not to put the value or (leaf node = LN) into the cache.
    Relying on the file system cache is more unpredictable unless you control everything else that happens on the system since it's easy for parts of the BDB database to get evicted. To keep this from happening, I would recommend reading the files more frequently than every 15 minutes. If the files are in the file system cache, then cat'ing them should be fast. (During one test we ran, "cat *.jdb > /dev/null" took 1 minute when the files were on disk, but only 8 seconds when they were in the file system cache.) And if the files are not all in the file system cache, then you want to get them there sooner rather than later. By the way, if you're using Linux, then you can use "echo 1 > /proc/sys/vm/drop_caches" to clear out the file system cache. This might come in handy during testing. Something else to watch out for with ZFS on Solaris is that sequentially reading a large file might not pull it into the file system cache. To prevent the cache from being polluted, it assumes that sequentially reading through a large file doesn't imply that you're going to do a lot of random reads in that file later, so "cat *.jdb > /dev/null" might not pull the files into the ZFS cache.
    That sums up our experience with using the file system cache for BDB data, but I don't know how much of it will translate to your application.

  • Linux Cluster File System partitions for 10g RAC

    Hi Friends,
    I planned to install 2 Node Oracle 10g RAC on RHEL and I planned to use Linux File system itself for OCR,Voting Disk and datafiles (no OCFS2/RAW/ASM)
    I am having SAN storage.
    I would like to know how do i create shared/cluster partitions for OCR,Voting Disk and Datafiles (common storage on SAN).
    Do i need to install any Linux cluster file system for creating these shared partitions (as we have sun cluster in solaris)?
    If so let me know what versions are supported and provide the necessary Note / Link
    Regards,
    DB

    Hi ,
    Below link may be useful to you:
    ORACLE-BASE - Oracle 10g RAC On Linux Using NFS

  • Is possible control tape library slot 1 - 10 for file system backup

    hi ..
    i am new to osb , i just install and setup osb , i have a question as below , hope expert can help me
    env: testing
    rhel 5.5
    tape library with 20 slot
    file system backup
    1. is possible osb only use slot 1 - 10 for file system backup ? amanda can control slot x - slot y for the configuration .
    2. how do i label the tape for slot 1 - slot 10 by obtool ? how to control osb auto load the tape for next backup ? where to check the log say that next tape is tape-02 ?
    thanks ..

    hi dcooksey
    how do i use list for a tape drive...for example, if you want tape drive A to only use slots 1-10 from obtool or webtool ?
    becoz i new to backup solution & osb ( i always use ghost or acronis to clone the image ) , my thinking as below , pls correct me if i am wrong
    slot 1 - 10 for daily backup
    slot 11-16 for full system backup
    slot 17 - 20 reserve ( this tape only use for full system backup before perform any application upgrade patches , )
    daily backup mon - fri ( 2 week ) ( no backup on saturday and sunday ) , server application offline
    full system backup friday ( 1 , 14 on calendar ) every 2 week perform full system backup after daily backup completed
    for upgrade application ,
    perform full system backup after daily backup , then release the server to application team to perform upgrading .
    so how to set my media family for the above setting ? the slot configuration is control by media family ?
    hope you can help ...

  • FILE SYSTEM BACKUP IS HANGING ON OSB

    Hi all
    i am having a test setup for OSB ( using windows 2003 ent 32 bit , administrative server + mediaserver ) , i am using HP c1537a scsi Tape drive ( driver details Microsoft 10/1/2012 version 5.2.3790.0 ) , and while firing the file system backup the job is getting hanged on the webtool,
    i am attaching the logs
    WEB TOOL LOG
    admin/5.1     
    Type     backup osb-server
    Level     full
    Family     FS-BKP-MF
    Encryption     off
    Scheduled time     04/19.11:29
    Introduction time     2012/04/19.16:59
    Earliest exec time     04/19.11:29
    Last update time     2012/04/19.17:23
    Expire time     never
    State     running since 2012/04/19.17:23
    Priority     100
    Privileged op     yes
    Run on host     osb-server
    Requires     host osb-server and family FS-BKP-MF and any device
    Attempts     2
    Log     
         2012/04/19.16:59:49 Job created.
         2012/04/19.16:59:49 Dispatching job to run on administrative server.
         2012/04/19.17:04:03 Drive or volume on which mount attempted is unusable.
    ob> lsdev
    drive tc-tape in service
    ob> pingd tc-tape
    Info: drive tc-tape accessible.
    ob>
    ob> lsdev -lg
    tc-tape:
    Device type: tape
    Model: [none]
    Serial number: [none]
    In service: yes
    Automount: yes
    Error rate: 8
    Query frequency: 1024KB (1048576 bytes) (from driver)
    Debug mode: no
    Blocking factor: (default)
    Max blocking factor: (default)
    UUID: 26934f4f-fe7b-49ce-a3ed-034a7c9ad046
    Attachment 1:
    Host: osb-server
    Raw device: //./obt0
    Connection type: SCSI
    Inquiry data:
    Vendor: HP
    Product: C1537A
    Firmware: L111
    Tape state: online, write protected
    Hardware compression: on
    Last read was: compressed
    Hardware encryption: not available
    Maximum block size: 61440
    Remaining tape: 3615300 (uncompressed) blocks (3.70GB)
    ob>
    please guide me resolve this issue ..
    Regards
    ASA
    Edited by: 927832 on Apr 19, 2012 5:10 AM

    Hi Rich
    Thanks for your reply.
    since i am using a stand alone Tape drive, we will not able to take the inventory , isn't it ?
    i hope that my media family parameters are correct only .please find the media family details below.
    media family name :- test-media-family
    volume id used :- unique to this media family
    volume expiration :- 3 years
    appendable :- yes
    rotation policy & volume duplication policy are ;- none
    ====================================
    While i am mounting the Tape to the drive is gives the following error
    ob> mountdev -w test-tape
    Error: can't mount volume in test-tape - timed out waiting for bot
    ob>
    Edited by: 927832 on Apr 20, 2012 3:04 AM

  • OS File System Backup

    Hi,
    Apps: 12.1.3
    DB-11gR1
    OS - RHEL 5.6
    How i can take OS file system backup manually?-
    which is the best mothod?
    I know, we can use tar?
    could anyone tell me which is the best method..
    Thanks
    KSV.DBA

    In addition to the link mentioned above, do you have to rely on free solutions, or do you have a budget for a solution?
    You can perform a complete OS level backup of your Oracle database when it is shutdown, or put the database into online backup mode, but need to review the Oracle documentation for limitations. The tools of choice for Oracle is RMAN, which does not require extra logging or backup mode.
    You can use tar or rsync if you prefer to backup data to a shared storage through NFS or use tar to stream the data to an attached tape drive. There is plenty of information and tutorials on how to perform incremental backups on the web, and here tar incremental backup

  • Unable to get the file system information for: \\****servername\E$\; error = 64 Unable to distribute content to DP

    One of our DPs has stopped loading content. 
    I've research for quite a bit and cannot find a clear cut reason to this.  This server only has a DP role, I verified sharing permissions, all looked good. This DP has been running just fine for the last year or so and all sudden it will no longer load
    packages.  The disk drive is still present I can still reach the hidden share \\servername.com\E$
    Verified that the SMSSIG$ folder is there and the last entry is from 4/23/2015 
    SCCM 2012 R2 
    OS 2008 R2 Standard
    Any help is greatly appreciated!
    Here's a snipit from the distmgr.log
    Start updating the package on server ["Display=\\*****.com\"]MSWNET:["SMS_SITE=1AB"]\\*****.com\...
    Attempting to add or update a package on a distribution point.
    Will wait for 1 threads to end.
    Thread Handle = 0000000000001E48
    STATMSG: ID=2342 SEV=I LEV=M SOURCE="SMS Server" COMP="SMS_DISTRIBUTION_MANAGER" SYS=***.com SITE=1AB PID=2472 TID=8252 GMTDATE=Thu Apr 30 19:12:01.972 2015 ISTR0="SYSMGMT Source" ISTR1="["Display=\\*****.com\"]MSWNET:["SMS_SITE=1AB"]\\*****.com\"
    ISTR2="" ISTR3="" ISTR4="" ISTR5="" ISTR6="" ISTR7="" ISTR8="" ISTR9="" NUMATTRS=2 AID0=400 AVAL0="CAS00087" AID1=404 AVAL1="["Display=\\*****.com\"]MSWNET:["SMS_SITE=1AB"]\\*****.com\"
    SMS_DISTRIBUTION_MANAGER 4/30/2015 2:12:01 PM
    8252 (0x203C)
    The current user context will be used for connecting to ["Display=\\*****.com\"]MSWNET:["SMS_SITE=1AB"]\\*****.com\.
    Successfully made a network connection to \\*****.com\ADMIN$.
    Ignoring drive \\*****.com\C$\.  File \\*****.com\C$\NO_SMS_ON_DRIVE.SMS exists.
    Unable to get the file system information for: \\*****.com\E$\; error = 64.
    Failed to find a valid drive on the distribution point ["Display=\\*****.com\"]MSWNET:["SMS_SITE=1AB"]\\*****.com\
    Cannot find or create the signature share.
    STATMSG: ID=2324 SEV=E LEV=M SOURCE="SMS Server" COMP="SMS_DISTRIBUTION_MANAGER" SYS=sccmprdpr1sec2.mmm.com SITE=1AB PID=2472 TID=8252 GMTDATE=Thu Apr 30 19:12:55.206 2015 ISTR0="["Display=\\*****.com\"]MSWNET:["SMS_SITE=1AB"]\\*****.com\"
    ISTR1="CAS00087" ISTR2="" ISTR3="30" ISTR4="94" ISTR5="" ISTR6="" ISTR7="" ISTR8="" ISTR9="" NUMATTRS=2 AID0=400 AVAL0="CAS00087" AID1=404 AVAL1="["Display=\\*****.com\"]MSWNET:["SMS_SITE=1AB"]\\*****.com\"
    Error occurred. Performing error cleanup prior to returning.
    Cancelling network connection to \\*****.com\ADMIN$.

    Error 64 is being returned which is simply "the network name is no longer available".
    There can be a number of reasons for this from SMB compatibility issues (2003 servers wont support SMB2), to the expected and actual computer name of the boxes don't match (tries to authenticate with server.tld.com when the actual name is srv-01.tld.com and
    you just put a C-name in). I'd start from the top:  Try opening said share from the Primary Site server as that's the box doing the work.  Verify the IP and computer name is legit and that no one has played ACL games between the two systems (remember
    RPC only initiates/listens on port 135 but established connections are up in the dynamic port range).
    At the end of the day it's an issues "underneath" SCCM, and not an SCCM problem specifically. 

  • File System backup and restore command

    Hi All,
    I require File System Backup and Restore command in AIX to Trivoli Storage Manager.Please send me the commands.
    Thanks in Advance
    Regards,
    Soumya

    Please check this out
    Restore
    http://www.ahinc.com/aix/backup.htm
    restore -xvf/dev/rmt0 Restores all the files located on the tape device that were backed up using the backup command.
    Backup
    http://publib.boulder.ibm.com/infocenter/systems/index.jsp?topic=/com.ibm.aix.cmds/doc/aixcmds1/backup.htm
    Hope this will help you.
    Regards
    dEE

  • Integration Directory Objects Transport Using the File System

    Hi
    I am trying to do the IR and ID Objects Transport Using the File system... I am able to Transport the IR Objects and IMport in to QA ... but when i try to Import ID objects its giving me the error...bcoz all the ID objects pointing to theSAP R/3 DEV system ( Interface Detremination, receiver determination ...)... do i have to config something in SLD ..so when i import the ID objects in to QA system it points to correct SAP R/3 QAS system...
    and we are using single SLD for all the DEV , QAS and PRD system
    SV
    Edited by: S V on Oct 15, 2008 4:08 AM

    Hi,
    You will have to configure transport targets like for different business systems in your dev environment, there shouls be corresponding Business systems in QUA environment. For more details refer the link
    http://help.sap.com/saphelp_nw2004s/helpdata/en/ef/a21e3e0987760be10000000a114084/frameset.htm
    Thanks!

  • How to make bank management system using java file system

    Hi, I have some fields
    1.ID
    2. Deposite
    3. Withdraw
    4. Balance
    Now how can i manage this Bank Management System using java file system.
    Thanks in advance.

    Then we're back to (1): Do your own homework. Google has zillions of links on handling files in Java. When you have written some code and have an actual problem, we'll be happy to help you with it.
    (edit) Incidentally, this sounds suspiciously like the sort of problem they set for the certification programs. In which case, don't bother; they're not worth the virtual paper they're printed on.

  • Oracle Database Point intime recovery with File System backup.

    Hi.. Experts...
    We have shutdown the database and have take a file system backup on
    09.10.2010 @ 10 AM.
    We started the database and all the  archive logs generated after that are already
    available in archive directory till now.
    We discovered there were some  deletions took place. Yesterday evening at 4 PM.
    Requirement.
    We want to recover the database till  yesterday 3.30 PM.
    Can any one send steps to recover with above backup and available archive logs till 3.30 PM Yesterday.
    Regards

    Experts.....Nick Loy, Varadharajan M, Mark & Volker.............Thanks U Very Much.
    TAIL END MESSAGE OF RECOVERY COMMAND
    ORA-00279: change 11331655931 generated at 10/10/2010 11:30:56 needed for
    thread 1
    ORA-00289: suggestion : /oracle/D10/oraarch/D10arch1_123487_607825009.dbf
    ORA-00280: change 11331655931 for thread 1 is in sequence #123487
    ORA-00278: log file '/oracle/D10/oraarch/D10arch1_123486_607825009.dbf' no
    longer needed for this recovery
    Log applied.
    Media recovery complete.
    SQL>
    SQL>
    SQL> ALTER DATABASE OPEN RESETLOGS;
    Database altered.
    SQL>
    SYSTEM RECOVERED TO DEMAND & RELEASED
    T H A N K S    --   T O    --   E V E R Y O N E .
    RGDS

  • Help needed. I have a 3TB external hard drive partitioned into 1TB's. I used one partition as backup for the main computer hard drive. But now it is greyed out as shown in the disk utility and its name changed, i dont know how that happened, i tried verif

    Help needed. I have a 3TB external hard drive partitioned into 1TB's. I used one partition as backup for the main computer hard drive. But now it is greyed out as shown in the disk utility and its name changed, i dont know how that happened, i tried verifying and repairing it but had no luck. I also tried mounting and unmounting it but still no solution. Anyone to help please?

    Looks bad. I would strongly advise backing up the stuff on the other two partitions as soon as possible in case the rest of the drive goes wrong.
    Beyond that, Disk Utility has limited abilities to repair disks, DiskWarrior uses a different approach and might, emphasis might be more successful. It has certainly done the job for me in the past when Disk Utility could not.
    See http://www.alsoft.com/diskwarrior/

  • How can i use JS files in ADF for language translation.

    Hi,
    I have JS for different languages and dn't want to convert them to property files(resource bundle files). How can i use JS files in ADF for language translation.
    Thanks

    Hi ILya Cyclone,
    Thanks alotfor the reply. Can you tell me where should i include this in the jspx page.
    Step 1)
    I have the js file as js/ifl_messages_US.js and i created a resource file as you mentioned: JS_FILE_PATH=js/ifl_messages_US.js
    Step 2)
    Then added the entry in faces-config.xml for the resource file as follow:
    <resource-bundle>
    <base-name>resource_en.properties</base-name>
    <var>resource</var>
    </resource-bundle>
    <locale-config>
    <supported-locale>en</supported-locale>
    </locale-config>
    Step 3) This is my jspx page. In which a table is dynamically created on page load. Can you help me where should i enter the "#{resource.JS_FILE_PATH}"
    <?xml version='1.0' encoding='UTF-8'?>
    <jsp:root xmlns:jsp="http://java.sun.com/JSP/Page" version="2.1"
    xmlns:f="http://java.sun.com/jsf/core"
    xmlns:h="http://java.sun.com/jsf/html"
    xmlns:af="http://xmlns.oracle.com/adf/faces/rich">
    <jsp:directive.page contentType="text/html;charset=UTF-8"/>
    <f:view>
    <af:document id="d1">
    <af:messages id="m1"/>
    <af:resource type="javascript" source="/js/pdfFile.js"/>
    <af:form id="f1">
    <input type="hidden" name="checkRadio" id="checkRadio" value=""/>
    <af:panelGroupLayout id="pgl1" halign="left" layout="vertical">
    <af:image source="/images/BRAND_IMAGE.gif" id="i1"/>
    </af:panelGroupLayout>
    <af:spacer width="10" height="10" id="s1"/>
    <af:table varStatus="rowStat" summary="table"
    value="#{backingBeanScope.DummyBean.collectionModel}"
    rows="#{backingBeanScope.DummyBean.collectionModel.rowCount}"
    rowSelection="none" contentDelivery="immediate" var="row"
    rendered="true" id="t1" styleClass="AFStretchWidth"
    binding="#{backingBeanScope.DummyBean.myTableBinding}"
    columnResizing="disabled">
    <af:column id="c2" headerText="Actions">
    <af:activeOutputText value="#{row.Actions}" id="aot2"/>
    <af:goLink text="#{row.Actions}" id="gl1"
    clientComponent="true" visible="false"/>
    <af:selectBooleanRadio text="" id="sbr1"
    valueChangeListener="#{backingBeanScope.DummyBean.checkselectbox}">
    <af:clientListener method="selectCheckBox" type="click"/>
    </af:selectBooleanRadio>
    </af:column>
    <af:forEach items="#{backingBeanScope.DummyBean.columnNames}" end="#{backingBeanScope.DummyBean.columnSize}"
    var="name" begin="1">
    <af:column sortable="false" sortProperty="#{name}"
    rowHeader="unstyled" headerText="#{name}"
    inlineStyle="width:100px;" id="c1">
    <af:activeOutputText value="#{row[name]}" id="aot1" escape="false">
    </af:activeOutputText>
    <!-- <af:outputFormatted value="#{row[name]}" id="of1"/>-->
    <!--<af:goLink text="goLink 1" id="gl1"
    destination="#{row.bindings.url.inputvalue}"/>-->
    </af:column>
    </af:forEach>
    </af:table>
    </af:form>
    </af:document>
    </f:view>
    </jsp:root>
    Thanks in advance

Maybe you are looking for