Reduce export import time

Hl All,
I want to reduce the total time spent during export-import.
Restrictions:
=========
1)Cross Platform
2)exp and imp from lower version to upper version
3)No 10G database
Basically i want to do exp and imp in parallel so it reduces the total time spent on this activity. I thought of doing schema level exp-imp in parallel but i am afraid of the dependencies.
Is there any other way to achieve the same or if i go with the above specified approach, can anyone provide some valuable inputs to that.
I am trying to automate the above so that it becomes one time effort and rest all the times it(script) should do at his own.
Thanks and regards
Neeraj

Hi All,
Data volume is not less than 60GB and not more than 150GB.
If i use a pipe on unix in between exp-imp what if my exp is slower at any
point of time and import is faster(Any reason)?
is import going to wait for the contents coming into the pipe through export or import will fail.
I can consider creating indexes using the flat file.
is there any way to get only the indexes in the flat file, i mean if i use indexfile
option for import it gives me "CREATE TABLE..." statements too which means import utility is reading full
dumpfile and giving me the output, i want only the "CREATE INDEX..." statement in the flat file.
What about the schema level export and import?
Any valuable inputs or proper steps from anyone out there
Any restrictions while importing the schema's.
Thanks and Regards

Similar Messages

  • Export / Import Time

    Hi All
    While exporting (EXP) and Importing (IMP) from oracle 10g, can we know how much time remaining for the export/import.
    regards,
    Mudassir

    i can only estimate the size and not the timeWell if you have a size estimate... start DP again in normal mode monitor the size and once you have worked out the MB/per second rate you will be able to use the estimate size to get an end time.
    Kind regards and I hope this helps.

  • Estimating Unicode Export / Import time

    hithere,
    Is there a good methog of estimating the unicode export/import runtime . for a 9cpu system with 500 GB of database size , what is the thumbrule of calculating the runtime.
    welcome your ideas and suggestion..
    Thanks

    Hi,
    You can also use [SAP Downtime Assessment Service|https://websmp105.sap-ag.de/upgradeservices] from SAP.
    Regards,
    Somar

  • How to Reduce import time

    Hi All,
    Import time of a specific dump file is 1.5 Hr on oracle server 9i realese2 OS is United Linux. While the same file is imported on another machine having oracle server 9i realese2 OS is windows XP in 15 min.
    Although the specification of Linux machine is better than windows machine.
    (Import Utility is executed from a machine having windows2000 for both servers)
    Now how can i reduce the import time for Linux server.

    Importing from linux machine takes more time than it imported from a client machine.
    Also i case of window server import is done from a network machine from same machine.

  • Export/Import  Object configuration

    hi gurus!
    I have the next landscape, and I´ll want to stop make manually the export of configuration objects:
    XI DEV    -
        XI QAS     -
        XI CORP   
    BS_DSP060          BS_QSP060               BS_SIF060
    The problem is that in each enviroment each Business system has diferent name. so that´s my problem, how could I make
    that this Business systems recognize in automatically at the export/import time?
    Thanks

    Hi,
    Refer this..
    How to handle SLD for XI
    https://websmp108.sap-ag.de/~form/sapnet?_SHORTKEY=01200252310000071155&_SCENARIO=01100035870000000202
    Preparation of SLD for Transports Using CMS in XI
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/b01876ac-abbb-2a10-788e-95a7c5a5aeca
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/300e3ca7-31cd-2a10-a6ba-87c7447d5fd7
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/9e76e511-0d01-0010-5c9d-9f768d644808
    Might help you..
    Regards,
    vasanth.

  • Is time to export-import location dependent

    I had a doubt that does the export-import process depend on the location from where it is performed. I will try to explain the scenario. There is export-import to be done of a server in Africa as the migration to a new server is required. My question is does it make a difference in the time it takes to export-import, if the SecureCRT session is logged on from here, Africa itself or if the process is done from a far off location, say, India as some of our DBA team is in India.
    So, the doubt is does the export-import process depend on the location from where it is performed?
    I hope, my question is clear.
    Please, help in solving the doubt.
    regards

    You need to clarify from an Oracle perspective what the client machine is in this scenario (i.e. the machine that is connecting to the Oracle database). What machine is the export utility running on and what machine is the export utility writing to?
    If the export utility is running on the same machine and writing to a dump file on the same machine regardless of where the DBA is sitting, then where the DBA is sitting is irrelevant. If the export utility runs on a different machine or writes to a dump file on a different machine depending on where the DBA is sitting, then where the DBA is sitting is relevant.
    Justin

  • I have purchased Asus STRIX-GTX980. Now what should I change settings in Premier Pro and After Effects to reduce export time

    I have recently purchased Asus STRIX-GTX980 4GB with i7 4790K.
    Now what settings are to be optimized in Pr  (CS6) to reduce Export time . Please explain in Detail Please explain in Detail step by step
    Now what settings to be optimized in Ae (CS6) to reduce Export time . Please explain in Detail step by step .
    The main Motto of noto is to reduce expot times.
    Please help me set values .

    As Dave says, GPU does very little in AE, see this: GPU (CUDA, OpenGL) features in After Effects
    Now, some third-party plugins (such as Red Giant Universe and Video Copilot's Element 3d) do make use of the GPU in AE, but their requirements have nothing to do with AE or its settings.
    Premiere, on the other hand, does make use of a GPU, you should ask in that forum what they suggest.
    Also, in the future, I would suggest checking on system requirements and benefits BEFORE making a purchase...

  • HT201250 How do I restore photos from Time Machine?  All of my photos have been imported within iPhoto and therefore are exported to Time Machine via an iPhoto file.  I can't figure out how to break individual photo files out of the iPhoto file.  Any idea

    How do I restore photos from Time Machine?  All of my photos have been imported within iPhoto and therefore are exported to Time Machine via an iPhoto file.  I can't figure out how to break individual photo files out of the iPhoto file.  Any ideas?

    You might post this query on the iPhoto forum and perhaps someone there can provide an answer.
    I tried this on my MBP and it seems that you have to restore the entire iPhoto Library.  What I did was put the current library in a separate folder on the desktop and then I restored the library from a prior date.  If you were to do that, you then could extract the desired photos and then delete the restored library.  Then place the current library back to the original location and import the desired photos.
    Not tidy but it seems it will work.
    Ciao.

  • Export/Import Project still buggy in v2.1.3

    After seeing the update notes for v2.1.3 and noticing references to project importing, I started to get excited that maybe the bugs have been ironed out. Unfortunately my longstanding bug still remains.
    If anyone can think of any workarounds I'd be extremely grateful...
    I have a 2.16GHz Intel Core Duo MacBook Pro with 2GB RAM, running OS X 10.5.6 and now Aperture 2.1.3
    I have a project with many images, arranged into stacks, a pick in each stack with many adjustment and a 62-page book created from these picks. There are also some smart albums and a light-table.
    Now I export the project (i.e. in the Project Inspector I right-click the project and choose Export > Project) to a file on the desktop and rename the existing project in Aperture. I now import the saved .approject file and I find that all my Picks are no longer Picks and my book is therefore no longer made up of my desired images. (I've tried this many times now and ruled out the possibility of a corrupt export file).
    As a result, when I select the Book I get the "Non-Pick Images" dialog with the dreaded "One or more placed items are no longer the pick of a stack." message. "Use Current Pick" ruins the book - I have to go through and work out all the picks again (and also re-pan all the images within the Photo Boxes as this data has also been lost). "Create new Version" appears to preserve my book's original look, but my project now contains a couple of hundred more images than it used to and my other albums and light-table still have incorrect images shown as the Picks.
    Does anybody have any ideas of what I can do to ensure the stack-picks are preserved during the export/import process?
    (By the way, the reason I'm exporting and then re-importing is because I actually want to do the export from my laptop where the project is and then the import on my main work machine where the rest of my Aperture library lives, but that fails for the same reason, so I'm doing the export+import on my laptop for now to reduce the number of variables in the problem.)

    I go with the assumption that you now know how to create transport sets. After having created a transport set containing the objects you want to be exported, go to the administer tab under the main page of portal. There you will find the Export/Import portlet. Here you need to choose your transport set from the first LOV and then click EDIT to choose the "Security Option". Now you export out the stuff. This will export all the user page customizations for you.
    Thanks.

  • Export/Import subpartition stats

    I hope someone can give me a workaround for this, because it's causing our reports to take longer than they should!
    Background:
    We have some sub-partitioned tables on a 10.2.0.3 database, partitioned daily on the date column, with the subpartitions based on a list of values.
    Overnight, various reports are run. Each report loads its data into the table, and then produces a file based on the data that's been loaded for that report. It is not practical (IMO) to analyze the tables after each report has loaded its data, due to other reports loading their data at the same time.
    As the amount of data loaded into the tables each night does not vary significantly, we export the stats from a previous partition and import them into the new partition as part of the partition housekeeping job (stats imported from old partition, old partition gets dropped, new partition created with same name as the old one, and stats imported). This is done using dbms_sql.export_table_stats and dbms_sql.import_table_stats.
    However, one report which currently loads 43million rows is taking 4.5 hours to run. The size of the load file increases daily, but looking at the history of the report, each relatively small increase causes the report to run a disproportional amount longer (ie. an increase of a similar amount of rows on one night can add twice as much time onto the length of the report than the increase the previous night did).
    We've just implemented some changes to improve the buffer sizes, etc, on the database, in a bid to reduce some of the waits, but this has not improved matters much - the report now runs in 4 hours.
    We know this report can run faster, because in testing, we saw the report run in 60 minutes! Subsequent investigation shows that this was after the partitions had been analyzed, whereas the slow report ran prior to the partitions being analyzed, despite the stats being there for the partition.
    I have now tested the export/import stats process and found that they do not import the stats for the subpartitions. This looks like it is a large part of why the report takes longer before the relevant partitions/subpartitions have been analyzed than it does afterwards.
    Does anyone know of anyway that I can export/import the stats at a subpartition level? (I tried putting a subpartition name in the partition parameter, but I just got an error about it being an unknown partition name.)
    Any help, ideas or workarounds on this will be gratefully received!

    *** Duplicate Post - Please Ignore ***

  • System copy using R3load ( Export / Import )

    Hi,
    We are testing System copy using R3load ( Export / Import ) using our production data.
    Environment : 46C / Oracle.
    while executing export, it takes more than 20 hours, for the data to get exported, we want to reduce the export time drastically. hence, we want to achieve it by scrutinizing the input parameters.
    during export, there is a parameter by name splitting the *.STR and *.EXT files for R3load.
    the default input for the above parameter is No, do not split STR and EXT files.
    My question 1 : If we input yes instead of No and split the *.STR and *.EXT files, will the export time get reduced ?
    My Question 2: If the answer is yes to Question 1, will the time reduced will be significant enough ? or how much percentage of time we can expect it to reduce, compare to a No Input.
    Best Regards
    L Raghunahth

    Hi
    The time of the export depends on the size of your database (and the size of your biggest tables) and your hardware capacity.
    My question 1 : If we input yes instead of No and split the *.STR and *.EXT files, will the export time get reduced ?
    In case you have a few very large tables, and you have multiple cpu's and a decent disk storage, then splitting might significantly reduce the export time.
    My Question 2: If the answer is yes to Question 1, will the time reduced will be significant enough ? or how much percentage of time we can expect it to reduce, compare to a No Input.
    As you did not tell us about your database size and hardware details there is no way to give you anything but very basic metrics. Did you specify a parallel degree at all, was your hardware idling for 20 hrs or fully loaded already?
    20 hrs for a 100gb database is very slow. It is reasonable (rather fast in my opinion)  for a 2tb database.
    Best regards, Michael

  • Project duplication:  Copy/Paste vs. Export/Import

    Anyone have any thoughts on project duplication?
    Presently I am copying a project and then pasting it into multiple new locations then renaming them. Recently I was in a training session where the instructor suggested exporting the project and then importing it claiming database size savings. Can anyone shed any insight?
    Thanks in advance.

    I can't imagine why an export/import would be a smaller footprint within the database than a project copy/paste. Not to mention that if you are talking about a relatively large project it can take a substantial amount of time to export and import. One of our schedules that is about 18k activities takes approximately 1.5 hours to export/import where as a project copy/paste only takes about 10 minutes. Note that we do not copy the baseline schedules when replicating projects.

  • Is there a capability to save/export the time capsule settings file when using the iphone/ipad airport utility. the "file" button does not exist on the latest airport utility app.

    is there a capability to save/export the new airport 2TB time capsule settings file when using the iphone/ipad airport utility. set-up wasn't a problem but the "file" button does not exist on the latest airport utility app v6.3 to save the configuration file.

    the "file" button does not exist on the latest airport utility app v6.3 to save the configuration file.
    Sounds like you are a bit confused with version numbers.
    Latest AirPort Utility version for the iPhone / iPad is 1.3.3.  There is no option or capability to export/import settings on the iOS version(s) of AirPort Utility.....although you could take a series of screen shots and save them for future reference.
    AirPort Utility 6.3.x is found on a Mac.....not on iPhone / iPad. Export and Import options are found under the File menu in 6.3.x.

  • Problem with EXPORT IMPORT PROCESS in ApEx 3.1

    Hi all:
    I'm having a problem with the EXPORT IMPORT PROCESS in ApEx 3.1
    When I export an application, and try to import it again. I get this error message
    ORA-20001: GET_BLOCK Error. ORA-20001: Execution of the statement was unsuccessful. ORA-06550: line 16, column 28: PLS-00103: Encountered the symbol "牃慥整㈰㈯⼴〲㐰〠㨷㐵㈺′䵐" when expecting one of the following: ( - + case mod new not null <an identifier> <a double-quoted delimited-identifier> <a bind variable> avg count current exists max min prior sql stddev sum variance execute forall merge time timestamp in
    As a workaround, I check the exported file and found this
    wwv_flow_api.create_flow
    p_documentation_banner=> '牃慥整⠤㈰㈯⼴〲㠰〠㨷㠵㈺′äµ
    And when I replace with this
    p_documentation_banner=> ' ',
    I can import the application without the error.
    somebody knows why I have to do this??
    Thank you all.
    Nicolas.

    Hi,
    This issue seems to have been around for a while:
    Re: Error importing file
    I've had similar issues and made manual changes to the file to get it to install correctly. In my case, I got:
    ORA-20001: GET_BLOCK Error. ORA-20001: Execution of the statement was unsuccessful.<br>ORA-02047: cannot join the distributed transaction in progress<br>begin execute immediate 'alter session set nls_numeric_characters='''||wwv_flow_api.g_nls_numeric_chars||'''';end;There are several suggestions, if you follow that thread, about character sets or reviewing some of the line breaks within pl/sql code within your processes etc. Not sure what would work for you.

  • Regarding Distribution Monitor for export/import

    Hi,
    We are planning to migrate the 1.2TB of database from Oracle 10.2g to MaxDB7.7 . We are currently testing the database migration on test system for 1.2TB of data. First we tried with just simple export/import i.e. without distribution monitor we were able to export the database in 16hrs but import was running for more than 88hrs so we aborted the import process. And later we found that we can use distribution monitor and distribute the export/import load on multiple systems so that import will get complete within resonable time. We used 2 application server for export /import but export completed within 14hrs but here again import was running more than 80hrs so we aborted the import process. We also done table splitting for big tables but no luck. And 8 parallel process was running on each servers i.e. one CI and 2 App servers. We followed the document DistributionMonitorUserGuide from SAP. I observerd that  on central system CPU and Memory was utilizing above 94%. But on 2 application server which we added on that servers the  CPU and Memory utilization was very low i.e. 10%. Please find the system configuration as below,
    Central Instance - 8CPU (550Mhz) 32GB RAM
    App Server1 - 8CPU (550Mhz) 16GB RAM
    App Server2 - 8CPU (550Mhz) 16GB RAM
    And also when i used top unix command on APP servers i was able to see only one R3load process to be in run state and all other 7 R3load  process was in sleep state. But on central instance all 8 R3load process was in run state. I think as on APP servers all the 8 R3load process was not running add a time that could be the reason for very slow import.
    Please can someone let me know how to improve the import time. And also if someone has done the database migration from Oracle 10.2g to MaxDB if they can tell how they had done the database migration will be helpful. And also if any specific document availble for database migration from Oracle to MaxDB will be helpful.
    Thanks,
    Narendra

    > And also when i used top unix command on APP servers i was able to see only one R3load process to be in run state and all other 7 R3load  process was in sleep state. But on central instance all 8 R3load process was in run state. I think as on APP servers all the 8 R3load process was not running add a time that could be the reason for very slow import.
    > Please can someone let me know how to improve the import time.
    R3load connects directly to the database and loads the data. The quesiton is here: how is your database configured (in sense of caches and memory)?
    > And also if someone has done the database migration from Oracle 10.2g to MaxDB if they can tell how they had done the database migration will be helpful. And also if any specific document availble for database migration from Oracle to MaxDB will be helpful.
    There are no such documents available since the process of migration to another database is called "heterogeneous system copy". This process requires a certified migration consultant ot be on-site to do/assist the migraiton. Those consultants are trained specially for certain databases and know tips and tricks how to improve the migration time.
    See
    http://service.sap.com/osdbmigration
    --> FAQ
    For MaxDB there's a special service available, see
    Note 715701 - Migration to SAP DB/MaxDB
    Markus

Maybe you are looking for