Choose a database among various databases for export/import using.

Hello,
I am using forms 6i and I want to export several databases using
forms.
With regards to that what is the technique/code I could use to select
the database of my choice for export.
How could we get the name/services of a particular database among
various databases using Forms 6i.
Could some one give me the idea/code to get the above requirement?
Thanks
Amit

Why would you want to use Forms (a client tool) to import or export a database? Imp and exp are command line tools meant to be run on the database server.
You will probably be hitting other problems, like different database versions. For every database version you need the correct imp and exp.
If you really want to use Forms, than just make a table that hold the names of the databases.

Similar Messages

  • How to use lodercli for export import

    Hello,
    I want to migrate my content server data  , that is on SAPDB version 7.3 (32 bit)  to new server with MAX DB 7.6. (64 bit) .
    I wanted to know:
    1)  how to do export an d import using lodercli. i wanted to use loder of my new server i.e. MAXDB loder to do the export immport. Can any one tell me the steps and the command to do this ?
    2) I have refered SAP note 962019, but need some more help on the steps that is given in this note for hetrageneous copy.
    Note : i dont have any SAP system on the server i have only content server and hence i dont have user SAPR3 on database.
    Regards,
    Bhavesh

    > Please check the following (point 7) in note 962019:
    Ok...
    > 7. The procedure described in this note supports the following migrations:
    > and I am dealing with the same situation hence choose the export import rather than normal backup recovery.
    Well. it says: the procedure supports these migrations, which is true.
    It does not say: "please do use the export/import migration in these situations".
    Please do read the note point 1 where it says;
    "Homogeneous system copies (backup/recovery) can be executed only in
    systems that fulfill  the following conditions:
    a) The database versions in the source system and the target system
        do not differ.
    b) The processor uses the same byte sorting sequence (Little Endian
        -> Little Endian; Big Endian -> Big Endian).
    *If the specified operating system prerequisites are met in the
    source and target system, you MUST use the homogeneous system copy
    and not the heterogeneous system copy.*"
    So in your case export import is not even supported!
    > 2) I know the Cs 6.10 is out of support and it need to upgrade to new CS version that is 6.4 , but if i am already installing the new server for content server with CS 6.4 and MAXDB 7.6 why then why cant i use the export import from older server to new server??
    > Other wise i need to upgrade the MAXDB then upgrade CS and then take the backup and restore, isnt it more logivcal to do export import using loder as we have all prerequisites available??
    1. you don't meet the prerquesites (see above) - you can use backup/restore, so you have to !
    2. export/import is a terribly slow and clumsy procedure
    3. backup/restore is a easy, safe and fast method to perform the system copy
    If you still really, really want to stick with export/import, feel free to try it out.
    The note contain all required steps and the documentation for MaxDB even includes tutorials for the loader.
    regards,
    Lars

  • Regarding Distribution Monitor for export/import

    Hi,
    We are planning to migrate the 1.2TB of database from Oracle 10.2g to MaxDB7.7 . We are currently testing the database migration on test system for 1.2TB of data. First we tried with just simple export/import i.e. without distribution monitor we were able to export the database in 16hrs but import was running for more than 88hrs so we aborted the import process. And later we found that we can use distribution monitor and distribute the export/import load on multiple systems so that import will get complete within resonable time. We used 2 application server for export /import but export completed within 14hrs but here again import was running more than 80hrs so we aborted the import process. We also done table splitting for big tables but no luck. And 8 parallel process was running on each servers i.e. one CI and 2 App servers. We followed the document DistributionMonitorUserGuide from SAP. I observerd that  on central system CPU and Memory was utilizing above 94%. But on 2 application server which we added on that servers the  CPU and Memory utilization was very low i.e. 10%. Please find the system configuration as below,
    Central Instance - 8CPU (550Mhz) 32GB RAM
    App Server1 - 8CPU (550Mhz) 16GB RAM
    App Server2 - 8CPU (550Mhz) 16GB RAM
    And also when i used top unix command on APP servers i was able to see only one R3load process to be in run state and all other 7 R3load  process was in sleep state. But on central instance all 8 R3load process was in run state. I think as on APP servers all the 8 R3load process was not running add a time that could be the reason for very slow import.
    Please can someone let me know how to improve the import time. And also if someone has done the database migration from Oracle 10.2g to MaxDB if they can tell how they had done the database migration will be helpful. And also if any specific document availble for database migration from Oracle to MaxDB will be helpful.
    Thanks,
    Narendra

    > And also when i used top unix command on APP servers i was able to see only one R3load process to be in run state and all other 7 R3load  process was in sleep state. But on central instance all 8 R3load process was in run state. I think as on APP servers all the 8 R3load process was not running add a time that could be the reason for very slow import.
    > Please can someone let me know how to improve the import time.
    R3load connects directly to the database and loads the data. The quesiton is here: how is your database configured (in sense of caches and memory)?
    > And also if someone has done the database migration from Oracle 10.2g to MaxDB if they can tell how they had done the database migration will be helpful. And also if any specific document availble for database migration from Oracle to MaxDB will be helpful.
    There are no such documents available since the process of migration to another database is called "heterogeneous system copy". This process requires a certified migration consultant ot be on-site to do/assist the migraiton. Those consultants are trained specially for certain databases and know tips and tricks how to improve the migration time.
    See
    http://service.sap.com/osdbmigration
    --> FAQ
    For MaxDB there's a special service available, see
    Note 715701 - Migration to SAP DB/MaxDB
    Markus

  • Database full compile while doing schema import using impdp

    hi,
    oracle 10g
    Database full compile while doing schema import using impdp..
    what is happening here..
    regards,
    Deepak

    My scenario
    I need to import the particular schema from the full export dump. which has taken by using expdp. while importing i am using remap_schema for a single schema.
    But it try to import the full dump and compile all the schema objects.
    regards,
    Deepak

  • Script request for  exporting, importing, inserting entry to Oracle 9i LDAP

    Do anyone has the scripts for exporting, importing, inserting entry to Oracle 9i LDAP server? Thanks for help.

    you can use ldapsearch utility to generate ldif files. Perform a search on the node you want to export and specify the output to be in LDIF format ( -L - this is a ldapsearch option). Once you have the ldif file, you can import it to any LDAPv3 complaint server such as OID. for OID, use ldapadd/ldapmodify utility to import the data.
    These utilities are present under IAS_HOME/bin or ORACLE_HOME/bin.

  • Low Hit Ratio for Export/Import Buffer

    Hi,
    Do low hit ratio (less than 70%) for Export/Import buffer causes performance issue?
    I didnt get any Note related to this.
    Any help will be appreciated.
    Thanks,
    Saurabh

    Thanks everyone,
    Basicaly all the buffers(Program/Nametab/CUA/Screen/Genric Key/Single record) has hit ratio around 99% , but for export/import buffer its 60%. which is indeed less.
    This is what i intended to ask, if this low hit ratio for export/import buffer has much impact ?
    Any note if anyone remembers?
    Thanks,
    Saurabh

  • Export import - using TOAD FOR ORACLE  and ORACLE DATABASE 10G EXPRESS or s

    Hi all,
    Could you please kindly help me?
    I am using TOAD FOR ORACLE to export a table to flat file A. The tool just supports to distinguish fields by "spaces".
    And the web page of ORACLE DATABASE 10G EXPRESS to import the data from flat file A to another database. To load data to a table from a text file, the web page bases on "comma" to distinguish fields.
    So could you have any suggestion for me? I need to export data via TOAD FOR ORACLE. And then import it to another database by the home page of ORACLE DATABASE 10G EXPRESS or sqlplus.
    Thank you so much for your help!

    Dont use TOAD for exporting your data. Use PL/SQL. Below is the code given in Asktom.com that does what you want.
    create or replace function  dump_csv( p_query     in varchar2,
                                          p_separator in varchar2 default ',',
                                          p_dir       in varchar2 ,
                                          p_filename  in varchar2 )
    return number
    is
        l_output        utl_file.file_type;
        l_theCursor     integer default dbms_sql.open_cursor;
        l_columnValue   varchar2(2000);
        l_status        integer;
        l_colCnt        number default 0;
        l_separator     varchar2(10) default '';
        l_cnt           number default 0;
    begin
        l_output := utl_file.fopen( p_dir, p_filename, 'w' );
        dbms_sql.parse(  l_theCursor,  p_query, dbms_sql.native );
        for i in 1 .. 255 loop
            begin
                dbms_sql.define_column( l_theCursor, i, l_columnValue, 2000 );
                l_colCnt := i;
            exception
                when others then
                    if ( sqlcode = -1007 ) then exit;
                    else
                        raise;
                    end if;
            end;
        end loop;
        dbms_sql.define_column( l_theCursor, 1, l_columnValue, 2000 );
        l_status := dbms_sql.execute(l_theCursor);
        loop
            exit when ( dbms_sql.fetch_rows(l_theCursor) <= 0 );
            l_separator := '';
            for i in 1 .. l_colCnt loop
                dbms_sql.column_value( l_theCursor, i, l_columnValue );
                utl_file.put( l_output, l_separator || l_columnValue );
                l_separator := p_separator;
            end loop;
            utl_file.new_line( l_output );
            l_cnt := l_cnt+1;
        end loop;
        dbms_sql.close_cursor(l_theCursor);
        utl_file.fclose( l_output );
        return l_cnt;
    end dump_csv;
    /Here is the link to this thread in asktom.
    http://asktom.oracle.com/pls/asktom/f?p=100:11:0::::P11_QUESTION_ID:95212348059

  • Issues with Export/Import using Database & Shared buffer

    Hi All,
    I have a method that calls a program via a job and I am having issues passing data between the two.
    Note that two different users process the method and program (via job) resp. This is how I am calling the second prog-
    SUBMIT ZPROGRAM
                 VIA JOB     l_jobname
                     NUMBER  l_jobcount
                     USER    i_user
                 AND RETURN.
    I need to pass data from method to  the second program and vice versa and then the method continues its processing with the data acquired from the second program.
    I have tried using Import/Export using database and also shared buffer. What I have found is that most of the times I am able to pass data from method to program. However the job takes a couple of min to execute and I think that the data is not making back to the method in time.
    I have looked at some useful forum links-
    Problem with export/import in back ground
    Re: EXPORT/IMPORT  to MEMORY
    but havent been able to find an answer yet. Any solution? Thanks in advance for your help!
    Liz

    Hi Suhas, Subhankar
    I have tested the scenario without the job previously itself and it works. Thats the reason, i am trying with the job now as my requirement is that I need to change the user while executing the second report.
    Here is an example of my import/export - I am passing the return value from the second report to the first.
    Code in second report-
    DATA: INDXKEY LIKE INDX-SRTFD VALUE 'RET1'.
    INDX-AEDAT = SY-DATUM.
    INDX-USERA = SY-UNAME.
    EXPORT RETURN1 TO SHARED BUFFER INDX(ST) ID INDXKEY.
    Code in first report -
    SUBMIT ZPROGRAM
                     VIA JOB     l_jobname
                         NUMBER  l_jobcount
                         USER    i_user
                     AND RETURN.
    Once Job close FM is executed successfully, I import the values as follows
    IMPORT RETURN1 TO RETURN1 FROM SHARED BUFFER INDX(ST) ID INDXKEY3.
    INDXKEY is having value RET1.
    However Return1 is not having any values in first report. It has some value in executed without the job
    Please note that I have tried Export/import with Database too and I am getting the same results.
    Thanks for your suggestions.
    Regards, Liz

  • Database Migration 10g, difference in export/import from 8.1.6 and DBUA

    i would like to know if there is any difference in migrating using export/import methods from 8.1.6 and migration using dbua from 8.1.6 -> 8.1.7 -> 10g. In terms of tablespaces, datafiles performance .
    Will the DBUA convert the tablespaces automatically to take advantage of 10g ?
    Thanks .

    Hi,
    If we refer to the Oracle doc, it's depend of your 10g release.<br>
    Into Upgrade Paths 10.2, you need to upgrade your 8i firstly in 8.1.7.4.<br>
    Into Upgrade Paths 10.1, you you can directly migrate.<br>
    Anyway, you need to use 8iexp utility, and 10g imp utility : Using Different Releases of Export and Import 10.1 or Using Different Releases of Export and Import 10.2.<br>
    <br>
    Nicolas.

  • Best choice for exporting / importing EUL

    Hi all
    I have been tasked with migrating an EUL from a R11 to a R12 environment. The Discoverer version on both environments is 10.1.2 and the OS is Solaris on oracle db's.
    I am unfortunately not experienced with Discoverer and there seems to be no one available to assist for various reasons. So I have been reading the manual and forum posts and viewing metalink articles.
    I tried exporting the entire eul via the wizard and then importing it to the new environment but i was not successfull and experienced the system hanging for many hours with a white screen and the log file just ended.
    I assumed this was a memory problem or slow network issues causing this delay. Someone suggested I export import the EUL in pieces and this seemed to be effective but I got missing item warnings when trying to open reports. This piece meal approach also worried me regarding consistency.
    So I decided to try do the full import on the server to try negate the first problem I experienced. Due to the clients security policies I am not able to open the source eul and send it to our dev. I was able to get it from their dev 11 system but I dismissed this as the dev reports were not working and the only reliable eul is the Prod one. I managed to get a prod eex file from a client resource but the upload to my server was extremely slow.
    I asked the dba to assit with the third option of exporting a db dump of the eul_us and importing this into my r12 dev environment. I managed this but had to export the db file using sys which alleviated a priviledge problem when logging in, I have reports that run and my user can see the reports but there are reports that were not shared to sysadmin in the source enviroment are now prefixed with the version 11 user_id in my desktop and the user cannot see her reports only the sysadmin ones.
    I refreshed the BA's using a shell script I made up which uses the java cmd with parameters.
    After some re reading I tried selecting all the options in the validate menu and refreshing in the discover admin tool.
    If I validate and refresh the BA using the admin tool I get the hanging screen and a lot of warnings that items are missing( so much for my java cmd refresh!) and now the report will not open and I see the substitute missing item dialogue boxes.
    My question to the forum is which would be the best approach to migrate the entire eul from a R11 instance to a R12 instance in these circumstances?
    Many thanks
    Regards
    Nick

    Hi Srini
    The os and db details are as follows:
    Source:
    eBus 11.5.2
    Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bi
    SunOS 5.10 Generic_142900-11 sun4u sparc SUNW,Sun-Fire-V890
    Target:
    ebus 12.1.2
    Oracle Database 11g Enterprise Edition Release 11.1.0.7.0 - 64bit Production DEV12
    SunOS 5.10 Generic_142900-11 sun4u sparc SUNW,Sun-Fire-V890
    Yes the DBA initially did an exp for me using EUL_US as the owner but a strange thing happened with privileges and also, some of the imported tables appeared in the target environment under the apps schema(21 tables) even though the eul_us exp had been 48 tables.
    I also had a problem on the db with "eul_us has insufficient privileges on table space discoverer" type errors.
    I checked the eul_us db privileges and was unable to resolve this initial privilege error even though the privileges were granted to eul_us.
    The dba managed to exp as system and then import it with the full=y flag in the import command which seems to bring in the privileges.
    Then I ran the eul5_id.sql and then made up a list of the business areas and made a sh script to refresh the business areas as follows:
    java -jar eulbuilder.jar -connect sysadmin/oracle1@dev -apps_user -apps_responsibility "System Administrator" -refresh_business_area "ABM Activities" -log refresh.log
    This runs successfully and I can log in select business area and grant access to the users. The reports return data.
    Then one of the users said she can't see all her reports. I noticed some if I opened desktop that were sitting there prefixed with a hash and her version 11 user id.
    So back to the manuals and in the disco admin help the instructions are to first go to view > validate > select all options then go to the business area and click file refresh. This gives me a lot of warnings about items that are missing. I assume this is because the item identifiers brought across in the db dump are the version 11 ones and thus not found in the new system.
    Any suggestions?
    Many thanks
    Nick

  • Suggestions needed for Export/Import Business Area Problem

    I'm trying to export/import a Business Area from one instance to another and I get the following error during import processes.
    Database Error - ORA-00001: unique constraint (EUL4_US.EUL4_EXP_PK) violated
    The Business Area I'm trying to import doesn't exists in new instance. During the import wizard I've checked refresh existing object and match by identifiers.
    I think my target instance has objects with same identifiers, so I went ahead and changed the identifiers in my source, but still problem persists. How can I correct this error?
    Thanks

    Did you happen to copy the EUL from one database to another before you were importing/exporting the BA?
    Discoverer recognizes EULs using unique reference numbers. However, if you use the database export and import utilities to copy an EUL, the new EUL (including its reference number) will be identical to the original EUL. When EULs have the same reference number, EUL consistency issues can arise if you do both of the following:
    * If you modify objects in both the original EUL and in the new EUL
    * Having modified objects in both EULs, if you then attempt to copy objects between the two EULs using the Discoverer Export Wizard and Import Wizard (or the Discoverer /export and /import commands)
    To avoid potential EUL consistency issues, run the eul5_id.sql script as the owner of the new EUL. The eul5_id.sql script gives a new refererence number to the new EUL and thereby avoids any potential EUL consistency issues.
    Reference
    * How to import an EUL using the standard database import utility
    http://download-west.oracle.com/docs/html/B13916_04/maintain_eul.htm#i1008632
    For 4i, you will ned to run:
    sql> update EUL4_VERSIONS set VER_EUL_TIMESTAMP =TO_CHAR(SYSDATE, 'YYYYMMDDHH24MISS');
    sql> commit;
    Some similar EUL inconsistencies are documented in:
         Note.267476.1     Discoverer Workbook Fails With 'Cannot join tables.Item dependency "" not found in EUL' After EUL Migration, Importing Or Cloning
    You might want to check that if the EUL version timestamps are the same.
    ~Steve.

  • Custom directories for export/import for Integration Builder

    Hi,
    XI uses default directories for exporting and importing Integration Builder objects, and there is no option to change the directory or to pick up objects from another directory.
    Does anybody know of any way we can change the default directory that XI uses? Maybe some system setting hidden somewhere deep?
    Thanks and Regards
    Manish

    Hi Manish,
    when there's no way there's always an <b>XI way</b>:)
    you can help your basis team using XI
    set you file adapter to monitor export directory
    and try sending your xi export files to a different server/location
    you can have a look how Felix did similar scenario: 
    /people/sap.user72/blog/2005/07/15/copy-a-file-with-same-filename-using-xi
    isn't this a good way ?
    Regards,
    michal

  • Plug-In for exporting/importing links

    I need a plug-in that can export/import all links from a pdf.
    Until now I found only one http://www.evermap.com/autobookmark.asp
    But it's kinda pricy... does anybody know of anything else?

    I use Jeffrey Friedl's “jf Zenfolio” plug-in and highly recommend it. Unlike most plug-ins of this sort, it get syncing right. Most will allow you to upload once, and then if you change things in Lr, the syncing goes to heck. Not JfZenfolio. It's a pleasure to use.
    Jeffrey Friedl's Blog » Jeffrey’s “Export to Zenfolio” Lightroom Plugin

  • Same user for export import utility

    oracle database 10g
    OS windows 2000 server
    for import Is it necessary to create user with same name as of export file?

    No need to create a user. but while importing you have to import as system with from user and touser clause specifying from where we export and where we are going to import
    eg:
    imp system/manager file=xyz.dmp fromuser=scott touser=matt

  • Automated Solution for Exporting SQL used in Crystal Reports

    My company has created over 400 reports that are running on Crystal Reports XI. Due to proposed changes in our database, we occasionally need to review the SQL statements of all 400+ reports to determine if any of them would be impacted by the proposed changes. In order to review the SQL used in these 400+ reports, we have had to open each report manually. We would like to know how we can review the SQL used in the reports without having to manually look at each report. I suspect there is a way to either write a program or use a tool that will export the SQL that is used in all reports. In a similar vain, we would be interested in knowing if there is an automated way to identify table names and fields that are used in each SQL statement.

    I see two directions to go:
    1. Review functionalities of .rpt Inspector
    [http://www.softwareforces.com/Product/ri/pro/3/rptInspector.htm]
    - it allows rpt mass analyses
    2. Post your question to ".Net Development - Crystal Reports" forum to see if it is possible to access required information through custom application

Maybe you are looking for