Dbms_datapump/ import job/ mode=table /11g

Hi
I'm working with dbms_datapump API.
Situation:
I need to import a dump in a schema, but only 4/2000 tables must have the data imported. I did not find a way to make this in only with one import command, so i make it by 2 consecutive imports.
First import with the data_filter INCLUDE_ROWS at 0: it works well with 10g and 11g
The second import tries to only import the content of the 4 tables: it works well on 10g, but fails on 11g.
Code:
DECLARE
lv_jobHandle NUMBER;
BEGIN
lv_jobHandle := dbms_datapump.open(operation => 'IMPORT'
,job_mode => 'TABLE'
,remote_link => NULL
,job_name => 'TESTIMP_EMPTY');
dbms_datapump.add_file(handle => lv_jobHandle
,filename => 'mydump_%U.dp'
,directory => 'MYDIR_DP');
dbms_datapump.add_file(handle => lv_jobHandle
,filename => 'TESTIMP_EMPTY_'||to_char(SYSTIMESTAMP,'YYYYMMDDHH24MISS')||'.log'
,directory => 'MYDIR_LOG'
,filetype => dbms_datapump.ku$_file_type_log_file);
dbms_datapump.set_parameter(handle => lv_jobHandle
,name => 'TABLE_EXISTS_ACTION'
,value => 'APPEND');
dbms_datapump.metadata_filter(handle => lv_jobHandle
*,name => 'NAME_LIST'*
*,value => '''TMP_EVENT_TRIGGERS'',''SPLDBBUILDHIERARCHY'',''SPLDBBUILDLOG'',''SPLMIGRATIONLOG''');*
dbms_datapump.metadata_remap(handle => lv_jobHandle
,name => 'REMAP_SCHEMA'
,old_value => 'SEB_DEV800SP3'
,value => 'TSTIMPORTEMPTY');
dbms_datapump.metadata_remap(handle => lv_jobHandle
,name => 'REMAP_TABLESPACE'
,old_value => 'SRC_DATA'
,value => 'DEST_DATA'
,object_type => NULL);
dbms_datapump.start_job(lv_jobHandle);
dbms_datapump.detach(handle => lv_jobHandle);
END;
*11g Error:*
[oracle@qcdvcn1004 dbadev]$ cat TESTIMP_EMPTY_20110427153505.log
Master table "DBADEV"."TESTIMP_EMPTY" successfully loaded/unloaded
ORA-31655: no data or metadata objects selected for job
Starting "DBADEV"."TESTIMP_EMPTY":
Job "DBADEV"."TESTIMP_EMPTY" successfully completed at 15:35:11
Someone has an idea ?
Thanks!
jonathan

Hello,
I need to import a dump in a schema, but only 4/2000 tables must have the data imported. I did not find a way to make this in only with one import command, so i make it by 2 consecutive imports.There's a nice trick so as to do that in just *1 Export/Import* by using the QUERY parameter.
Please scroll down on this link and read the solution given by Dean Gagne:
Export with all objects and only some table data
But, this is true, it's not expressed by using API.
Hope this help.
Best regards,
Jean-Valentin

Similar Messages

  • Datapump network import from 10g to 11g database is not progressing

    We have an 11.2.0.3 shell database in RedHat Linux 5 64-bit platform, and we are pulling data from its 10g database (10.2.0.4) HP-UX platform using datapump via NETWORK_LINK. However, the import seems to be not progressing and completing. We have left it run for almost 13 hours and did not even import a single row to the 11g database. We even try to import only one table with 0 rows but still it is not completing, the logfile just continue to be lopping on the following:
    Worker 1 Status:
    Process Name: DW00
    State: EXECUTING
    Estimate in progress using BLOCKS method...
    Job: IMP_NONLONG4_5
    Operation: IMPORT
    Mode: TABLE
    State: EXECUTING
    Bytes Processed: 0
    Current Parallelism: 1
    Job Error Count: 0
    Worker 1 Status:
    Process Name: DW00
    State: EXECUTING
    We also see this:
    EVENT SECONDS_IN_WAIT STATUS STATE ACTION
    SQL*Net message from dblink 4408 ACTIVE WAITING IMP_NONLONG4_5
    Below is our par file:
    NETWORK_LINK=DATABASE_10G
    DIRECTORY=MOS_UPGRADE_DUMPLOC1
    LOGFILE=imp_nonlong_grp-4_5.log
    STATUS=300
    CONTENT=DATA_ONLY
    JOB_NAME=IMP_NONLONG4_5
    TABLES=SYSADM.TEST_TBL
    Any ideas? Thanks.

    Thanks a lot for all who have looked and responded to this, I appreciate you giving time for suggestions. As a recap, datapump export and import via dumpfile works on both 10g and 11g databases, we are only having issue pulling data from the 10g database by executing datapump network import (using dblink) from the 11g database.
    SOLUTION: The culprit was the parameter optimizer_features_enable='8.1.7' that is currently set on the 10g database, we have taken that out of the parameter file of the 10g database and network import worked flawlessly.
    HOW DID WE FIGURE OUT: We turned on a trace to the datapump sessions and trace file shows something about optimizer (see below). So we have taken one by one any parameter that has something to do with optimizer and found out it was optimizer_features_enable='8.1.7' parameter that when removed makes the network import successful.
    SELECT /* OPT_DYN_SAMP */ /*+ ALL_ROWS opt_param('parallel_execution_enabled',
    'false') NO_PARALLEL(SAMPLESUB) NO_PARALLEL_INDEX(SAMPLESUB) NO_SQL_TUNE
    */ NVL(SUM(C1),0), NVL(SUM(C2),0), NVL(SUM(C3),0)
    FROM
    (SELECT /*+ NO_PARALLEL("SYS_IMPORT_TABLE_02") INDEX("SYS_IMPORT_TABLE_02"
    SYS_C00638241) NO_PARALLEL_INDEX("SYS_IMPORT_TABLE_02") */ 1 AS C1, 1 AS C2,
    1 AS C3 FROM "DBSCHEDUSER"."SYS_IMPORT_TABLE_02" "SYS_IMPORT_TABLE_02"
    WHERE "SYS_IMPORT_TABLE_02"."PROCESS_ORDER"=:B1 AND ROWNUM <= 2500)
    SAMPLESUB
    SELECT /* OPT_DYN_SAMP */ /*+ ALL_ROWS IGNORE_WHERE_CLAUSE
    NO_PARALLEL(SAMPLESUB) opt_param('parallel_execution_enabled', 'false')
    NO_PARALLEL_INDEX(SAMPLESUB) NO_SQL_TUNE */ NVL(SUM(C1),0), NVL(SUM(C2),0)
    FROM
    (SELECT /*+ IGNORE_WHERE_CLAUSE NO_PARALLEL("SYS_IMPORT_TABLE_02")
    FULL("SYS_IMPORT_TABLE_02") NO_PARALLEL_INDEX("SYS_IMPORT_TABLE_02") */ 1
    AS C1, CASE WHEN "SYS_IMPORT_TABLE_02"."PROCESS_ORDER"=:B1 THEN 1 ELSE 0
    END AS C2 FROM "DBSCHEDUSER"."SYS_IMPORT_TABLE_02" "SYS_IMPORT_TABLE_02")
    SAMPLESUB

  • BODS 4.2 Cannot import the metadata table, RFC_ABAP_INSTALL_AND_RUN syntax error

    Hi all, we installed BODS 4.2 server to substitute a 4.1, but we are facing the error:
    Error: Cannot import the metadata table <name=T001>
    RFC CallReceive error <Function /BODS/RFC_ABAP_INSTALL_AND_RUN: RFC_ABAP_RUNTIME_FAILURE -(Exception Key: Syntax error in program /BODS/SAPLBODS....
    We already tried the solution for when people get the error related to unicode.
    Also, we are able to pull data via extractors, it only fails when loading Tables....
    Any help is greatly appreciated.

    Dear,
    You will have to import the new ABAP Function Group "BODS/BODS".
    Here are some details:
    Installing Functions on the SAP Server
    SAP BusinessObjects Data Services provides functions that support the use of the ABAP, BAPI, and
    IDoc interfaces on SAP servers. You will need some or all of these functions in the normal operation
    of the software in an SAP environment. These functions perform such operations as dynamically loading
    and executing ABAP programs from Data Services, efficiently running preloaded ABAP programs,
    allowing for seamless metadata browsing and importing from SAP servers, and reporting the status of
    running jobs. Some of these functions read data from SAP NetWeaver BW sources.
    You must upload the provided functions to your SAP server in a production environment. It is
    recommended that you always upload the functions to your SAP server whether you are in a
    development, test, or production environment. The functions provide seamless integration between
    Data Services and SAP servers.
    The default installation places two function module files for SAP servers in the ...\Data
    Services\Admin\R3_Functions\transport directory. You then upload these files to SAP servers
    using the SAP Correction and Transport System (CTS) or manually. Using CTS allows for version
    control as the functions evolve across releases.
    The installation provides two versions of transport files (depending on the server version you are using)
    to install the functions on the SAP server. To obtain the names of the latest transport files for installing
    or upgrading these SAP server functions, see the readme.txt file
    And I've found those files and text files in the local install folder....in:
    Program Files\SAP BusinessObjects\Data Services\admin\R3_Functions
    (that's where I've installed it).
    There you'll find some descriptive txt as how to proceed.
    After installing, it might happen that the executing user is missing some authorizations.
    Here my authorizations team helped me by tracing the user and then adding the necessary rights.
    Sure hope this will help you.

  • Import / Export Specific tables from DB

    Hi,
    I am trying to export all 'AD_' tables from my database from another database with their constraints.
    I tried a command just for export one table from my cmd :
    "exp adempiere/adempiere@simext15:1521/compiere2 file=emp.dmp tables=(AD_ACCESSLOG)"
    it export fine but when try to import this file from using command
    "imp adempiere/adempiere@simext15:1521/compiere2 file=emp.dmp fromuser=adempiere touser=adempiere tables=(AD_ACCESSLOG)"
    it gives me so many errors:
    Connected to: Oracle Database 11g Enterprise Edition Release 11.1.0.6.0 - Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    Export file created by EXPORT:V11.01.00 via conventional path
    import done in WE8MSWIN1252 character set and UTF8 NCHAR character set
    . importing ADEMPIERE's objects into ADEMPIERE
    . . importing table "AD_ACCESSLOG" 0 rows imported
    IMP-00015: following statement failed because the object already exists:
    "CREATE UNIQUE INDEX "AD_ACCESSLOG_KEY" ON "AD_ACCESSLOG" ("AD_ACCESSLOG_ID""
    " ) PCTFREE 10 INITRANS 2 MAXTRANS 255 STORAGE(INITIAL 65536 FREELISTS 1 FR"
    "EELIST GROUPS 1 BUFFER_POOL DEFAULT) LOGGING"
    IMP-00017: following statement failed with ORACLE error 2264:
    "ALTER TABLE "AD_ACCESSLOG" ADD CONSTRAINT "AD_ACCESSLOG_KEY" PRIMARY KEY ("
    ""AD_ACCESSLOG_ID") USING INDEX PCTFREE 10 INITRANS 2 MAXTRANS 255 STORAGE(I"
    "NITIAL 65536 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT) TABLESPACE "
    ""USERS" LOGGING ENABLE "
    IMP-00003: ORACLE error 2264 encountered
    ORA-02264: name already used by an existing constraint
    IMP-00017: following statement failed with ORACLE error 2264:
    "ALTER TABLE "AD_ACCESSLOG" ADD CONSTRAINT "ADTABLE_ADACCESLOG" FOREIGN KEY "
    "("AD_TABLE_ID") REFERENCES "AD_TABLE" ("AD_TABLE_ID") ENABLE NOVALIDATE"
    IMP-00003: ORACLE error 2264 encountered
    ORA-02264: name already used by an existing constraint
    IMP-00017: following statement failed with ORACLE error 2264:
    "ALTER TABLE "AD_ACCESSLOG" ADD CONSTRAINT "ADCOLUMN_ADACCESSLOG" FOREIGN KE"
    "Y ("AD_COLUMN_ID") REFERENCES "AD_COLUMN" ("AD_COLUMN_ID") ON DELETE CASCAD"
    "E ENABLE NOVALIDATE"
    IMP-00003: ORACLE error 2264 encountered
    ORA-02264: name already used by an existing constraint
    About to enable constraints...
    IMP-00017: following statement failed with ORACLE error 2430:
    "ALTER TABLE "AD_ACCESSLOG" ENABLE CONSTRAINT "ADTABLE_ADACCESLOG""
    IMP-00017: following statement failed with ORACLE error 2430:
    "ALTER TABLE "AD_ACCESSLOG" ENABLE CONSTRAINT "ADCOLUMN_ADACCESSLOG""
    Import terminated successfully with warnings.
    I tried a lot for this I just want to export and restore all 338 tables related to Application Dictionary from one database to another database with their constraints and data.
    I tried also using datalink but it only copy the structure and data but not the constraints, I want all three should be copy in to another database.
    I know their is dependency on a table to another table but i don't know exactly it..
    tried it onces again i am able to do export and import tables but only primary key related to every table gets restore neither Unique key nor Foreign key , please help me out from this i can export and import with full table with data and its all constraints.
    Please help me out from this problem, waiting for your valuable suggestions.
    Regards,
    Manu

    ORA-02264: name already used by an existing constraint
    The constraint is already exists with the same name. So drop that before issuing the import command.
    And ,
    1. Check if all tables get imported
    2. If all tables get imported then create the constraints manually
    3. If tables do not get imported then import tables manually by specifying their name in imp statement
    Thanks

  • Oem: import job error

    i get the following message when i import any object in database.
    couldn't open "d:\import.log" : no such file or directory
    please help
    thanks

    Are you running the DataGatherer service?
    <BLOCKQUOTE><font size="1" face="Verdana, Arial">quote:</font><HR>Originally posted by rafael:
    if i try to export/import any object of a remote database the job works fine, but if then export/import job is on local database i get the same error. I try to change all parameters and credential, always get the same error.
    if i try to import an object (a table, for example) i get this error.
    "@EXP_USEROLE@"."D"
    "@EXP_USERNAME@"."SYSTEM"
    "@EXP_MODE@"."R"
    "@EXP_TYPE@"."USERS"
    COULDN'T OPEN "D:\ORANT\ORA81\WORK\547125707.LOG" : NO SUCH FILE OR DIRECTORY
    THANKS FOR ALL.
    <HR></BLOCKQUOTE>
    null

  • Export/import (user mode)

    can you help me please,
    1- I exported the SCOTT schema (user mode):
    exp scott/tiger FILE=D:\scott.dat LOG=D:\scott_exp.log OWNER=SCOTT
    then I imported to new created before schema SCOTT_COPY:
    imp scott_copy/tiger FILE=D:\scott.dat LOG=D:\scott_imp.log
    fromuser=scott
    touser=scott_copy
    I got in scott_copy only tables from scott (all existed types were not successful imported) WHY only tables (see log)!!.
    2- What is the best way to export user and import it with all (objects and privilegs granted to origin user).
    thanks
    jamil
    see scott_imp.log
    =================
    import done in WE8ISO8859P1 character set and WE8ISO8859P1 NCHAR character set
    IMP-00017: following statement failed with ORACLE error 2304:
    "CREATE TYPE "ADDRESS_OBJ" TIMESTAMP '2002-02-08:10:15:23' OID '8AB6D00504B0"
    "47A9BDAC02E58E4298A0' as object("
    "pobox varchar2(20),"
    "street varchar2(20),"
    "city varchar2(20),"
    "state varchar2(20),"
    "country varchar2(20))"
    IMP-00003: ORACLE error 2304 encountered
    ORA-02304: invalid object identifier literal
    IMP-00017: following statement failed with ORACLE error 2304:
    "CREATE TYPE "BREAD_LIST" TIMESTAMP '2002-02-08:10:13:03' OID 'D9896A6992394"
    "4F58D608A3ED01791D6' as array(10) of bread_obj;"
    IMP-00003: ORACLE error 2304 encountered
    ORA-02304: invalid object identifier literal
    IMP-00017: following statement failed with ORACLE error 2304:
    "CREATE TYPE "BREAD_OBJ" TIMESTAMP '2002-02-08:10:10:55' OID '24F457B9807C47"
    "B9B753CEC5E88D584A' as object("
    "bread_id number(2),"
    "bread_name varchar2(20),"
    "sale_price number(5,2),"
    "in_stock number(4));"
    IMP-00003: ORACLE error 2304 encountered
    ORA-02304: invalid object identifier literal
    IMP-00017: following statement failed with ORACLE error 2304:
    "CREATE TYPE "INGREDIENT_OBJ" TIMESTAMP '2002-02-08:10:11:39' OID '8A39AD416"
    "8DF411995F0B2B2020BF731' as object("
    "ing_id number(2),"
    "name varchar(20),"
    "price number(5,2),"
    "in_store number (8,3))"
    IMP-00003: ORACLE error 2304 encountered
    ORA-02304: invalid object identifier literal
    . . importing table "A" 13 rows imported
    . . importing table "ACCOUNT" 5 rows imported
    . . importing table "B" 5 rows imported
    . . importing table "BONUS" 0 rows imported
    . . importing table "DEPT" 5 rows imported
    . . importing table "EMP" 13 rows imported
    . . importing table "EMP1" 14 rows imported
    . . importing table "EMP_SNAPSHOT" 14 rows imported
    . . importing table "RECEIPT" 1 rows imported
    . . importing table "SALGRADE" 5 rows imported
    . . importing table "T" 12 rows imported
    . . importing table "TABLE_RESULT" 5 rows imported
    . . importing table "TEMP" 3 rows imported
    . . importing table "XXX" 0 rows imported
    About to enable constraints...
    Import terminated successfully with warnings.

    Hi Basu,
    The source and target database is the same 8.1.7.0.0
    more infot about what do i want?export one user with all his objects,priviliges....... by simple words, I would like to have the same copy of source user with new namw of the user one.
    thanks.

  • The SSP Timer Job Distribution List Import Job was not run.

    The SSP Timer Job Distribution List Import Job was not run.   Reason: Logon failure: the user has not been granted the requested logon type at this computer
    We are facing the above mentioned problem . I have added the service account in backup operates and did everything like update farm credentils , update managed account etc , but still no luck . Please note that we haven't done any chnages to the account
    Application pool is keeping going into stop mode when we try to access the site . Both CA and webapplication are down with the error as Service Unavailble . Please help
    Srini

    sounds like the account can only be used on certain COMPUTERS ("has not been granted the requested logon type
    at this computer")... check active directory.
    Scott Brickey
    MCTS, MCPD, MCITP
    www.sbrickey.com
    Strategic Data Systems - for all your SharePoint needs

  • Importance of Families Table

    Hi experts,
    Im new to MDM and would like to know the importance of families table.
    Where it is used and how, as I find it is shown as disable all the times.
    Regards
    Hari

    Hi Hari,
    Families can be used to group together records, which have the same value or the same attribute in certain fields.
    The families table contains one record for each family. So attributes valid for all family members can be assigned directly to the family and need not be maintained individually for each family member.
    This means that instead of having to store group-specific information for every single record, the main table can be more manageable. This group-specific information (for example, for product categories) is stored in the families table and therefore applies to every family member.
    The families table must also contain the lookup field in the main
    table, which is used to define family membership. This stored lookup field must be of type taxonomy or hierarchy. This allows the family records to be generated automatically.
    For eg: if you have a taxonomy for suppliers with various attributes
    supplier
    |_Tools
    |_Hand Tools
    |_Machine Tools
    |_Services
    |_Repair
    Now you want to further categorize this hierarchy based on say Manufacturer or Catalog Name.This can be done using Families table.
    In console create a families table as type families using Category as family field.
    Add a field say Image (look up Image).
    After importing the taxonomy table category.
    Go to Data manager select Family mode.Select a node .Under partition tab select one of the look up values or Attributes based on which u want to categorize further.Then go to Family detail here u can link a image or PDF to this family.
    These are used in case of publishing and printing.
    Hope it will help you.
    Regards
    Richa

  • Error while importing a DW table into DAC

    Hi,
    We are on OBIEE 7.9.6 and we have a requirement of adding new DW table. I created a new table in DW and getting error "Could connect to data source process. Process failed during creation of connection pool", while trying to import this new table into DAC. I would like to know why I am getting this error.
    Any help is greatly appreciated.
    Thanks,
    Chandra

    Thank you for your response.
    I added this table in DW database and have imported into informatica. When I look at the traffic light it is red. Looks like thats the problem.
    Thanks again.

  • I'm trying to open a 900kb Word doc (240pages) in Pages but get this error message:  Import Warning - A table was too big. Only the first 40 columns and 1,000 rows were imported.

    I'm trying to open a 900kb Word doc (240pages) in Pages but get this error message:  Import Warning - A table was too big. Only the first 40 columns and 1,000 rows were imported.

    Julian,
    Pages simply won't support a table with that many rows. If you need that many, you must use another application.
    Perhaps the originator could use a tabbed list rather than a table. That's the only way you will be able to contain a list that long in Pages. You can do the conversion yourself if you open the Word document in LibreOffice, Copy the Table, Paste the Table into Numbers, Export the Numbers doc to CSV, and import to Pages. In Pages Find and Replace the Commas with Tabs.
    There are probably other ways, but that's what comes to mind here.
    Jerry

  • Error when Importing a External table

    When I am trying to import a external table I am getting a error as SQL Exception. Any ideas as to how to fix this error.
    The external table is under one schema. I have a locator in OWB pointing to that schema and I am trying to Import the external table. I am able to see this table in TOAD.
    Thanks

    Repository Error:SQL Exception..
    Class Name: CacheMediator.
    Method Name: getDDEntryFromDB.
    Repository Error Message: ORA-22992: cannot use LOB locators selected from remote tables
         at oracle.wh.repos.pdl.foundation.CacheMediator.getSQLResult(CacheMediator.java:504)
         at oracle.wh.repos.impl.CMPWBAPIFactory.getSQLResult(CMPWBAPIFactory.java:2137)
         at oracle.wh.service.sdk.integrator.RepositoryUtils.getSQLResult(RepositoryUtils.java:448)
         at oracle.wh.service.sdk.integrator.RepositoryUtils.getSQLResult(RepositoryUtils.java:413)
         at oracle.wh.ui.integrator.oracle.WBOracleDatabaseEntityAccessor.getAccessParametersSQLResult(WBOracleDatabaseEntityAccessor.java:518)
         at oracle.wh.ui.integrator.oracle.WBOracleDatabaseEntityAccessor.getAccessParameters(WBOracleDatabaseEntityAccessor.java:459)
         at oracle.wh.ui.integrator.common.ImportEntityAlgorithm.importAccessParameters(ImportEntityAlgorithm.java:2724)
         at oracle.wh.ui.integrator.common.ImportEntityAlgorithm.importExternalTable(ImportEntityAlgorithm.java:1499)
         at oracle.wh.ui.integrator.common.ImportEntityAlgorithm.dispatchElement(ImportEntityAlgorithm.java:583)
         at oracle.wh.ui.integrator.common.ImportEntityAlgorithm.importElement(ImportEntityAlgorithm.java:366)
         at oracle.wh.ui.integrator.sdk.EntityAccessor.importElement(EntityAccessor.java:80)
         at oracle.wh.ui.integrator.common.ImportService.importElement(ImportService.java:1046)
         at oracle.wh.ui.integrator.common.wizards.ImportElementTransaction.run(ImportWizardDefinition.java:578)

  • How to do import of only table data without procedures and synonyms

    Hello All,
    I want to import only the tables of one database to a new database. I did a table level export of one schema. But along with it the stored procedures,synonyms, views etc are exported. Is there any way to stop them. Or is it possible to import only the tables from the dump created?
    Regards
    Satish

    Hi Maran,
    Thanks for the feedback. I think I didnot specify the parafile with the table list while exporting. I am redoing it now. I shall update this post within 20 minutes.
    Regards
    Satish

  • Error ORA-02375 while trying to export/import JTF.JTF_PF_REPOSITORY table

    We have already created an SR, In the mean time, trying to see whether anyone else has come across this issue.  Thanks.
    On : 11.2.0.3 version, Data Pump Import
    Error ORA-02375 while trying to import JTF.JTF_PF_REPOSITORY table
    We are getting the below error while performing the full db
    import.
    ORA-02375: conversion error loading table
    "JTF"."JTF_PF_REPOSITORY" partition "EBIZ"
    ORA-22337: the type of accessed
    object has been evolved
    ORA-02372: data for row: SYS_NC00040$ :
    0X'8801FE000004AD0313FFFF0009198401190A434F4E4E454354'
    This issue is
    stopping our upgrade of database from 10.2.0.4 to 11.2.0.3. This is very
    critical for us to be resolved.

    Hi,
    seems this is Character set issue fo source and Target DB check this doc:Unable to Export Table WF_ITEM_ATTRIBUTE_VALUES due to errors ORA-02374, ORA-22337, and ORA-02372 (Doc ID 1522761.1)
    HTH

  • Import from internal table

    Dear Experts,
    I need a solution for importing from an internal table.
    Two variables I_EXP01 and I_EXP02 containing an xml stream have been exported to the database using the following syntax
    and depending on the size of I_EXP01 and I_EXP02 two or more entries are created in table LTEX.
    EXPORT
             L_SAPRL
             I_EXP01
             I_EXP02
    TO    DATABASE LTEX(LT)
                           ID LS_LTEXKEY.
    But now i have a few entries of table LTEX in an internal table and they need to be imported from the internal table.
    What would be the syntax to import from internal table so that after importing I can de-serialise the data from XML to ABAP.
    Thanks & Regards,
    Ashwini

    Try this and let me know if it works:
    IMPORT
    L_SAPRL
    I_EXP01
    I_EXP02
    To
    L_SAPRL1 (same type as above)
    I_EXP011 (same type as above)
    I_EXP021 (same type as above)
    FROM DATABASE LTEX(LT)
    ID LS_LTEXKEY.
    if it is not working. try using seperate statements for each internal table and see.
    Thanks,
    Venkatesh.
    Edited by: venkatesh333 on Jul 27, 2011 6:23 AM

  • Writing a stored procedure to import SQL Server table data into a Oracle table

    Hello,
    As a new DBA I have been tasked with writing a stored procedure to import SQL Server table data into an Oracle table. I have been given many suggestions on how to do it from SQL Server but I I just need to write a stored procedure to run it from the Oracle side. Suggestions/guidance on where to start would be greatly appreciated! Thank you!
    I started to write it based on what I have but I know this is not correct :/
    # Here is the select statement for the data source in SQL Server...
    SELECT COMPANY
    ,CUSTOMER
    ,TRANS_TYPE
    ,INVOICE
    ,TRANS_DATE
    ,STATUS
    ,TRAN_AMT
    ,CREDIT_AMT
    ,APPLD_AMT
    ,ADJ_AMT
    ,TRANS_USER1
    ,PROCESS_LEVEL
    ,DESCRIPTION
    ,DUE_DATE
    ,OUR_DATE
    ,OUR_TIME
    ,PROCESS_FLAG
    ,ERROR_DESCRIPTION
      FROM data_source_table_name
    #It loads data into the table in Oracle....   
    Insert into oracle_destination_table_name (
    COMPANY,
    CUSTOMER,
    TRANS_TYPE,
    INVOICE,
    TRANS_DATE,
    STATUS,
    TRANS_AMT,
    CREDIT_AMT,
    APPLD_AMT,
    ADJ_AMT,
    TRANS_USER1,
    PROCESS_LEVEL,
    DESCRIPTION,
    DUE_DATE,
    OUR_DATE,
    OUR_TIME,
    PROCESS_FLAG,
    ERROR_DESCRIPTION)
    END;

    CREATE TABLE statements would have been better as MS-SQL and Oracle don't have the same data types.
    OUR_DATE, OUR_TIME will (most likely) be ONE column in Oracle.
    DATABASE LINK
    Personally, I'd just load the data over a database link:
    insert into oracle_destination_table_name ( <column list> )
    select ... <transform data here>
    from data_source_table@mssql_db_link
    As far as creating the database link from Oracle to MS-SQL ... that is for somebody else to answer.
    (most likely you'll need to use an ODBC driver)
    EXTERNAL TABLE
    If the data from MS-SQL is in a CSV file, just use and external table.
    same concept:
    insert into oracle_destination_table_name ( <column list> )
    select ... <transform data here>
    from data_source_external_table
    MK

Maybe you are looking for

  • How do I install Flash on OS X Jaguar (10.2)?

    I have a iMac G3 (Snow) which can connect to the internet through Ethernet, and has Firefox 1.0.0.8. (I will install Firefox 2) I downloaded Flash Player 9 from Adobe, and mounted the DMG, and it had 2 files inside: Flash Player.plugin, and flashplay

  • Dynamic Filename from source Idoc

    Hi All,   I need to give the filename from the field in the source idoc. the field value need to be picked up as below ZSSV_GENERIC_IDOC -IDOC --Z1SSV_ITEMS_SUMMARY ---HOST_FILENAME Note: - represent the level in the datastructure. here the value of

  • Migrate wls 5.1 war/ejb .jar to 6.0

    Accidentially, I put this posting in an old thread. Let me create a new thread and ask agian. Basically, I want to clarify the directory structure. I have an application with the following directory structure in WLS 5.1 web application: c:\weblogic\m

  • Can you track itouch with the serial number?

    Can I track an Itouch with the Serial Number? My Son Itouch got stolen in Fam and Friends BBQ .

  • Solid alpha shape to vector solid shape

    Please can someone help me this is driving me totally nuts, ive tried all sorts of tracing and everything but cant seem to get just the outline of the shape i want. I have an animated logo ( white), on a background(black) ..( and tried inverting & tr