Prolblems on Loading the data using DataMovement

Hello Sir,
I have created a .dat file and want to load in the table which is created by user.
Now using the Data Movement when i gives the username and password for host credentials this error is popuped.
install_driver(Oracle) failed: Can't load 'D:/app/Metamorph/product/11.1.0/db_1/perl/site/5.8.3/lib/MSWin32-x86-multi-thread/auto/DBD/Oracle/Oracle.dll' for module DBD::Oracle: load_file:The specified module could not be found at D:/app/Metamorph/product/11.1.0/db_1/perl/5.8.3/lib/MSWin32-x86-multi-thread/DynaLoader.pm line 229. at (eval 10) line 3 Compilation failed in require at (eval 10) line 3. Perhaps a required shared library or dll isn't installed where expected at D:/app/Metamorph/product/11.1.0/db_1/perl/site/5.8.3/lib/MSWin32-x86-multi-thread/Oraperl.pm line 58 Compilation failed in require at D:\app\Metamorph\product\11.1.0\db_1\sysman\admin\scripts/emd_common.pl line 37. BEGIN failed--compilation aborted at D:\app\Metamorph\product\11.1.0\db_1\sysman\admin\scripts/emd_common.pl line 37. Compilation failed in require at D:\app\Metamorph\product\11.1.0\db_1/sysman/admin/scripts/db/db_common.pl line 73. Compilation failed in require at - line 1.
how to prceed can i get the guide lines.
Thanks and Regards,
Binney

Hmmm I'm not sure that I know what is "DataMovement" but as I searched metalink and found a doc that says you sould use oracle shipped perl binary instead of any other one (perl from OS). So check that $ORACLE_HOME/perl/bin/perl used by your scripts if you want to use oracle supplied perl modules and tools.
BTW why don't you use SQL*Loader instead of other tools to load data into database...

Similar Messages

  • TSV_TNEW_PAGE_ALLOC_FAILED error while loading the DATA using DTP

    Hi,
    While loading the data using DTP for 2  DSO's we are gettig the error
    TSV_TNEW_PAGE_ALLOC_FAILED
    can any one kindly help me out regarding the same.
    Thank You,
    Poornima.

    Hi Soundarya,
    Thanks a lot for the reply. But i found that its running fine in development, where as coming to quality its throwing an error. These happened for Two DSO's. In both the transformations i have identified that the Transformation names are different from Development and Quality..
    There are no routines written for them and no select statements have been used
    Can you please suggest me regarding the same.
    Edited by: Poornima Gayatri on Mar 22, 2010 7:00 AM

  • How to load the data using a plsql table in ODI.

    Hi All ,
    Can anyone help me on this ?
    We have a PLSQL procedure which returns a plsql table as out parameter.
    We are supposed to load the data in to a file using this plsql table (Table type) in ODI.
    Can this be done using ODI?
    Regards,
    Karthik+

    Hi,
    We have one process with a ref cursor (Oracle) as a source and remote Oracle DB as a target. I ended up writing my own KM that populates a global temporary table from the cursor first and then transfers the data to target. If temporary table is an option for you, the rest is pretty easy.
    Regards.

  • Idoc issue while loading the data into BI

    Hello Gurus,
    Initially i am having the Source System connection problem. After it was fixed by the basis i had followed the below process.
    I am loading the data using the Generic extractor specifying the selection and its a full load. When i load the data from R3 PP application  i found the below result in the monitor screen of BW.
    1. Job was completed in the R3 system and 1 million records are fetched by extractor.
    2. the records are not posted to the BW side Because of TRFC issue.
    3. i got the idoc number and process it in the BD87 tCode. But it was not process sucessfully it gives the following error " Error when saving BOM. Please see the application log."
    4. when i check the Application Log using the Tcode SLG1 with the help of time and date of that particular process it was in the yelow status.
    Kindly let me know i can i resolve this issue. i have already tried by repeating the infopackage job still i am facing the same issue. I have also check the connection its is ok.
    Regards

    hello veerendra,
    Thanks for your quick response. yes i am able to process it manually. after processing  it was ended with the status 51 application not posted.
    could you pls help me out with the same
    regard
    Edited by: KK on Nov 4, 2009 2:19 AM
    Edited by: KK on Nov 4, 2009 2:28 AM

  • Unable to load the data into Cube Using DTP in the quality system

    Hi,
    I am unable to load the data from PSA to Cube using DTP in the quality system for the first time
    I am getting the error like" Data package processing terminated" and "Source TRCS 2LIS_17_NOTIF is not allowed".
    Please suggest .
    Thanks,
    Satyaprasad

    Hi,
    Some Infoobjects are missing while collecting the transport.
    I collected those objects and transported ,now its working fine.
    Many Thanks to all
    Regards,
    Satyaprasad

  • How to write a procedure to load the data into a table using xml file as input to the procedure?

    Hi,
    Iam new to the xml,
    can u please anyone help me how to write procedure to load the data into a table using xml as input parameter to a procedure and xml file is as shown below which is input to me.
    <?xml version="1.0"?>
    <DiseaseCodes>
    <Entity><dcode>0</dcode><ddesc>(I87)Other disorders of veins - postphlebitic syndrome</ddesc><claimid>34543></claimid><reauthflag>0</reauthflag></Entity>
    <Entity><dcode>0</dcode><ddesc>(J04)Acute laryngitis and tracheitis</ddesc><claimid>34543></claimid><reauthflag>0</reauthflag></Entity>
    <Entity><dcode>0</dcode><ddesc>(J17*)Pneumonia in other diseases - whooping cough</ddesc><claimid>34543></claimid><reauthflag>0</reauthflag></Entity>
    </DiseaseCodes>.
    Regards,
    vikram.

    here is the your XML parse in 11g :
    select *
      from xmltable('//Entity' passing xmltype
    '<?xml version="1.0"?>
    <DiseaseCodes>
    <Entity><dcode>0</dcode><ddesc>(I87)Other disorders of veins - postphlebitic syndrome</ddesc><claimid>34543></claimid><reauthflag>0</reauthflag></Entity>
    <Entity><dcode>0</dcode><ddesc>(J04)Acute laryngitis and tracheitis</ddesc><claimid>34543></claimid><reauthflag>0</reauthflag></Entity>
    <Entity><dcode>0</dcode><ddesc>(J17*)Pneumonia in other diseases - whooping cough</ddesc><claimid>34543></claimid><reauthflag>0</reauthflag></Entity>
    </DiseaseCodes>
    ') columns
      "dcode" varchar2(4000) path '/Entity/dcode',
      "ddesc" varchar2(4000) path '/Entity/ddesc',
      "reauthflag" varchar2(4000) path '/Entity/reauthflag'
    dcode                                                                            ddesc                                                                            reauthflag
    0                                                                                (I87)Other disorders of veins - postphlebitic syndrome                           0
    0                                                                                (J04)Acute laryngitis and tracheitis                                             0
    0                                                                                (J17*)Pneumonia in other diseases - whooping cough                               0
    SQL>
    Using this parser you can create procedure as
    SQL> create or replace procedure myXMLParse(x clob) as
      2  begin
      3    insert into MyXmlTable
      4      select *
      5        from xmltable('//Entity' passing xmltype(x) columns "dcode"
      6                      varchar2(4000) path '/Entity/dcode',
      7                      "ddesc" varchar2(4000) path '/Entity/ddesc',
      8                      "reauthflag" varchar2(4000) path '/Entity/reauthflag');
      9    commit;
    10  end;
    11 
    12  /
    Procedure created
    SQL>
    SQL>
    SQL> exec myXMLParse('<?xml version="1.0"?><DiseaseCodes><Entity><dcode>0</dcode><ddesc>(I87)Other disorders of veins - postphlebitic syndrome</ddesc><claimid>34543></claimid><reauthflag>0</reauthflag></Entity><Entity><dcode>0</dcode><ddesc>(J04)Acute laryngitis and tracheitis</ddesc><claimid>34543></claimid><reauthflag>0</reauthflag></Entity><Entity><dcode>0</dcode><ddesc>(J17*)Pneumonia in other diseases - whooping cough</ddesc><claimid>34543></claimid><reauthflag>0</reauthflag></Entity></DiseaseCodes>');
    PL/SQL procedure successfully completed
    SQL> select * from MYXMLTABLE;
    dcode                                                                            ddesc                                                                            reauthflag
    0                                                                                (I87)Other disorders of veins - postphlebitic syndrome                           0
    0                                                                                (J04)Acute laryngitis and tracheitis                                             0
    0                                                                                (J17*)Pneumonia in other diseases - whooping cough                               0
    SQL>
    SQL>
    Ramin Hashimzade

  • Loading the data from a packed decimal format file using a sql*loader.

    Hi ,
    In one of the project i'm working here i have to load the data into oracle table from a file using a Sql*loader but the problem is the data file is in the packed decimal format so please let me know if there is any way to do this....I search a lot regarding this ..If anybody faced such type of problem ,then let me the steps to solve this.
    Thanks in advance ,
    Narasingarao.

    declare
    f utl_file.file_type;
    s1 varchar2(200);
    s2 varchar2(200);
    s3 varchar2(200);
    c number := 0;
    begin
    f := utl_file.fopen('TRY','sample1.txt','R');
    utl_file.get_line(f,s1);
    utl_file.get_line(f,s2);
    utl_file.get_line(f,s3);
    insert into sampletable (a,b,c) values (s1,s2,s3);
    c := c + 1;
    utl_file.fclose(f);
    exception
    when NO_DATA_FOUND then
    if utl_file.is_open(f) then utl_file.fclose(f); ens if;
    dbms_output.put_line('No. of rows inserted : ' || c);
    end;SY.

  • Loading the data from a text file to a table using pl/sql

    Hi Experts,
    I want to load the data from a text (sample1.txt) file to a table using pl/sql
    I have used the below pl/sql code
    declare
    f utl_file.file_type;
    s varchar2(200);
    c number := 0;
    begin
    f := utl_file.fopen('TRY','sample1.txt','R');
    loop
    utl_file.get_line(f,s);
    insert into sampletable (a,b,c) values (s,s,s);
    c := c + 1;
    end loop;
    exception
    when NO_DATA_FOUND then
    utl_file.fclose(f);
    dbms_output.put_line('No. of rows inserted : ' || c);
    end;
    and my sample1.txt file looks like
    1
    2
    3
    The data is getting inserted, with below manner
    select * from sampletable;
    A     B     C
    1     1     1
    2     2     2
    3     3     3
    I want the data to get inserted as
    A     B     C
    1     2     3
    The text file that I have is having three lines, and each line's first value should go to each column
    Please help...
    Thanks

    declare
    f utl_file.file_type;
    s1 varchar2(200);
    s2 varchar2(200);
    s3 varchar2(200);
    c number := 0;
    begin
    f := utl_file.fopen('TRY','sample1.txt','R');
    utl_file.get_line(f,s1);
    utl_file.get_line(f,s2);
    utl_file.get_line(f,s3);
    insert into sampletable (a,b,c) values (s1,s2,s3);
    c := c + 1;
    utl_file.fclose(f);
    exception
    when NO_DATA_FOUND then
    if utl_file.is_open(f) then utl_file.fclose(f); ens if;
    dbms_output.put_line('No. of rows inserted : ' || c);
    end;SY.

  • How  to load the data from excel  file  into table in oracle using UTL_FI

    How to load the data from excel file into table in oracle
    and from table to excel file
    using UTL_FILE package
    Please give me some example

    This is something i tried in oracle apex
    http://avdeo.com/2008/05/21/uploading-excel-sheet-using-oracle-application-express-apex/
    Regards,
    CKLP

  • Steps to load the data by using flat file for hierarchies in BI 7.0

    Hi Gurus,
    steps to load the data by using flat file for hierarchies in BI 7.0

    hi ,
    u will get the steps int he following blog by Prakash Bagali
    Hierarchy Upload from Flat files
    regards,
    Rathy

  • Large Schema file registered and using folders to load the data (WebDAV)

    I have a very large schema file that I have registered using binary (Schema file > 30000 lines).
    My data has been in MarkLogic and it is organized in a folder structure providing reuse of pieces of the xml.
    It seemed the most logical way to load the data was to create the folder structure in XDB and move the files via WebDAV. We also have extensive XQueries written.
    My question is how do I query the data back our when it is loaded this way. I have read and experimented with resource_view but that is not going to get me what I need. I would like to make use of my xqueries that I have.
    Can i do this, and if so, HOW??

    Sure. Let me lay out a with some more specific information.
    My schema has an overall root level of "Machine Makeup".
    All of these items are defined under it with a lot of element reuse and tons of attribute groups that are used throughout the xml schema. I can do a lot of things but what I cannot do is change the structure of the schema.
    The data is currently in a "folder structure" that more closely resembles the following. I have tried to annotate number of files and keep in mind all of these are working documents and additional "record" (xml files) can be added.
    Composite - contains 12 folders and each of these folders contain xml documents
    compfolder 1
    - compfolder 12
    Most of these folders contain < 200 .xml files (each with id.xml) as the name of the file however one of these directories currently contains 1546 files. They all belong... no real way to split them up further.
    At the same level as composite
    About half of these folders at this level contain > 1000 but less than 3000.
    Like
    PartsUse > 1000 but less than 3000
    transfer of parts 8000 and can continue to grow
    people > 3000 and these would be used in the transfer of parts like "who sold it" Everytime someone new a new one is added etc.
    There are about 12 folders at this level.
    Now, the way the system works is a new person that is not in our list is involved the users get a "new" empty xml document to fill in and we assign it an id and place it in the folder.
    So it would something like this
    Composite
    folder1 - 1000 xml file
    folder 2 - 200 xml files
    folder 12 - < 2000
    Locations
    contains < 2000 xml files
    Activities < 3000 xml files
    PartsTransfer 8000 xml files and growing
    materials < 1000 xml files
    setup < 1000 xml files
    people 3000 xml files and growing
    groupUse < 1000
    and so forth.
    All of the folders contain the links I had previously layed out
    So Activities would have materialLink id = "333" peopleLink ="666"
    And so of and so forth.
    So because my file numbers are > than the optimum about half the time, how would I have 3 separate xmltype tables (I understand mine would be more than 3...)the one schema file that is intertwined with others. Can you show me an example. Would I annotate it in the schema at just those levels. This schema file is so huge I would not want to have to annotate all elements with a table. Can I pick and choose to mimic the folder structure?
    THANKS SO MUCH for your help. I really want to show that Oracle is the way. I can implement it with a simpler structure but not the size of complexity that I need. I haven't been able to find a lot of examples with the new binary registration.
    Edited by: ferrarapayne on Nov 16, 2011 1:00 PM

  • Extracter/datasource that used last time to load the data

    All,
      Can some one help me to find out a table where it stores log/detail of a datasource or extract structure (in BW) when was the last time it was used to load the data ? Help would be highly appreciated.
    Thanks,
    Hari

    Hi Hari, Welcome to SDN.
    In bw side select datasource and monitor. Increase monitor selection to get the details.
    Hope it Helps
    Srini
    [Dont forget to close the call by assigning poings.... - Food for Points: Make a Difference through Community Contribution!|https://www.sdn.sap.com/irj/sdn/index?rid=/webcontent/uuid/007928c5-c4ef-2a10-d9a3-8109ae621a82]

  • Unable to load the data into HFM

    Hello,
    We created new HFM app configured that with FDM, generated output file through FDM and loaded that file through HFM directly 5-6 times, there was no issue till here.
    Then I loaded the file through FDM 4 times successfully, even for different months. But, after 4 loads I start getting Error. Attached is the error log .
    Please help us earliest..
    ** Begin fdmFM11XG6A Runtime Error Log Entry [2013-10-30-13:44:26] **
    Error:
    Code............-2147217873
    Description.....System.Runtime.InteropServices.COMException (0x80040E2F): Exception from HRESULT: 0x80040E2F
       at HSVCDATALOADLib.HsvcDataLoadClass.Load(String bstrClientFilename, String bstrClientLogFileName)
       at fdmFM11XG6A.clsFMAdapter.fDBLoad(String strLoadFile, String strErrFile, String& strDelimiter, Int16& intMethod, Boolean& blnAccumFile, Boolean& blnHasShare, Int16& intMode)
    Procedure.......clsHPDataManipulation.fDBLoad
    Component.......E:\Opt\Shared\Apps\Hyperion\Install\Oracle\Middleware\EPMSystem11R1\products\FinancialDataQuality\SharedComponents\FM11X-G6-A_1016\AdapterComponents\fdmFM11XG6A\fdmFM11XG6A.dll
    Version.........1116
    Identification:
    User............fdmadmin
    Computer Name...EMSHALGADHYFD02
    FINANCIAL MANAGEMENT Connection:
    App Name........
    Cluster Name....
    Domain............
    Connect Status.... Connection Open
    Thanks,'
    Raam

    We are working with the DB team but they have confirmed that they is no issue with the TB, the process we have followed
    As a standard process – while loading the data from FDM or manually to HFM – we don’t write any SQL query. Using the web interface – data would be loaded to HFM application. This data can we viewed by different reporting tools (smart view(excel)/HFR Report/etc.)
    There is no any official documents on oracle website which talk about Insert SQL query which is used to insert data to HFM tables. Even, Hyperion does not provide much details on its internal tables used. Hyperion does not provide much insight on internal structure of HFM system.
    As per Hyperion blogs/forums on internet –HFM stores the base level data in so called DCE tables (for example EMHFMFinal _DCE_1_2013 where EMHFMFinal  is application name, 1 identifies the Scenario and 2013 the Year).  Each row in the DCE table contains data for all periods of a given combination of dimensions (also called an intersection).
    We are trying to load same data file with a replace option( it should delete the existing data before loading the data file).

  • Regarding  Loading the data from DSO to cube.

    Hello Experts,
    I have DSO which loads data from psa  using 7.0 tranformation (using DTP).  And i have a cube which loads the data from that DSO using 3.x transfer rules. Now i have deleted the init request for a infopack before load the data into DSO. But when i load the data from DSO to cube by  right click on DSO -> click On Additional Functions -> update the 3.x data to targets, It is giving me an error like 'Delete init. request REQU_4H7UY4ZXAO72WR4GTUIW0XUKP before running init. again with same selection
    Please help me with this.
    i want to load the data in the init request to cube..
    Thanks

    Hi Shanthi,
    Thanks For reply. I have already deleted the init request from source system to DSO and then i have tried still i am getting the error.
    Thanks

  • SQL Interface - Error in Loading the data from SQL data source

    Hello,
    We have been using SQl data source for loading the dimensions and the data for so many years. Even using Essbase 11.1.1.0, it's been quite a while (more than one year). For the past few days,we are getting the below error when trying to load the data.
    [Mon Jan 10 11:02:56 2011]Local/{App Name}/{DB Name}/{User Id}/Info(1021013)
    ODBC Layer Error: [S1000] ==> [[DataDirect][ODBC DB2 Wire Protocol driver][UDB DB2 for Windows, UNIX, and Linux]CURSOR IDENTIFIED IN FETCH OR CLOSE STATEMENT
    IS NOT OPEN (DIAG INFO: ).]
    [Mon Jan 10 11:02:56 2011]Local/{App Name}/{DB Name}/{User Id}/Info(1021014)
    ODBC Layer Error: Native Error code [4294966795]
    [Mon Jan 10 11:02:56 2011]Local/{App Name}/{DB Name}/{User Id}/Error(1021001)
    Failed to Establish Connection With SQL Database Server. See log for more information
    [Mon Jan 10 11:02:56 2011]Local/{App Name}/{DB Name}/{User Id}/Error(1003050)
    Data Load Transaction Aborted With Error [7]
    [Mon Jan 10 11:02:56 2011]Local/{App Name}///Info(1013214)
    Clear Active on User [Olapadm] Instance [1]
    Interestingly, after the job fails thru our batch scheduler environment, when I run the same script that's being used in the batch scheduler, the job completes successfully.
    Also, this is first time, I saw this kind of error message.
    Appreciate any help or any suggestions to find a resolution. Thanks,

    Hii Priya,
    The reasons may be the file is open, the format/flatfile structure is not correct, the mapping/transfer structure may not be correct, presence of invalid characters/data inconsistency in the file, etc.
    Check if the flatfile in .CSV format.
    You have to save it in .CSV format for the flatfile loading to work.
    Also check the connection issues between source system and BW or sometimes may be due to inactive update rules.
    Refer
    error 1
    Find out the actual reason and let us know.
    Hope this helps.
    Regards,
    Raghu.

Maybe you are looking for