Incremental Load with ODI

Hi All,
I have a few questions related to ODI.
1. ODI can be used to migrate data from DB2/4000 to Oracle 10g on AIX. This is possible right?
2. My DB2 is a core system. Can ODI handle incremental loads to the target source? For example, if a particular tables sees updates every 10 minutes, will ODI be able to handle the changes on this table and reflect the same on the target database?
Any suggestions would be appreciated.
Thanks

Yes , you can by using trigger but the trade off is the source table must be installed a trigger , and several temp tables are created in source database .
CDC by using native journal , we are on the way of testing it .
Edited by: oracletruedb on Oct 7, 2008 2:33 AM

Similar Messages

  • Problem using SQL Loader with ODI

    Hi,
    I am having problems using SQL Loader with ODI. I am trying to fill an oracle table with data from a txt file. At first I had used "File to SQL" LKM, but due to the size of the source txt file (700MB), I decided to use "File to Oracle (SQLLDR)" LKM.
    The error that appears in myFile.txt.log is: "SQL*Loader-101: Invalid argument for username/password"
    I think that the problem could be in the definition of the data server (Physical architecutre in topology), because I have left blank Host, user and password.
    Is this the problem? What host and user should I use? With "File to SQL" works fine living this blank, but takes to much time.
    Thanks in advance

    I tried to use your code, but I couldn´t make it work (I don´t know Jython). I think the problem could be with the use of quotes
    Here is what I wrote:
    import os
    retVal = os.system(r'sqlldr control=E:\Public\TXTODI\PROFITA2/Profita2Final.txt.ctl log=E:\Public\TXTODI\PROFITA2/Profita2Final.txt.log userid=MYUSER/myPassword @ mySID')
    if retVal == 1 or retVal > 2:
    raise 'SQLLDR failed. Please check the for details '
    And the error message is:
    org.apache.bsf.BSFException: exception from Jython:
    Traceback (innermost last):
    File "<string>", line 5, in ?
    SQLLDR failed. Please check the for details
         at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:146)
         at com.sunopsis.dwg.codeinterpretor.k.a(k.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSqlC.treatTaskTrt(SnpSessTaskSqlC.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java)
         at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand(DwgCommandSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandBase.execute(DwgCommandBase.java)
         at com.sunopsis.dwg.cmd.e.i(e.java)
         at com.sunopsis.dwg.cmd.h.y(h.java)
         at com.sunopsis.dwg.cmd.e.run(e.java)
         at java.lang.Thread.run(Unknown Source)

  • DAC Incremental load with new instance

    we have our daily incremental load running from an instance, but now we would like to run load from another instance( not full load) how can this be acheived

    One possible way to do this is to create in awm a cube script with a Load Command with a where clause filter that would filter what records were loaded into the cube.  This cube script could then be run to load only partial data from the instance.

  • HFM DATA LOAD WITH ODI HANGS LONG TIME

    Hi all,
         There's a very strange problem when I loading data from MS SQLServer to HFM with ODI. Specifically, there are 2 interfaces to 2 applications on the same HFM server. Data amount is about 1,300,000 and 650,000 separately.
    The strange thing is when I execute interface individually, it sometimes works well. However when I execute the package contains 2 interfaces, the larger one almost hangs on about 10+ hours every time whether I use an agent or not.
    After some research, it seems that the session hangs on because it cannot get return info from HFM but loading data has already completed. I found some similar problems on OTN like 64bit driver and jre compatible error OR deadlock on table. Different with this one. So, can anyone help on this? Much appreciate in advance!!!
    BTW, ODI and HFM are on the same server but ODI repositary and source of interface are on another MS SQL data server. The version is as below:
    HFM 11.1.1.3.0.956
    ODI 11.1.1.6.0
    win server 2003 x86
    MS SQLServer 2008 R2
    win server 2008 x64
    Regards,
    Steve

    Hi SH,
         source is MS SQLServer 2008 R2, staging area is on the source side, target is a HFM 11.1.1.3.0.956 based on SQLSERVER.
         KM is a standard 'IKM SQL to Hyperion Financial Management Data'.
         No transformation logic but only a filter to select data in current year.
    Besides, I have do some performance tuning as guide tolds:
    REM #
    REM # Java virtual machine
    REM #
    set ODI_JAVA_HOME=D:\oracle\Java\jdk1.6.0_21
    REM #
    REM # Other Parameters
    REM #
    set ODI_INIT_HEAP=512m
    set ODI_MAX_HEAP=1024m
    set ODI_JMX_PROTOCOL=rmi
    In Regedit:
    EnableServerLocking: 1
    MaxDataCacheSizeinMB :1000
    MaxNumDataRecordsInRAM: 2100000
    MultiServerMaxSyncDelayForApplicationChanges:300
    MultiServerMaxSyncDelayForDataChanges:300
    After some reaserch, I think the problem can be located at the HFM-ODI adapter or HFM side(maybe HFM cannot respond a completed info to ODI), do you have any idea? Thanks in advance

  • How to load with ODI secondary hierarchy on HFM

    Hi,
    I tried to load on HFM application a secondary hierarchy for entity dimension (like Essbase shared members), but the flow is aborted and I found on log file the follow message:
    Error: Member defined more than once in load file
    +>>>>>>Ent_1_1+
    How can I solve this problem?
    Thans in advance
    Ant

    Hi,
    Please use Windows 8 driver from the following link:
       http://h20000.www2.hp.com/bizsupport/TechSupport/D​riverDownload.jsp?prodNameId=447377&locale=en_us&t​...
    Regards.
    BH
    **Click the KUDOS thumb up on the left to say 'Thanks'**
    Make it easier for other people to find solutions by marking a Reply 'Accept as Solution' if it solves your problem.

  • W_gl_balance_a duplicating/growing data with incremental load plan

    Hi, all!
    I'm executing incremental load plan for Financial Analytics Offering and I'm facing the data doubling/aggregating in every execution.
    Any one faced this problem?
    Solutions Versions:
    - OBIA 11.1.1.7.1
    - OBIEE 11.1.1.7.131017
    - ODI 11.1.1.7.0

    Hi, Sher Ullah Baig!
    I was openned a SR to Oracle, but I still waiting response about 2 weeks.
    Thanks and Regards!
    Fernando Gonçalves
    E-Novasys Business Intelligence Team

  • Problem loading metadata with ODI : "Single sign on validation failed"

    Hi,
    When I load metadata in Planning using ODI, the load fails at the step "Prepare for loading" with this error message : "Single sign on validation failed".
    Can you help me please ?
    Thanks.
    Virgile.

    "com.hyperion.odi.planning.ODIPlanningException: Single sign on validation failed."
    Means you have put the incorrect account details in the topology manager, update the username/password
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Time Dimension Population with ODI 11.1.1.5

    Hi All,
    I am planning to build a time/Date dimension to use with the HR Fact tables (sourced from Oracle HRMS).
    Customer would like to supply start and end dates (01-JAN-05 TO 31-DEC-11) and i should be able to generate and load the time data between these dates.
    Down the line, if the customer to append the Time Dimension for future data, he should be able to specify the new start and end dates (01-JAN-12 To 31-DEC-15), i should be able to incrementally load the new time data, instead of truncating and regenerating all the time data.
    Any ideas on implementing this requirements using ‘Oracle Data Integrator 11.1.1.5’?
    Thanks in advance
    Regards,
    Venkat

    This is what I did recently,
    Download the 'SQL_as_source' knowledge module from the Oracle ODI Blog here : http://blogs.oracle.com/warehousebuilder/entry/odi_11g_simple_flexible_powerful
    Used a temporary interface with the following simple query :
    with row_generator as (
    SELECT to_date('01011980','ddmmyyyy') + LEVEL - 1 DT
    FROM sys.dual
    CONNECT BY LEVEL <=
    /* End Date in here -> */ to_number(to_date('31122030','ddmmyyyy') -
    /* Start Date in here -> */ to_date('01011980','ddmmyyyy')) +1 )
    SELECT DISTINCT
    'YEAR' as TIME_LEVEL,
    NULL as YEAR_MONTH,
    NULL as YEAR_QTR,
    to_char(DT,'YYYY') as YEAR
    from row_generator
    UNION
    SELECT DISTINCT
    'QUARTER',
    NULL,
    to_char(DT,'YYYY')||' / '||to_char(DT,'Q'),
    to_char(DT,'YYYY') from row_generator
    UNION
    SELECT DISTINCT
    'MONTH',
    to_char(DT,'YYYY')||' / '||to_char(DT,'MM'),
    to_char(DT,'YYYY')||' / '||to_char(DT,'Q'),
    to_char(DT,'YYYY')from row_generator
    order by 1,2,3,4
    As you can see in the above code, it would be possible to pass in ODI variables as the start and end dates , you could either allow the user to type these in at run time or read them from a control table.
    Use the temporary interface as the source data for your Time dimension interface (you might want to add more Time Attributes - Mine was very simple) , then use an IKM Incremental update approach (Im using IKM Oracle Incremental update (Merge)).
    Hope this helps
    Rgrds
    Alastair
    p.s - my time dimension is loading multiple levels (Years, Quarters, Months) into the same table - the SQL contains a row for each day though so easily adapted to report days.
    Edited by: PeakIndicators_Alastair on Jan 24, 2012 12:05 PM

  • Problem getting MySQL driver loaded in ODI Studio (ODI 11g)

    Hello -
    I am trying to create a new data server in the physical architecture under the MySQL technology. Since there is no MySQL JDBC driver included as there was in ODI 10g, I looked into the documentation to find a similar drivers directory that I had used in ODI 10g for loading other JDBC drivers. I would like to be able to select the MySQL Connector/J driver from the drop-down list of usable drivers in the Driver dialog window. I found this link, that explains how to install drivers:
    http://download.oracle.com/docs/cd/E14571_01/core.1111/e16453/install.htm#CHDBIFAJ
    I followed the documents instructions, added the file "mysql-connector-java-5.1.13-bin.jar" to my directory "C:\Documents and Settings\<my username>\Application Data\odi\oracledi\userlib" and restarted ODI Studio. I still cannot get the app to recognize the new driver to use, am I missing a step or is there a known issue with this method? More importantly has anyone been able to get this to register with the Driver dialog box successfully?
    With ODI Studio open, I attempted to delete the mysql jar file from my userlib directory and there was an error that said the file was in use... so I assume that ODI Studio is using the file but not registering it within the Driver selection dialog.

    I think we found our own answer... apparently the ODI Studio interface does not dynamically load that drop-down list from registered drivers in the classpath. We can still use the driver name by manually typing it into the "JDBC Driver:" text box.

  • Duplicate rows in Hierarchy Table created when running incremental load

    I copied an out of the box dimension and hierarchy mapping to my custom folders (task hierarchy) this should create the same wids from dimension to hierarchy table and on full load does this using the sequence generator. The problem I am getting is whenever I run a incremental load instead of updating, a new record is created. What would be the best place to start looking at this and testing. A full load runs with no issues. I have also checked the DAC and run the SDE trunc always and SIL trunc for full load only.
    Help appreciated

    Provide the query used for populating the child records. Issue might be due to caching.
    Thanks
    Shree

  • OBIA Financial Analytics - ETL Incremental Load issue

    Hi guys
    I have an issue while doing ETL Incremental load in DEV. Source, Target ORACLE.
    issue with these two tasks: SDE_ORA_GL_JOURNALS and SDE_ORA_ImportReferenceExtract
    incremental load is holding at SDE_ORA_GL_JOURNALS.....on Database sessions the query is completed and also the session is done but no update in the session log for that session in informatica. It says hust ' SQL Query issued to database' . no progress from there. and the task in both informatica session monitor and DAC says running and keeps running for ever. No errors seen in any of the log files.
    Any idea on whats happening? I checked session logs, DAC servr logs, Database alert logs, Exception logs on source and found nothing.
    I tried to run these Informatica generated queries in SQL developer and they ran well and I did see the results. More over, weird thing is from past three days and about 10 runs....the statis tics are
    both thses tasks run in parallel and most of the times Import references extract is completed first and then GL_Journals is running for ever
    In one run, GL_Journals is done but then Import reference extract is running for ever. I dont exactly understand whats happening. I see both the queries running in parallel on source database.
    Please give me some idea on this. And this same stuff is running good on QA. I dont know how this is working.....any idea on this is appreciated. Thank you. let me know of ay questions. Thank you in advance.

    Please refer this:
    http://gerardnico.com/wiki/obia/installation_7961

  • How to Read file name which we are dealing with ODI File tool

    Hi,
    We are using ODi10g version and we have requirement to move file from one place to another place. We are using ODIFileMove utility but we also want to read file name.
    Any help.
    Thanks in Advance.

    You can accomplish this with a fairly simple Jython script.  Use the os.listdir(<directory>) command to get the name of files in a given directory. 
    You can then (still in the Jython script) loop through the files and move them to a desired location (bypassing the OdiFileMove tool) OR use the Jython script to write the file names to a SQL table.  Then, use an ODI procedure to loop through the newly inserted records and store the file name in an ODI variable that you can then use in your OdiFileMove tool etc.
    I often refer to this blog entry from Gurcan Orhan as a starting point for this kind of task: Loading multiple files with ODI | Gurcan Orhan&amp;#039;s Oracle Data Integrator Blog

  • Incremental Loading of a Tree Component

    I'm working on an explorer-type interface in Flex 2 for
    browsing a remote file repository. It's a standard split-pane
    affair with a directory tree on the left and a listing on the
    right.
    Because the entire directory tree can be rather large, I need
    to load it incrementally as the user expands nodes rather than all
    at once. I failed to find a relevant example so I wrote my own. It
    works, but I'm new to Flex and am not sure if there's an easier way
    or if there are any pitfalls to the way I did it.
    I posted my code here:
    http://xocoatl.blogspot.com/2007/01/incremental-loading-of-tree-in-flex-2.html
    Any comments here or on the blog are appreciated; I'm
    guessing that having a good example of the "right" way to do this
    will be useful to many others.
    Thanks.

    i am also using another workaround using CSS trick(inlineStyle). if #{node.children} is null, i am placing an empty 10x10 white png image file over expand icon on tree node.
    following code is nodeStamp of tree component.
    <f:facet name="nodeStamp">
      <af:group id="g1">
        <af:image source="/images/spacer.png" id="i1" inlineStyle="border: 2px solid white; position: absolute; margin-left:-14px;" rendered="#{node.children == null}"/>
        <af:commandLink text="#{node.name}" id="cl1" partialSubmit="true"/>
      </af:group>
    </f:facet>

  • Hyperion Planning- Account dimension - Sort Order Issue with ODI 11.1.1.5

    Hello Experts,
    First of all, I thank you everybody in this forum for posting their valuable suggestions.
    I am facing a strange issue with ODI (Version 11.1.1.5) while loading metadata to Planning ACCOUNT dimension (Version 11.1.2.x). I am using a Oracle view to load metadata to this dimension. I have defined the sort order in the view and enabled SORT_ORDER_BY_INPUT. For the fresh load to the dimension, everything is working fine. But, when I move a member from one parent to another parent, and define it as the top member of new parent, it is missing its sort order.
    But it is working fine if you define it as the second, third...or last member under new parent. I have tried all the possible ways by flipping all the properties but unable to fix it.
    Please help me here to resolve this issue.
    Your help is greatly appreciated!!
    Thanks
    Edited by: user11965311 on Sep 20, 2011 7:07 PM

    To be honest I think you are best logging it with Oracle because it sounds like you are doing everything correctly so it could be a bug.
    Have you tested this on any previous versions of ODI or is it a new installation?
    If it does get logged with Oracle please keep me informed of the outcome as I would be interested to hear.
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • 11g (11.2.0.1) - dimension operator very slow on incremental load

    Dimension operator very slow in porcessing incremental loads on 11.2.0.1 Have applied cumulative patch - still same issue.
    Statistics also gathered.
    Initial load into empty dimension performs fine (thousands of records in < 1 min) - incremental load been running over 10 mins and still not loaded 165 records from staging table.
    Any ideas?
    Seen in 10.2.0.4 and applied patch which cured this issue.

    Hi,
    Thanks for the excellent suggestion.
    Have run other mapings which maintain SCD type 2 using dimesnion operator behave similary to 10g. - have raised issue with this particular mapping with Oracle - awaiting response.
    One question - when look at the mappings which maintain SCD Type 2s looks to join on dimension key and the surrogate ids.
    What is best practice regarding indexing of such a dimension, e.g is it recommended to index dimension key and the surrogate ids along with the natural/nbusiness keys?
    Thanks

Maybe you are looking for