Metadata load using ODI

Hi
I am trying to load new Employees using ODI.Our Employee members are in Alphabetical order.Is there any way i can load new employees in Alphabetical order.
Please suggest,
Thanks,

Then I believe it will add them to end.
Say you had
Par
-----A1
-----A3
Then you tried to load
Par | A1
Par | A2
Par | A3
You should end up with
Par
-----A1
-----A2
-----A3
If you tried to load just
Par | A2
Then you would end up with
Par
-----A1
-----A3
-----A2
which is pretty much the way planning works, if you tried it through the outlineload utility you would end up with the same.
Cheers
John
http://john-goodwin.blogspot.com/

Similar Messages

  • Metadata loading using ODI

    Hi John
    I am loading Employee metadata using ODI(10g). In IKM SQL to Hyperion Planning, i put the option "LOAD_ORDER_BY_INPUT" option to "NO.I expected that, ODI will sort the employee members and load it. But its not sorting, it loaded in the same sequence as in my load file.
    Please suggest, how can i sort the members
    Thanks,

    Hi
    This post would be more suited to the ODI forum.
    Anyway, check out ODI Order By Issue

  • Metadata load using ERPi

    Hi All,
    I am new to ERPi, Our client is willing to implement ERPi Metadata load. The source is People soft 9.1 and the target Hyperion EPMA planning 11.1.2.2.300.20.
    My source COA tree structure is different with my target Hyperion COA structure, can we implement metadata load using ERPi, if so how can we do it?.
    and what are the considerations to be taken care in the metadata load.
    Any thoughts on these would really appreciate your response. Thanks in advance.
    Best regards,
    Vishy

    The response from support is correct.

  • Can metadata be loaded using ODI into classisc applcn and upgraded to epma

    Hi John,
    As ODI is not compatible with EPMA, cane we load metadata into classic using ODI and then upgrade the applcns to EPMA?
    Is this method suggestible? When are the adapters for Planning to connect epma will be released. any idea?

    Hi John,
    Thanks for the information.
    As per your blog you have loaded the entity dimension using the Hierarchy and members tables. Is the structure of these two tables same as interface tables?
    And can i insert all the dimensions and their members with only these two tables or do i need to create two sets of tables for every dimension i have.

  • GL Data load using ODI to Essbase

    Hi, I am trying to load GL actual data to essbase application using ODI. Source file is having 10 columns and the Target is having 11 columns. We are using rules file to load data into essbase. Rules file will split 9th column as two columns and will load the data into Essbase. When we test the rules file in essbase data is getting load into application. when we use the same rule file in ODI interface data is not getting load and giving error as "Unknown member" for the member which we are spliting in 9th column. Source file: HSP_RATES ACCOUNT PERIOD YEAR SCENARIO VERSION CURRENCY ENTITY SFUND PROGRAM DATA HSP_InputValue 611101 Jul FY13 ACTUAL Final Local 0000 SBNR AC0001PS0001 25000 AC0001PS0001 is the concatenated string from GL. we will split this as two columns using rules file to load into essbase application. Please suggest what might be the reason for the error. How to do the mapping between source and target. I have mapping one column(AC0001PS0001) to two dimensions (Program, Activity) in Essbase. Please suggest. Thanks Sri

    In ODI, what you have to do is to split it in the ODI itself. While you are mapping, you can use SQL functions to map it to two different columns. Similar to the way you are doing it in Rule file.
    Regards
    Amarnath
    ORACLE | Essbase

  • Data loading using ODI

    Hi,
    I am using ODI 10g for loading data into oracle 10g database from a flat file. The flat file is in fixed format, but the data is in a single row, so the data is not loaded correctly. For example, 1000 records are shown in 1 line.
    Can someone please suggest, what customization should I make to the IKM to achieve this.
    Please let me know the questions you may have.
    Thanks a lot.

    Hi John,
    I reversed my Planning model and able to see those columns now. I tried executing my interface and got the below error
    The source result set contains column [EMPLOYEE_ACCOUNT] which has no corresponding column on the planning side.
    My source is an oracle table and i have seperate columns for each dimension member. What I did is, mapped the Data column (JobTitle) from source to the Target column(JobTitle) and rest all dimension member columns from source to Point of View column seperated by a comma. My Planning cube name is EMPLOYEE, so in th data cube name column i manually entered "EMPLOYEE" and tried executing it.
    Please guide me as always
    Thanks,
    Sravan

  • BINARY FILE loading using ODI

    Hi ,
    is it possible to load a binary file using  ODI ?
    Basically i'm looking for file to table and file to file conversion .
    Regards.

    DBMS_LOB.LOADBLOBFROMFILE (
      dest_lob IN OUT NOCOPY BLOB,
      src_bfile IN BFILE,
      amount IN INTEGER,
      dest_offset IN OUT INTEGER,
      src_offset IN OUT INTEGER);
    you can specify the src. Also, you can ftp the file using ODI sftp tool and then load.
    There are other ways to load as well,
    1. Sql loader aslo supports loading BLOBs. Odi has got IKM Sql loader however I have never tried with BLOB and need to check if it supports. The below link might be helpful,
         ORACLE-BASE - Loading LOB Data Using SQL*Loader
    2. External tables could be used.
    3. Java program could be called in ODI proc.

  • Data load using ODI 11.1.1.5 to hyperion.

    Hi All,
    I have installed ODI and we have already installed Hyperion.
    After installation of ODI , I have created Master repository and Work repository.
    Now i am planning to load data to hyperion planning applications dimensions(Eg. Entity).
    1) In Topology section, i did right click on Hyperion planning and clicked on "New Data Server". Filled the required infomartion and saved it.
    2) Then i right clicked on newly created Data server and opted "New physical schema", but after that it is not showing any application in Application(Catalog) and Application (Work Catalog) drop downs.
    Before those steps we had created planing applications in Hyperion.
    Please let me know what is wrong/missing in above steps or share the steps to load the data in hyperion.
    -PM

    Yes , i have entered the values in text boxes/drop downs . Afterward i tried to save it but it ask to fill "context".Then i seletected Context option, In that i have only one option global, I opted that.
    And in second field logical schema , doesnt have any value. System doesnt allow to save without giving these values.
    Can you please me further steps.
    -PM

  • Is this Possible using ODI / ODQ ?

    Hi ALL,
    Is the below requirement possible with ODI / ODQ?
    (right now I am doing it mannually, and its like a monthly/weekly process)
    If we can do the same using ODQ how I need to start and proceed?
    My requirement is some thing like this. I have two columns
    Address1 (street address)
    Address2 (additional Information)
    In some cases the address is not in the proper format. So I need to do some data cleansing (data scrubbing) before loading using ODI
    Use Address2 if Address1 is PO Box, building name, or any info aside from a street address. Only one address column is needed in the output (Address1)
    For example :
    INPUT
    ADDRESS1--------------------------------------------------ADDRESS2
    HIGHWAY # ------------------------------------------------5010944 MARSH RD
    PO BOX 40964---------------------------------------------1100 AA CORP PARKWAY
    DBA VANTAGE PP---------------------------------------11031 PERRY HIGHWAY
    HOME INFUSION THERAPY SERVICES--------11031 PERRY HIGHWAY
    7275 BUFFALO ROAD----------------------------------(WAS 6855 BUFFALO ROAD)
    7601 N FEDERAL HWY---------------------------------#2506
    MCAS ---------------------------------------------------------(MARINE CORPS AIR STATION)
    212 BRADDOCK AVENUE------------------------------
    212 NINTH STREET----------------------------------------
    2138 JONATHON DRIVE---------------------------------
    OUTPUT (Note : we don’t need the Address2 column in the output)
    ADDRESS1----------------------------------ADDRESS2
    10944 MARSH RD------------------------10944 MARSH RD
    1100 AA CORP PARKWAY-----------1100 AA CORP PARKWAY
    11031 PERRY HIGHWAY-------------11031 PERRY HIGHWAY
    11031 PERRY HIGHWAY-------------11031 PERRY HIGHWAY
    7275 BUFFALO ROAD------------------(WAS 6855 BUFFALO ROAD)
    7601 N FEDERAL HWY----------------#2506
    MCAS-----------------------------------------(MARINE CORPS AIR STATION)
    212 BRADDOCK AVENUE------------
    212 NINTH STREET----------------------
    2138 JONATHON DRIVE---------------
    I have posted a similar kind of question in DATA Quality (ODQ) forum
    Forum Home » Master Data Management (MDM) » Data Quality
    ADDRESS CLEANING(US) USING ODQ
    Thread: ADDRESS CLEANING(US) USING ODQ
    if any body is working on address cleansing using ODQ, Please guide me to start working on it.
    IS there any specific documents for ADDRESS CLEANING which gives more information about the process flows such as
    Transformer, Customer Data Parser, Sort for Postal Matcher, Postal Matcher, Window Key Generator,Sort for Linking, Relationship Linker, Commonizer, and Data Reconstructor.
    Thanks in Advance,
    Rathish

    Hi Rathish,
    I was building the java code but I got other idea to make more simple.
    1) load the values into a temp table with one column for each address (means 2 columns per line) plus 2 other columns to control.
    Something like:
    table temp_data
    columns: add_1 , control_1, add_2, control_2
    You can do that thru an interface. Let the fields "control" with no value when loaded
    2) create a table to storage the values you sent to me (Avenue, street...)
    table street_names
    columns: street_name, country (to your future use)
    3) Create an ODI procedure (or you can customize a KM with the same technique)
    3.1) Step 1 at target Tab, technology javaScript
    code:
    <@String vU=" ";@>
    3.2.1) step 2 - Tab Source - technology "your database" and the Logical Schema where the "street_names" table as created
    select street_name
    from street_names
    where country = 'US'
    3.2.2) step 3 - Tab Target - tecnology javaScript
    <@vU+="or instr(upper(add_1),upper('#street_name'))>0 ";@>
    3.3) step 3 - Target Tab - technology "your database" and the Logical Schema where the "temp_data" table as created
    <@vU=vU.substring(3);@>
    update temp_data
    set control_1 = 'Y'
    where <@=vU@>
    3.4) step 4 step 3 - Target Tab - technology "your database" and the Logical Schema where the "temp_data" table as created
    <@vU=vU.replaceAll("add_1","add_2")@>
    update temp_data
    set control_2 = 'Y'
    where <@=vU@>
    4) Now you have a table (temp_data) with a flag "Y" where there is a street name and "null" where doesn't exist. Just generate your file by a interface or odiSqlUnload Toll.
    Put all together into a package and you will have your process.
    I create in this way to minimize the access to the temp_table, just 2 updates.
    can be used some alternatives to tunning the hole process if the time doesn't match what you need.
    That is a fast solution. Something more elaborated can be done if necessary.
    Does it help you?
    Cezar Santos

  • Error while loading metadata into planning using ODI

    Hi John,
    While trying toload members into planning using ODI the following error occurred. Please suggest
    java.lang.Exception
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java)
         at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand(DwgCommandSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandBase.execute(DwgCommandBase.java)
         at com.sunopsis.dwg.cmd.e.i(e.java)
         at com.sunopsis.dwg.cmd.h.y(h.java)
         at com.sunopsis.dwg.cmd.e.run(e.java)
         at java.lang.Thread.run(Unknown Source)

    Hi John,
    While trying toload members into planning using ODI the following error occurred. Please suggest
    java.lang.Exception
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java)
         at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand(DwgCommandSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandBase.execute(DwgCommandBase.java)
         at com.sunopsis.dwg.cmd.e.i(e.java)
         at com.sunopsis.dwg.cmd.h.y(h.java)
         at com.sunopsis.dwg.cmd.e.run(e.java)
         at java.lang.Thread.run(Unknown Source)

  • Unable to load metada into Hyperion Planning using ODI

    HI All,
    I am using ODI 11g and trying to load the metadata for Hyperion Planning application 11.1.2.1. But I couldnt load all the Product members in Product Dimension.
    Could able to load few members but was able to run the interface successfully but thrown error while loading few members(Most of the members are shared members, verified that base members existed). Please find the below error log.
    2011-11-17 06:18:06,534 INFO [SimpleAsyncTaskExecutor-2]: Oracle Data Integrator Adapter for Hyperion Planning
    2011-11-17 06:18:06,550 INFO [SimpleAsyncTaskExecutor-2]: Connecting to planning application [xxxx] on [xxxx]:[11333] using username [xxxx].
    2011-11-17 06:18:06,675 INFO [SimpleAsyncTaskExecutor-2]: Successfully connected to the planning application.
    2011-11-17 06:18:06,706 INFO [SimpleAsyncTaskExecutor-2]: The load options for the planning load are
    Dimension Name: Product Sort Parent Child : false
    Load Order By Input : false
    Refresh Database : false
    2011-11-17 06:18:06,768 INFO [SimpleAsyncTaskExecutor-2]: Begining the load process.
    2011-11-17 06:18:06,878 DEBUG [SimpleAsyncTaskExecutor-2]: Number of columns in the source result set does not match the number of planning target columns.
    2011-11-17 06:18:07,440 INFO [SimpleAsyncTaskExecutor-2]: Load type is [Load dimension member].
    2011-11-17 06:18:07,487 ERROR [SimpleAsyncTaskExecutor-2]: Record [[D090, Mobile Mice Cordless_Prod type, D090-Flores DJ, null, Shared, null, null, null, null, null, null, null, +, null]] was rejected by the Planning Server.
    2011-11-17 06:18:07,503 ERROR [SimpleAsyncTaskExecutor-2]: Record [[D091, Mobile Mice Cordless_Prod type, D091-Fiji BT, null, Shared, null, null, null, null, null, null, null, +, null]] was rejected by the Planning Server.
    2011-11-17 06:18:07,518 ERROR [SimpleAsyncTaskExecutor-2]: Record [[D093, Mobile Mice Cordless_Prod type, D093-New York, null, Shared, null, null, null, null, null, null, null, +, null]] was rejected by the Planning Server.
    2011-11-17 06:18:07,550 ERROR [SimpleAsyncTaskExecutor-2]: Record [[D060, Mobile Mice Cordless_Prod type, D060-Corbusier Vista, null, Shared, null, null, null, null, null, null, null, +, null]] was rejected by the Planning Server.
    2011-11-17 06:18:07,565 ERROR [SimpleAsyncTaskExecutor-2]: Record [[D064, Mobile Mice Cordless_Prod type, D064-Sardinia Laser, null, Shared, null, null, null, null, null, null, null, +, null]] was rejected by the Planning Server.
    2011-11-17 06:18:07,581 ERROR [SimpleAsyncTaskExecutor-2]: Record [[D065, Mobile Mice Cordless_Prod type, D065-Ibiza Optical, null, Shared, null, null, null, null, null, null, null, +, null]] was rejected by the Planning Server.
    Regards,
    SP

    Can you check the error log file as you have posted the log file, it should give more information to why the records were rejected.
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Loading members to Essbase using ODI

    Hi All,
    We are planning to use ODI to load members from Oracle database to Essbase using ODI. Should we convert all the dimension tables in Oracle database to Parent-Child relation and then load it or can we use the existing hierarchial structure in Oracle database.
    Thanks in Advance.
    Kannan.

    Hi,
    If you reverse engineer an essbase cube it creates columns :- ParentName, MemberName.
    When you load metadata into essbase using ODI it will need to be mapped against Parent and Member, to make it easier it would be useful if your source data is in parent/child format.
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • SOS!!!!----Error for Loading data from Oracle 11g to Essbase using ODI

    Hi all.
    I want to load data from oracle database to essbase using ODI.
    I configure successfully the physical and logical Hyperion essbase on Topology Manager, and got the ESSBASE structure of BASIC app DEMO.
    The problem is.
    1. When I try view data right click on the essbase table,
    va.sql.SQLException: Driver must be specified
         at com.sunopsis.sql.SnpsConnection.a(SnpsConnection.java)
         at com.sunopsis.sql.SnpsConnection.testConnection(SnpsConnection.java)
         at com.sunopsis.sql.SnpsConnection.testConnection(SnpsConnection.java)
         at com.sunopsis.graphical.frame.b.jc.bE(jc.java)
         at com.sunopsis.graphical.frame.bo.bA(bo.java)
         at com.sunopsis.graphical.frame.b.ja.dl(ja.java)
         at com.sunopsis.graphical.frame.b.ja.<init>(ja.java)
         at com.sunopsis.graphical.frame.b.jc.<init>(jc.java)
    I got answer from Oracle Supporter It's ok, just omit it. Then the second problem appear.
    2. I create an interface between oracle database and essbase, click the option "staging area deffirent from target"(meaning the staging is created at oracle database), and using IKM SQL to Hyperion Essbase(metadata), execute this interface
    org.apache.bsf.BSFException: exception from Jython:
    Traceback (innermost last):
    File "<string>", line 61, in ?
    com.hyperion.odi.essbase.ODIEssbaseException: Invalid value specified [RULES_FILE] for Load option [null]
         at com.hyperion.odi.essbase.ODIEssbaseMetaWriter.validateLoadOptions(Unknown Source)
         at com.hyperion.odi.essbase.AbstractEssbaseWriter.beginLoad(Unknown Source)
         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
         at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
         at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
         at java.lang.reflect.Method.invoke(Unknown Source)
         at org.python.core.PyReflectedFunction.__call__(PyReflectedFunction.java)
         at org.python.core.PyMethod.__call__(PyMethod.java)
         at org.python.core.PyObject.__call__(PyObject.java)
         at org.python.core.PyInstance.invoke(PyInstance.java)
         at org.python.pycode._pyx1.f$0(<string>:61)
         at org.python.pycode._pyx1.call_function(<string>)
         at org.python.core.PyTableCode.call(PyTableCode.java)
         at org.python.core.PyCode.call(PyCode.java)
         at org.python.core.Py.runCode(Py.java)
         at org.python.core.Py.exec(Py.java)
    I'm very confused by it. Anybody give me a solution or related docs.
    Ethan.

    Hi ethan.....
    U always need a driver to be present inside ur <ODI installation directory>\drivers folder.....for both the target and the source......Because ojdbc14.jar is the driver for oracle already present inside the drivers folder ....we don't have to do anything for it.....But for Essbase ...you need a driver......If u haven't already done this.....try this.....
    Hope it helps...
    Regards
    Susane

  • Error regarding data load into Essbase cube for Measures using ODI

    Hi Experts,
    I am able to load metadata for dimensions into Essbase cube using ODI but when we are trying same for loading data for Measures encountring following errrors:
    Time,Item,Location,Quantity,Price,Error_Reason
    '07_JAN_97','0011500100','0000001001~1000~12~00','20','12200','Cannot end dataload. Essbase Error(1003014): Unknown Member [0011500100] in Data Load, [1] Records Completed'
    '28_JAN_97','0011500100','0000001300~1000~12~00','30','667700','Cannot end dataload. Essbase Error(1003014): Unknown Member [0011500100] in Data Load, [1] Records Completed'
    '28_JAN_97','0011500100','0000001300~1000~12~00','500','667700','Cannot end dataload. Essbase Error(1003014): Unknown Member [0011500100] in Data Load, [1] Records Completed'
    Can anyone look into this and reply quickly as it's urgent requiremet.
    Regards,
    Rohan

    We are having a similar problem. We're using the IKM SQL to Hyperion Essbase (DATA) knowledge module. We are mapping the actual data to the field called 'Data' in the model. But it kicks everything out saying 'Unknown Member [Data] in Data Load', as if it's trying to read that field as a dimension member. We can't see what we missed in building the interface. I would think the knowledge module would just know that the Data field is, um, data; not a dimension member. Has anyone else encountered this?
    Sabrina

  • ODI Planning metadata load takes very long time

    Hi,
    I am using ODI for Hyperion Planning metadata load.
    We are facing performance issues, as the process is taking a very long time.
    The process uses "LKM File to SQL" and "IKM SQL to Hyperion Planning"
    The number of rows to process in the file is around 70000. The file is generated from DRM. The ODI integration process takes around 2 hours to process and load the file to Planning. Even if we add 1 new row to the file and everything else remains same in the file, the process takes that long.
    So, the whole process takes around 3 hours to load to Planning for all dimensions.
    I tried increasing the fetch rows to 200 in source but there is no significant increase in performance. The Heap size is set to maximum of 285 MB in odiparams.bat.
    How can the performance be improved?
    Can I use different LKM or change any other setting?
    Thanks,
    RS

    Hi John,
    In my current implementation, the dimension hierarchies are maintained in DRM.
    So, business directly makes changes in DRM and exports the hierarchies in a text file.
    I agree that loading 70000 records is odd on regular basis, but it makes it easier for business to retain control of their hierarchies and maintainance is easy in DRM.
    On bulk loading to DB table and loading to Planning, I have 2 questions:
    1. Do you think that "LKM SQL to SQL" [Oracle to Planning] will have significant improvement in performance over "LKM File to SQL" [File to Planning], as we are still using "Sunopsis Memory engine" as staging area?
    2. I checked your blog, there you have suggested using "Sunopsis memory engine" for "LKM SQL to SQL".
    Is it mandatory to use "Sunopsis emory engine" as staging area, can we use any other user defined staging area [Oracle tables]?
    Cheers,
    RS

Maybe you are looking for