Multiple source columns from Oracle DB to Hyperion Essbase cube

Hi Experts ,
We have a Source as oracle table to the target Hyperion Essbase Cube(Planning) 11.1.2 using ODI 11g.
Right now my issue is We need to map multiple columns from Oracle DB to One Dimension in Essbase (Planning).
For example -> * In Oracle table i have columns as Name,Volume,unit and item *
In Essbase the dimension called "*CatDefinition has lower level members as Name,Volume,unit and item * "
Now using ODI ,* I want to map the oracle table columns(Multiple) to the dimension(Catdefinition) ? * Please suggest me with idea to achieve multiple column mapping to the single dimension in Essbase .
Thanks in Advance !
Keny Alex

Do you even need FDM? You could build a load rule in Essbase directly. If you need to map, then you would register the Essbase target adapter.

Similar Messages

  • FDM:how to load data from oracle DB to hyperion essbase

    Hi,
    I want to get data from the orcle database and load it into Hyperion Essbase.Do i need to have any source adapter for it or only by using essbase target adapter I can get it ?
    Pls guide me with this.
    Thanks

    You want an integration import script along with an Essbase target adaptor -- see page 197 of the fdm admin guide at:
    http://download.oracle.com/docs/cd/E12825_01/epm.111/fdm_admin.pdf
    You will need to use an ODBC compliant Oracle driver for FDM to connect to your db --
    the examples in the admin guide are SQL server connections. The Oracle connection strings are
    slightly different.
    The below wiki shows an ODBC connection to Oracle:
    http://www.orafaq.com/wiki/ODBC_FAQ
    Regards,
    John A. Booth
    http://www.metavero.com

  • Data Integration from Oracle GL to Hyperion  Planning

    Hi all,
    Can I knw the steps for extarcting the from Oracle GL Tables to Hyperion Planning?
    We have installed hyperion 9.3.1 and ODI 10.
    Please let me know how will we do this process by using ODI?
    Thanks

    Hi User##########,
    In a nutshell you need to do the following...
    1. Create a Project in ODI
    2. Define your Source and Target data stores in Topology
    3. Import the appropriate Knowledge Modules
    4. Reverse Engineer your Source and Target modules to create a model
    5. Create an integration to load from your Source (Oracle GL) to your Target (Hyperion Planning -- I actually prefer to load the data directly to Essbase)
    I just did this last month. When loading the data I used an Essbase load rule to ensure Essbase knew the column order.
    I was also able to load data via ODI from a flat file WITHOUT a load rule but the order of the data in the file was very important. I was also not able to substitute or alter the information in the integration when this was done. So in the end I still needed to have a load rule handy.
    I'd recommend reading up on the ODI documentation for this -- some of it is actually quite good. http://download.oracle.com/docs/cd/E10530_01/doc/nav/portal_6.htm Check out the "Oracle Data Integrator Adapter for Hyperion xxx" section.
    -Chris
    P.S. Check out John Goodwin's blog -- there's some neat stuff in there and you'll probably figure out some neat stuff on your own too.

  • Source System from Oracle 8.1.5  to Netweaver 2004s BI

    Dear Friends,
    Can somebody tell me how to define a source system for bringing data from Oracle 8.1.5 to Netweaver 2004s BI.
    Thanks in advance.
    Regards
    Lalit Kumar Jain

    Use DB Connect..
    Fill the following parameters:
    1.<b>DB Connection</b> --Database name
    2.<b>DBMS</b>: ORA
    <b>3.user name:
    4.DBpassword:</b>
    5.<b>Conn Info:</b> Give oracle TNS name of the database to which you want to connect.
    6.<b>Permanent</b> --check the box..
    There are other parameters like conn limits..and ptimum conns.
    Regards
    Manga(Assign points)

  • Importing multiple amount columns from a single text file

    I'm sure this question has been addressed many times. I have tried to search for an answer here and other areas, but I have not been able to find a clear answer yet. I am relatively new to HFM and FDM and thus do not have a lot of experience to fall back on. I am primarily a Planning/Essbase person. That being said, here is my question:
    I have a data source (text file) containing two amount columns that I need to load to HFM via FDM. One amount column consists of Average Exchange Rates and the other amount column consists of Ending Exchange Rates. I have been asked to develop a process to load both columns of data to HFM using a single process (one Import Format). I've been told this is possible by writing an Import DataPump script. It seems that I would need to create a temporary record set based on the original source file and modify it so that it contained a duplicate set of records where the first set would be used for the Average Rate and the second set would be used for the Ending Rate. This would be a piece of cake using SQL against a relational source, but that's obviously not the case here. I do have some experience with writing FDM scripts but from an IF... Then... Else... standpoint based on metadata values.
    If there is anyone out there that has time to help me with this, it would be most appreciated.
    Thanks,

    This is relatively easy to achieve with a single import script associated with the Account source field (assuming AverageRate and EndRate are accounts in your application) in your import format.
    Essentially your first amount say AverageRate would be set as the default field for Amount and these values would be loaded as if it were a single value file. For the second value, EndRate you would have to insert the second value directly into the FDM work table which is the temporary table populated when data is imported from a file during the import process. The example code snippet below suld gve you guidance on how this is done
    'Get name of temp import work table
    strWorkTableName = RES.PstrWorkTable
    'Create temp table trial balance recordset
    Set rsAppend = DW.DataAccess.farsTable(strWorkTableName)
    If IsNumeric(EndRateFieldValue Ref Goes Here) Then
              If EndRateFieldValue Ref Goes Here <> 0 Then
                   ' Create a new record, and supply it with its field values
                   rsAppend.AddNew
                   rsAppend.Fields("DataView") = "YTD"
                   rsAppend.Fields("PartitionKey") = RES.PlngLocKey
                   rsAppend.Fields("CatKey") = RES.PlngCatKey
                   rsAppend.Fields("PeriodKey") = RES.PdtePerKey
                   rsAppend.Fields("CalcAcctType") = 9
                   rsAppend.Fields("Account") = "EndRate"
                   rsAppend.Fields("Amount") = EndRateFieldValue Ref
                   rsAppend.Fields("Entity")=DW.Utilities.fParseString(strRecord, 16, 1, ",")
                   rsAppend.Fields("UD1") = DW.Utilities.fParseString(strRecord, 16, 2, ",")
                   rsAppend.Fields("UD2") = DW.Utilities.fParseString(strRecord, 16, 3, ",")
                   rsAppend.Fields("UD3") = DW.Utilities.fParseString(strRecord, 16, 16, ",")
                   'Append the record to the collection
                   rsAppend.Update
              End If
    End If
    'Close recordset
    rsAppend.close
    In addition the return value of this Import Script should be "AverageRate" i.e. name of ht eaccount associated with the first value field. The NZP expression also needs to be put on the Amount field in the import format to ensure that the EndRate Field value is always processed even if the value of AverageRate is zero.

  • Retrieve varchar column from Oracle query in a resulset

    Hi, I'am a begginer in JSP Technologies.
    I've do a small jsp that opens a db connection to an oracle, execute a query a show results.
    When the columns of the query are int I've no problem, the jsp show the columns of the resulset, but when the columns selected are varchars the jsp doesn't show anything....
    Can anyone help me?
    Example:
    Table: CUSTOMERS
    Col1 : ID_COSTUMER int
    Col2: CUSTOMER_NAME varchar
    <%
    Connection canal = null;
    ResultSet tabla = null;
    Statement instruccion=null;
    try { Class.forName("oracle.jdbc.driver.OracleDriver");
    canal=DriverManager.getConnection("jdbc:oracle:thin:@XXXXXXXXXXXXXXXXXXXp");
    instruccion = canal.createStatement(ResultSet.TYPE_SCROLL_SENSITIVE, ResultSet.CONCUR_UPDATABLE);
    } catch(java.lang.ClassNotFoundException e){};
    String Query = "SELECT ID_CUSTOMER FROM CUSTOMERS";
    try { tabla = instruccion.executeQuery(Query);
    out.println("<TABLE Border=10 CellPadding=5><TR>");
    out.println("<TD>CUSTOMER</TD></TR>");
    while(tabla.next()) {
    out.println("<TR>");
    out.println("<TD>"+tabla.getString(1)+"</TD>");
    out.println("</TR>"); };
    out.println("</TABLE></CENTER></DIV></HTML>");
    tabla.close(); instruccion.close(); canal.close();}
    catch(SQLException e) {};
    %>Results:
    CUSTOMER
    1
    2
    3
    4
    Doing the change in query:
    SELECT CUSTOMER_NAME FROM CUSTOMERS
    I have not results....
    Thank you.

    sorry, I misplaced an ending code bracket on last one
    I'm not really familiar with doing this inside a jsp.. but there are a few things that you should try to make sure that it isn't a database problem before assuming it is a problem on your jsp.... It is possible you have already tried the following but just to make sure.
    When you do the second query does the first two out.printlns that are before the while still output? If not then your query is incorrect.
    Second I would try
    String Query = "SELECT ID_CUSTOMER,CUSTOMER_NAME FROM CUSTOMERS";
    try { tabla = instruccion.executeQuery(Query);
    out.println("<TABLE Border=10 CellPadding=5><TR>");
    out.println("<TD>CUSTOMER_ID</TD><TD>CUSTOMER_NAME</TD></TR>");
    while(tabla.next()) {
    out.println("<TR>");
    out.println("<TD>"+tabla.getString(1)+"</TD>");
    out.println("<TD>"+tabla.getString(2)+"</TD>");
    out.println("</TR>"); };
    out.println("</TABLE></CENTER></DIV></HTML>");I suspect this will also return nothing.. if that is the case then you need to check your database. If it does infact return values for the ID and nothing for the NAME then I'm not sure at this point what the problem is.

  • Multiple source files from multiple directories in sender CC

    Dear all,
    My requirrement is that ,I need to use multiple directories in sender CC (only one) and send the files to receivers.
    Kindly let me know how it can be done.
    Thanks,
    Srinivasa

    Hi,
    under advanced tab.............If you set the Enable indicator, you can enter variables for the Target Directory and File Name Scheme. Enter the names of the variables and references in the table.
    &#9679;      Enter each variable that you reference in the Target Directory and File Name Scheme fields without the surrounding percentage sign under Name of Variables in the table.
    The variables can refer to attributes of the message header or elements of the message payload.
    &#9675;       If the variables are to refer to an attribute of the message header, add the prefix message: to the name of the variable under Reference. You can specify the following attributes of the message header:
    sender_party, sender_service, receiver_party, receiver_service, interface_name, interface_namespace,
    message_id (message ID with hyphens, for example 9fbe1ff1-9a0d-11d9-8665-cbf10a126331)
    message_id_hex (message ID in hexadecimal format, for example 9fbe1ff19a0d11d98665cbf10a126331)
    For example, if you want to specify the interface name from the message header in the target directory or in the file name scheme, enter message:interface_name as the reference.
    If one of the message attributes contains characters that are not permitted in a file name, for example \, /, :, *, ?, ", <, >, |, then these characters are replaced by an underscore ("_").
    &#9675;       If the variable refers to an element in XML schema, add the prefix payload: to the information under Reference. The reference then comprises a pseudo path description in the form of a comma-separated list with the schema namea,na,nameb,nb,....
    namea,nameb,...corresponds to the element name and na,nb,... corresponds to the occurrence of the element name at the respective level in the document.
    The description begins at the root of the document and ends at the respective element.
    To reference the element that is in bold in the example, the following expression is used: payload:root,1,e1,1,e2,2
    The parser searches for the first occurrence of the root element at the first level. It then searches for the first occurrence of e1 at the second level and for the second occurrence of e2 at the third level. The content of the last element (&#8220;Example Value&#8221;) is set as the value for a specified variable.
    Thanks,
    Madhu

  • INTERFACING FROM ORACLE GL TO HYPERION FOR BUDGETING

    As of November 26, 1997, the HYPERION Interface to Oracle General Ledger is not supported.
    Someone know if is there a plan to accomplishing it in the future?
    Best Regards,
    Emilio Perez Solla

    As of the latest release version 9.3 they have interface for SAP, as now Hyperion is being acquired by Oracle you can definatly expect an interface for Oracle GL in version 9.5

  • SQL Load from Oracle into 64 bit Essbase 9.3.1 - issue

    I have a sql load that works fine in 32 bit Essbase 7.1.5
    Now I have setup the same sql load in a new 64 bit Essbase 9.3.1 environment. SQL runs. However when it retrieves it is not retrieving some fields "correctly". For example I have some extract functions in the SQL to extract the month or year from a date field. These are returning "0" now for the month or year. (the function syntax is correct as it works in the 7.1.5 environment. Plus for all my fields that are returning numeric dollar values, they are returning a value of "0" instead of the correct value (which it returns in the 7.1.5 environment).
    The sql works correctly if run in SQLPlus. So Essbase is having the issue.
    Any ideas?
    Edited by: Sean V on Jun 5, 2009 7:48 AM

    Hi Sean,
    If you can do it yourself or someone else who can do it for you,
    create a view in your db schema you are connected to and use a tool like Oracles SQLDeveloper (it's for free) or something similar to check your query result by select from your view.
    This way you can hide your internal logic behind this view and no one has to worry about it.
    From my POV this approach has no drawback but only advantage for you.
    If it works as expected put the view into your rules file (I do assume you are using rules files)
    Example (you need the CREATE VIEW privilege for you account(schema)):
    CREATE v_dataload as
    SELECT DIM_COL1 as year , DIM_COL2 as month, DIM_COL3 as market , DATA_COL1 as profit, DATA_COL2 as costs .... etc.,
    In you rules file you then just put in the SELECT box: year,month, profit, costs ...
    and in the table box: v_dataload
    If you name you columns like the Dimension (Gen1) you have the comfort that essbase automatically maps the columns tho the appropriate dimension. (But remember a colum in Oracle can have only up to 30 Characters)
    Back to you question:
    If you get a correct result from you view when using SQLDeveloper but not in you rules file I would say it's very likely a ODBC driver problem.
    If so, please check if you have installed the correct driver for your system. MERANT driver is the best choice as fare as I know.
    In normal cases you never should have to cast anything into a VARCHAR, since exactly this is the job of the ODBC driver.
    However, if everything fails and a cast (like Glen said) is the only way you can make it work put this cast into your view too.
    CREATE v_dataload as
    SELECT TO_CHAR(DIM_COL1) as year , TO_CHAR(DIM_COL2) as month, TO_CHAR(DIM_COL3) as market ,TO_CHAR( DATA_COL1) as profit, TO_CHAR(DATA_COL2) as costs .... etc.,
    Regards
    Andre
    Edited by: andreml on Dec 16, 2009 12:52 PM

  • Can anyone explain how the LO data source flows from ECC to BI up to cubes?

    Dear all,
    Can anyone explain step by step (HOW THE DATA FLOWS)....By taking 2LIS_11_VAHDR  from SD , starting from Activate Data source in ECC  to Info Cube in BI7.0. Please this will be helpful -
    Thanks for the advance answers...
    Edited by: harishk.225 on Dec 23, 2011 9:56 AM

    Hi Harish,
    First go to RSA5 in ECC  select your datasource 2LIS_11_VAHDR and activate it.
    After activatin check in RSA6 whether the datasoure activated properly or not.
    Then log on  to BI sytem Select datasource and click on replicate.
    Then Create Infocube infopackage ,Transformation, DTP ...entire flow. But dont schedule it because there is no data in SETUP Tables.For full load first we need to run statistical set up , so that data should come to SETUP Table.
    Now go to ECC  enter T.CODE OLI7BW to filled up the SETUP Table  for datasource 2LIS_11_VAHDR.
    It will ask  you for Run Name etc give the run name and time limit and execute it.
    If you get some error then first delete the SETUP Table by using Transaction Code LBWG.
    In LBWG it will ask you for applicatiion number , give 11 i.e for sales datasourc and execute it.
    Data will get deleted from SETUP Table. To check data is deleted or not  in SET UP Tabe go to database table MC11VA0HDRSETUP. Always remember the name of SETUP Table will be Extract Structure followed by SETUP.
    EX : If extract structure name is MC11VA0HDR then name of the SETUP Table will MC11VA0HDRSETUP.
    If data gets deleted then again go to 0LI7BW and run the Statistical Setup then agan check in MC1VAOHDRSETUP table contains data or not.
    Then trigger the Infopackage and DTP in BI.  Above steps was for full load.
    Now to Load deltas follow below steps.
    First Create INIT at BI side which enable delta at ECC Side.
    Then go to ECC and execute T.CODE LBWE.
    LBWE is LO Cockpit Work Bench. There we have folowng functions 1.Maintan Datasource. 2.Maintain Extract Structure. 3. Job  Control.  4.Delta Type 5.. Activate/deactivate.
    If you want to add any new field to your datasource then to to Maintain Extract Strucutere remeber before adding or modififying datasource first we should deactivate datasource by using 5th function  Activate/Deactivate.
    Then select DELTA TYPE in delta type we have four types of delta. 1. Delta Queue (RSA7) 2. Queued Delta (LBWQ)
    3. Unserialized (SM13) , 4. Seriaized (SM13)
    If you select Delta queue then deltas drectly come to RSAT.
    if you select Queued Delta  deltas goes to LBWQ then you have perform V3 job by using Job Control  function to move data from LBWQ to RSA7. Remeber to BI  data  always goes from RSA7 only.
    If you select Unserealized then delta goes to SM13 again you have to perorm V3 Job to move deltas in RSA7.
    Serialize update is not in used in Lo-cockpit.
    There is a differerence between  DELTA QUEUE , QUEUED DELTA, UNSERIALIZE UPDATE. Pls find the differences on net.
    Then select your delta type and create deltas dtp in BI and start loading .
    Hope this help you.
    In RSA7 we have two tables delta and repeat delta, to know the functionality of this please search for the documents on  net.
    Regards,
    Asim

  • Multiple report fields from single physical column.

    Hi Experts,
    I have a physical table like below. Multiple records for a single customer id. The no of Address_type may increase in future.
    Customer_id     Address_Type         Address
    101                  Primary                   abcdefgh
    101                  Secondary               234oij3mb4
    101                  Billing                      mdmnweoij4
    102................etc.
    My Requirement on OBIEE is to display report as below.
    Customer_ID, Primary_Address, Secondary_Address, Billing_Address.
    101                 abcdefgh            234oij3mb4               mdmnweoij4
    Multiple report fields ( Primary_Address, Secondary_Address, Billing_Address ) are derived from same physical column ADDRESS.
    I tried to duplicate the Address column on the report thrice and added conditions like Address_Type = Primary/Secondary/Billing for those three fields respectively.
    But only of the field is getting correct data and rest of duplicate columns are null.
    Tried FILTER function on OBI Answers, but the function fails saying it should be used only on measures.
    Thought the alternative would be to define multiple logical columns at BMM layer. Created multiple alias for same table at physical layer and correspondingly created LTS for each on a single logical table. Defined Fragmentation/Where conditions.
    LTS1 - Address_Type = PRIMARY
    LTS2 - Address_Type - SECONDARY
    LTS3 - Address_Type - BILLING
    Created multiple logical columns from each LTS like Primary_Address, Secondary_Address, Billing_Address from LTS1, LTS2, LTS3 respectively.
    This time, am getting data for all the fields on the report but getting cartesian result.
    Can someone guide me whether my approach is correct or any other alternate best approach to achieve the desired result.
    Note: I have the data Address on multiple physical rows for a customer, rather Not in a single row. which is causing me problem.
    Thanks...
    Ramchand.

    Hi Srini VEERAVALLI,
    I have lot many other fields in the same report coming from other tables.
    Does the Pivot table option for above mentioned Address fields affect other columns?
    If Yes, Can there be any other alternative? Am thinking of Opaque View at Physical layer to bring all Addresses on a single row.
    Thanks...
    Ramchand.

  • Best Practive - One mapping reading multiple source files

    I want to develop a solution for one single mapping reading multiple similar source files that are stored on different directories on my OWB server. I want to be able to determine on runtime of my mapping from what location to load the source file from.
    Example:
    Mapping: Load_test_data
    source file 1: c:\input\loc1\test.dat
    source file 1: c:\input\loc2\test.dat
    When I run the mapping I would like to use an input parameter specifying the location loc1 or loc2. I would also like to use this input parameter in my mapping to populate one column in my target table with the value of this input parameter. This design would make it possible to dynamically load source files from different directories and also being able to see after loading where the data came from.
    Questions:
    - Is there a way to create such a design
    - If not, what alternative would be appropriate.
    Thanks in advance for the feedback

    Thanks for the feedback. Unfortunately I do not use workflow together with my OWB.
    I now indeed specified the file name and file location in the configuration of my mapping. However I am not able to change then upon executing the mapping. Data file name and file location are empty and greyed out when I execute my mapping. It always takes the values I specified in the configuration of my mapping
    What I would like to do is specify the location upon runtime when I execute my mapping, but I don't know if this is possible. In addition I'd also want to use the data file location as an input parameter for one of the columns I populated in my target table.
    Then in the end I would be able to use one mapping and read multiple sources files from different locations and also be able to check in the end where the data was loaded from.
    Hope you can give me some more feedback on how to set this up in OWB.
    Many Thanks!
    data file name parameter to ma

  • Insert into MSSQL from Oracle

    Hello together,
    I'm a little bit frustrated. The follwoing is my problem:
    I want to insert some values from an Oracle 10gR2 database into a MSSQL database (SQL Server 2008 R2).
    Therefor I'm using DG4ODBC 11.2.0.3. The connection is working absolutly fine.
    I want to insert five values from five columns from oracle.table-a into a msssql table-b.
    But I'm hitting the error:
    Invalid character value for cast specification.
    Two columns from the oracle table are chars size 12 and 40, and three are number size 5.1, 5.1 and 4.
    The MSSQL column definitions are varchar40, varchar50, 2x float8 and int.
    The column assignments are:
    Oralce -> MSSQL
    a. char 12 -> varchar40
    b. char 40 -> varchar50
    c. number 5.1 -> float8
    d. number 5.1 -> float8
    e. number 4 -> int4
    The insert works greate for a,b and e but nor for c and d.
    I tried so many solutions (cast, to_binary_float, nls_character_format, nls_theritpory etc..) but nothing helps.
    I always hitting this error.
    Can anyone help me with this problem?
    cheers
    Joe

    Hi together,
    I fixed the problem. I multiply the origin number with 10. The result is a number without any comma. During the insert into MSSql I divide by 10.
    I now, this is a very strange solution but it works.
    Yes I know this helps only, because the column can only save number with one decimal place. For higher decimal places you have to multiple it with 100 or 1000 or......
    Hope this helps also other peoples.
    kind regards
    Joe

  • Problem after upgradation from Oracle 8.1.7 to Oracle 9i

    Hello ,
    I currently installed Oracle 9i, and i m using Mapx 4.5,
    and in the executable I am adding serverlayer
    (a layer formed by querying the geoloc column from oracle
    table), which was working fine with Oracle 8.1.7, but after
    installation of Oracle 9i , i encounter the following error
    while adding a serverlayer
    ORA-13373 Element of type Extent is not supported for Geodetic
    data
    Cause: Element type extent for a polygon geometry is not
    supported for
    geodetic data.
    Action: Convert the extent type polygon to a regular 5 point
    polygon
    and set
    etype accordingly.
    ORA-06512 at string line string
    Cause: Backtrace message as the stack is unwound by unhandled
    exceptions.
    Action: Fix the problem causing the exception or write an
    exception
    handler
    for this condition. Or you may need to contact your application
    administrator or database administrator
    Can anybody tell me what's going wrong ?
    As I urgently got to fix it.
    Thanks in advance
    Binoy.

    Hi,
    Oracle9i has a geodetic data model. Your data has a geodetic
    srid, which means that the data will need to conform to the rules
    associated with geodetic data. These rules include:
    No optimized rectangles are supported (my guess is the app is
    using an optimized rectangle as a query window, but the data
    could be the issue also).
    No arcs or circle types are supported.
    A single line segment cannot be equal to or greater than 1/2 the
    great circle distance.
    A polygon cannot have an area equal to or greater than 1/2 the
    surface area of the earth.
    You should try to find what version of Mapx supports Oracle9i.
    Regards,
    dan

  • Error : Data load from relation source (SQL Server 2005) to Essbase Cube

    Hi All,
    I am looking help from you. I am trying to load data from SQLServer 2005 table to Essbase Cube using IKM SQL to Hyperion Essbase (Metadata) Modules.
    I am getting below error. Let me know if i am missing some things.
    org.apache.bsf.BSFException: exception from Jython:
    Traceback (innermost last):
    File "<string>", line 61, in ?
    com.hyperion.odi.essbase.ODIEssbaseException: Invalid value specified [RULES_FILE] for Load option [null]
         at com.hyperion.odi.essbase.ODIEssbaseMetaWriter.validateLoadOptions(Unknown Source)
         at com.hyperion.odi.essbase.AbstractEssbaseWriter.beginLoad(Unknown Source)
         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
         at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
         at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
         at java.lang.reflect.Method.invoke(Unknown Source)
         at org.python.core.PyReflectedFunction.__call__(PyReflectedFunction.java)
         at org.python.core.PyMethod.__call__(PyMethod.java)
         at org.python.core.PyObject.__call__(PyObject.java)
         at org.python.core.PyInstance.invoke(PyInstance.java)
         at org.python.pycode._pyx3.f$0(<string>:61)
         at org.python.pycode._pyx3.call_function(<string>)
         at org.python.core.PyTableCode.call(PyTableCode.java)
         at org.python.core.PyCode.call(PyCode.java)
         at org.python.core.Py.runCode(Py.java)
         at org.python.core.Py.exec(Py.java)
         at org.python.util.PythonInterpreter.exec(PythonInterpreter.java)
         at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:144)
         at com.sunopsis.dwg.codeinterpretor.k.a(k.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSqlI.treatTaskTrt(SnpSessTaskSqlI.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java)
         at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand(DwgCommandSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandBase.execute(DwgCommandBase.java)
         at com.sunopsis.dwg.cmd.e.i(e.java)
         at com.sunopsis.dwg.cmd.g.y(g.java)
         at com.sunopsis.dwg.cmd.e.run(e.java)
         at java.lang.Thread.run(Unknown Source)
    com.hyperion.odi.essbase.ODIEssbaseException: com.hyperion.odi.essbase.ODIEssbaseException: Invalid value specified [RULES_FILE] for Load option [null]
         at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:146)
         at com.sunopsis.dwg.codeinterpretor.k.a(k.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSqlI.treatTaskTrt(SnpSessTaskSqlI.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java)
         at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand(DwgCommandSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandBase.execute(DwgCommandBase.java)
         at com.sunopsis.dwg.cmd.e.i(e.java)
         at com.sunopsis.dwg.cmd.g.y(g.java)
         at com.sunopsis.dwg.cmd.e.run(e.java)
         at java.lang.Thread.run(Unknown Source)

    ODI Step: Prepare for Loading Step:
    from java.util import HashMap
    from java.lang import Boolean
    from java.lang import Integer
    from com.hyperion.odi.common import ODIConstants
    from com.hyperion.odi.connection import HypAppConnectionFactory
    # Target planning connection properties
    serverName = "HCDCD-HYPDB01"
    userName = "admin"
    password = "<@=snpRef.getInfo("DEST_PASS") @>"
    application = "BUDGET01"
    database = "PLAN1"
    portStr = "1423"
    srvportParts = serverName.split(':',2)
    srvStr = srvportParts[0]
    if(len(srvportParts) > 1):
    portStr = srvportParts[1]
    # Put the connection properites and initialize the essbase loader
    targetProps = HashMap()
    targetProps.put(ODIConstants.SERVER,srvStr)
    targetProps.put(ODIConstants.PORT,portStr)
    targetProps.put(ODIConstants.USER,userName)
    targetProps.put(ODIConstants.PASSWORD,password)
    targetProps.put(ODIConstants.APPLICATION_NAME,application)
    targetProps.put(ODIConstants.DATABASE_NAME,database)
    targetProps.put(ODIConstants.WRITER_TYPE,ODIConstants.DATA_WRITER)
    print "Initalizing the essbase wrapper and connecting"
    pWriter = HypAppConnectionFactory.getAppWriter(HypAppConnectionFactory.APP_ESSBASE, targetProps);
    tableName = "BUDGET01_PLAN1"
    rulesFile = r"ActLd"
    ruleSeparator = "Tab"
    clearDatabase = "None"
    calcScript = r""
    maxErrors = 1
    logErrors = 1
    errFileName = r"E:\OraHome_ODI\Error\Budget01Plan1.err"
    logEnabled = 1
    logFileName = r"E:\OraHome_ODI\Error\Budget01Plan1.log"
    errColDelimiter = r","
    errRowDelimiter = r"\r\n"
    errTextDelimiter = r"'"
    logHeader = 1
    commitInterval = 1000
    calcOnly = 0
    preMaxlScript = r""
    postMaxlScript = r""
    abortOnPreMaxlError = 1
    # set the load options
    loadOptions = HashMap()
    loadOptions.put(ODIConstants.CLEAR_DATABASE, clearDatabase)
    loadOptions.put(ODIConstants.CALCULATION_SCRIPT, calcScript)
    loadOptions.put(ODIConstants.RULES_FILE, rulesFile)
    loadOptions.put(ODIConstants.LOG_ENABLED, Boolean(logEnabled))
    loadOptions.put(ODIConstants.LOG_FILE_NAME, logFileName)
    loadOptions.put(ODIConstants.MAXIMUM_ERRORS_ALLOWED, Integer(maxErrors))
    loadOptions.put(ODIConstants.LOG_ERRORS, Boolean(logErrors))
    loadOptions.put(ODIConstants.ERROR_LOG_FILENAME, errFileName)
    loadOptions.put(ODIConstants.RULE_SEPARATOR, ruleSeparator)
    loadOptions.put(ODIConstants.ERR_COL_DELIMITER, errColDelimiter)
    loadOptions.put(ODIConstants.ERR_ROW_DELIMITER, errRowDelimiter)
    loadOptions.put(ODIConstants.ERR_TEXT_DELIMITER, errTextDelimiter)
    loadOptions.put(ODIConstants.ERR_LOG_HEADER_ROW, Boolean(logHeader))
    loadOptions.put(ODIConstants.COMMIT_INTERVAL, Integer(commitInterval))
    loadOptions.put(ODIConstants.RUN_CALC_SCRIPT_ONLY,Boolean(calcOnly))
    loadOptions.put(ODIConstants.PRE_LOAD_MAXL_SCRIPT,preMaxlScript)
    loadOptions.put(ODIConstants.POST_LOAD_MAXL_SCRIPT,postMaxlScript)
    loadOptions.put(ODIConstants.ABORT_ON_PRE_MAXL_ERROR,Boolean(abortOnPreMaxlError))
    #call begin load
    pWriter.beginLoad(loadOptions)
    Execution step from opartor:
    Read rows: 0
    Insert/Delete/updat rows: 0
    ODI Step: Load Data Into Essbase
    from com.hyperion.odi.common import ODIConstants
    from com.hyperion.odi.connection import HypAppConnectionFactory
    from java.lang import Class
    from java.lang import Boolean
    from java.sql import *
    from java.util import HashMap
    # Get the select statement on the staging area:
    sql= """select C1_ACCOUNT "Account",C3_TIMEPERIOD "TimePeriod",C4_LOBS "LOBs",C5_TREATY "Treaty",C6_SCENARIO "Scenario",C7_VERSION "Version",C8_CURRENCY "Currency",C9_YEAR "Year",C10_DEPARTMENT "Department",C11_ENTITY "Entity",C2_DIVLOC "DivLoc",C12_DATA "Data" from OdiMapping.dbo.C$_0BUDGET01_PLAN1Data where      (1=1) """
    srcCx = odiRef.getJDBCConnection("SRC")
    stmt = srcCx.createStatement()
    srcFetchSize=30
    stmt.setFetchSize(srcFetchSize)
    rs = stmt.executeQuery(sql)
    #load the data
    stats = pWriter.loadData(rs)
    #close the database result set, connection
    rs.close()
    stmt.close()
    ODI Step: Report Statistics
    org.apache.bsf.BSFException: exception from Jython:
    Traceback (innermost last):
    File "<string>", line 2, in ?
    Essbase Writer Load Summary:
         Number of rows successfully processed: 1
         Number of rows rejected: 0
         at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:146)
         at com.sunopsis.dwg.codeinterpretor.k.a(k.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSqlI.treatTaskTrt(SnpSessTaskSqlI.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java)
         at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand(DwgCommandSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandBase.execute(DwgCommandBase.java)
         at com.sunopsis.dwg.cmd.e.i(e.java)
         at com.sunopsis.dwg.cmd.h.y(h.java)
         at com.sunopsis.dwg.cmd.e.run(e.java)
         at java.lang.Thread.run(Unknown Source)
    Where am i doing wronge for data load in Essbase? How I will find correct information to load data into Essbase?

Maybe you are looking for

  • My images are not showing up on iDvd

    Recently tried cleaning out my start up disk. Opened iDVD and now I can't find my images to drag and create a slideshow. Can anyone help?? Please

  • Change Font in Master pages so that it changes the rest of the template?

    I cannot change my font in the Master pages so that it effect the rest of the template. I would have to go through each of the page and change the font. I just want the body paragraph a certain font and I don't know how to do that. I don't know if I

  • SAP-Script; Is it possible to handle the BOX -Statement dynamic?

    Hi Scripters! My consultant wants a box around some information in a SAP-Script form. Normally this is quite easy, BUT in this case i have the information (which needs the box around) printed after the positions. So problem is: the box can not have a

  • How to display Success Message in PDF

    Hello all I need to display a success messgae in ODF. but without he message box. It should be such that, after the success, the user should not click yet see the success message. I got only MEssageBox. But that gives a pop up. Please let me know abo

  • Paying balance before or after closing date?

    Hello, I'm trying to rebuild my credit after destroying it from bad decisions when I was 18. This forum has definitely been a huge help! I wanted to get clarity on when to pay the full balance on my CC. I've seen a lot of post regarding letting at le