Missing Standard Dimension Column for data load (MSSQL to Essbase Data)

This is similar error to one posted by Sravan -- however I'm sure I have all dimensions covered -- going from MS SQL to SunOpsys Staging to Essbase. It is telling me missing standard dimension, however I have all accounted for:
org.apache.bsf.BSFException: exception from Jython:
Traceback (innermost last): File "<string>", line 23, in ? com.hyperion.odi.essbase.ODIEssbaseException: Missing standard dimension column for data load
at com.hyperion.odi.essbase.ODIEssbaseDataWriter.loadData(Unknown Source)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
at java.lang.reflect.Method.invoke(Unknown Source)
I'm using multiple time period inputs -- BegBalance,Jul,Aug,Sep,Oct,Nov,Dec,Jan,Feb,Mar,Apr,May,Jun (target has all of those in place of Time Periods)
I'm using hard coded input mapping for Metric, Scenario, Version, HSP_Rates and Currencies. -> 'Amount', 'Actual', 'Final', 'HSP_InputValue','Local' respectively.
The only thing I can think of is that since I'm loading to each of the months in the Time Periods dimension (the reversal was set up to accomodate that)... and now its somehow still looking for that? Time Periods as a dimension does not show up in the reversal -- only the individual months named above.
Any ideas on this one??

John -- I extracted the data to a file and created a data load rule in Essbase to load the data. All dimensions present and accounted for (five header items as similar here) and everything loads fine.
So not sure what else is wrong -- still getting the missing dimension error.
Any other thoughts?? Here's the entire error message. Thanks for all your help on this.
org.apache.bsf.BSFException: exception from Jython:
Traceback (innermost last):
File "<string>", line 23, in ?
com.hyperion.odi.essbase.ODIEssbaseException: Missing standard dimension column for data load
     at com.hyperion.odi.essbase.ODIEssbaseDataWriter.loadData(Unknown Source)
     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
     at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
     at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
     at java.lang.reflect.Method.invoke(Unknown Source)
     at org.python.core.PyReflectedFunction.__call__(PyReflectedFunction.java)
     at org.python.core.PyMethod.__call__(PyMethod.java)
     at org.python.core.PyObject.__call__(PyObject.java)
     at org.python.core.PyInstance.invoke(PyInstance.java)
     at org.python.pycode._pyx8.f$0(<string>:23)
     at org.python.pycode._pyx8.call_function(<string>)
     at org.python.core.PyTableCode.call(PyTableCode.java)
     at org.python.core.PyCode.call(PyCode.java)
     at org.python.core.Py.runCode(Py.java)
     at org.python.core.Py.exec(Py.java)
     at org.python.util.PythonInterpreter.exec(PythonInterpreter.java)
     at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:144)
     at com.sunopsis.dwg.codeinterpretor.k.a(k.java)
     at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting(SnpSessTaskSql.java)
     at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
     at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
     at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt(SnpSessTaskSql.java)
     at com.sunopsis.dwg.dbobj.SnpSessTaskSqlI.treatTaskTrt(SnpSessTaskSqlI.java)
     at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java)
     at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java)
     at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java)
     at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand(DwgCommandSession.java)
     at com.sunopsis.dwg.cmd.DwgCommandBase.execute(DwgCommandBase.java)
     at com.sunopsis.dwg.cmd.e.k(e.java)
     at com.sunopsis.dwg.cmd.g.A(g.java)
     at com.sunopsis.dwg.cmd.e.run(e.java)
     at java.lang.Thread.run(Unknown Source)
Caused by: com.hyperion.odi.essbase.ODIEssbaseException: Missing standard dimension column for data load
     at com.hyperion.odi.essbase.ODIEssbaseDataWriter.validateColumns(Unknown Source)
     ... 32 more
com.hyperion.odi.essbase.ODIEssbaseException: com.hyperion.odi.essbase.ODIEssbaseException: Missing standard dimension column for data load
     at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:146)
     at com.sunopsis.dwg.codeinterpretor.k.a(k.java)
     at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting(SnpSessTaskSql.java)
     at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
     at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
     at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt(SnpSessTaskSql.java)
     at com.sunopsis.dwg.dbobj.SnpSessTaskSqlI.treatTaskTrt(SnpSessTaskSqlI.java)
     at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java)
     at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java)
     at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java)
     at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand(DwgCommandSession.java)
     at com.sunopsis.dwg.cmd.DwgCommandBase.execute(DwgCommandBase.java)
     at com.sunopsis.dwg.cmd.e.k(e.java)
     at com.sunopsis.dwg.cmd.g.A(g.java)
     at com.sunopsis.dwg.cmd.e.run(e.java)
     at java.lang.Thread.run(Unknown Source)

Similar Messages

  • Data load failures in Essbase Clusters ( MSCS)

    Hi,
    If there is a failure on the Active Essbase Cluster Node (call this Node A) and the Cube has to be rebuilt on Cluster Node B, how will the Cube be re-built on Cluster Node B.
    what will orchestrate the required activities in the correct order to rebuild the Cube)? Both Essbase nodes are mounted on to Microsoft Clustered Services.
    In essential, I want to know
          A)  How to handle the metadata load that failed on Node1 onto Essbase Node2?
          B)  Does the session continues to perform the meta data/ data load on the Essbase Second Node when the Essbase First Node fails?
    Thanks for your help in advance.
    Regards,
    UB.

    It would be built the same way as you either use the Essbase clustername or the MSCS VIP used for the Essbase cluster configuration,  you would not ever be referencing a host name so it does not matter which node it is running on.
    The session will be lost on a failover as it is the same as an essbase restart.
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Record count is different for Full load and Init with data transfer

    Hi all,
    We have a data source for which delta is enabled on Calendar Day.
    If I run the full load record count is 2670.But If I run the delta record count is 2665.5 records are missing.
    Would you please help on this.Why those records are missing.

    Hi,
    Actually for Full Load say you have 50 records. When you run it will upload 50 records.
    Now if you run delta if there is no records changes it wouldnt update because there is no update in the existing records.
    only delta is done when there is changes in the records.

  • Data Load behaviour in Essbase

    Hello all-
    I am loading data from Flat File using a server Rule File. In the rule file i have properties for a feild where in it replaces a name in flat file for member name in outline so it is somwhat like this:
    Replace With
    Canada 00-200-SE
    Belgium 00-300- SE
    and so on
    Now in my flat file there was a new member for example china & the replacement for it was not present in Rule File & when the data was loaded in the system it didnt rejected that record on the contrary it loaded the values for china in
    the region which was above it and overwrited the values for the original one.
    Is this the normal behavior of essbase , I was thinking that record should have been rejected .
    I know when we do a Lock & Send via Addin & if member is not present in outline it give you warning when you lock that sheet & eventually if you dont delete that member from the template it will load data against it in the member above it.
    Is there a waok around for this problem or this is what it is ?
    I am on Hyperion Planning / Essbase Version 9.3.1.
    Thanks

    Still thinking how does these properties effects the way data is being loaded right now. Have gone through DBAG & i dont see a reason y any of these peoperties might be affecting the load^^^Here's what I think is happening: China is not getting mapped, but the replacement for Belgium is occuring and resolves to a valid member name. Essbase sees China and doesn't recognize it (you knew all of this already).
    When the load occurs, Essbase says (okay, I am anthromorphizing, but you get the ida) "Eh, I have no idea what China is, but 00-300-SE is the last good Country member I have, I will load there." Essbase is picking the last valid member and loading to that. I liken it to a lock and send from Excel with nested dimensions and non-repeating members. Essbase "looks up" a row, finds the valid member, and loads there.
    And yes, this is in the DBAG: http://download.oracle.com/docs/cd/E12825_01/epm.111/esb_dbag/ddlload.htm#ddlload1034271
    Search for "Unknown Member Fields" -- it's all the way at the bottom of the above link.
    In fact, to save you the trip, per the DBAG:
    If you are performing a data load and Essbase encounters an unknown member name, Essbase rejects the entire record. If there is a prior record with a member name for the missing member field, Essbase continues to the next record. If there is no prior record, the data load stops. Regards,
    Cameron Lackpour

  • Data load error in essbase studio

    I get the following error when trying to load an ASO cube using Essbase Studio (EPM 11.1.2). This error doen't seem to be documented in any of the Essbase manuals. Question - does this error indicate an essbase server issue or a data source issue? I'm thnking it's datasource related, but my data source is an Oracle database, which I've used previously to load cubes without a problem. I've refreshed the source and can connect to it fine otherwise.
    Error:
    Data load started at: Fri Dec 03 08:52:21 EST 2010.      Data load elapsed time:  10 Minutes 23 Seconds.
    Failed to deploy Essbase cube.
    Caused by: Failed to load data into database: 8020.
    Caused by: Cannot execute a SQL query
    Caused by: Io exception: Socket read timed out
    Caused by: Socket read timed out
    Appreciate any hel with this issue.

    When I have issues with Studio I try to break it down slowly. I build my dimensions one at a time. If it breaks on a single dimension build I trace the issues backwards and usually find my issue in the schema.
    Studio's role in life is to create SQL load rules and as such depend on a good schema definition. Unforntunately, the dimension build rules can't be opened in EAS with the Dataprep Editor (regular load rules) because they're binary and can do things that a normal load rule cannot (text measures, date measures, time varying attributes, etc.). But that doesn't mean the .rul files are un-readable. If you're having trouble with a particular dimension build process, open the load rule it creates with something like Notepad and grab the SQL that Studio is generating and drop it into Toad (or equivalant) to see if it is generating usable code. If not, there's something wrong with your modeling and you need to go back to the mini-schema.
    When you're able to build all dimensions all at the same time, you're almost there. If your issues comes when you want to build and load data, the final debuging steps go quickly. Towards that end, the data load rules (ones that load data vs. building dimensions) generated by Studio can be edited in EAS using the Dataprep Editor. If you know SQL Load Rules, you should be able to figure out. If not, contact John Goodwin, OCS or a partner and set up a consulting visit.

  • Flat file data load - ODS - Look up data in Startroutine

    Hi All,
    There is a requirement that I have two ODS say 1) ABC and XYZ.
    For Both ODS , We load flat file data.
    First we load data to ABC for current fiscal period
    During the data load to XYZ, We lookup part nos data in ABC for current fiscal period , if data there , for those part nos we load XYZ Flat file data.
    My requirement is that , I need to fetch part nos data from ABC for Fiscal period < current fisacl period.  Then I need to extract already loaded data for those part nos from XYZ and Mark the flag field as Non reportable .
    How can we achieve this ?
    Please advice
    Thanks
    Ajay

    Hi ,
    Thanks for your reply.
    I have done so. When I add lines of My Internal table data to DATA_PACKAGE, It gives the syntax error that both structures are not unique. We have several fields in XYZ ods but they are not in Communication structure.
    Thats the problem. Also , if I write the code , will it be executed for each datapackage. I mean , for each datapacket process , my code will fetch whole data from ABC and then from XYZ. Repetion will be there?
    Any other logic can i use ?
    Thanks
    Ajay

  • Log on data load through a BW data flow

    Dears,
    I am requesting to all of you who have already implemented this type of functionality. I am trying to find the easiet way, with less complexity, to implement a log through an existing BW data flow.
    I mean data load by an infopackage give some log on right and wrong records within the monitor, how can I utilize this information? is there a specific table which stored each record and their message? Or a program has to be implemented which will publish laoding status in a specific table?
    Thanks for your quick feedback,
    LL

    Hi Ludovic
    The monitor messages are only written if there is some problem in the record processing. You can only find information for those records which have problem or if the processing during the routines encountered some problem.
    What you can do to capture messages is write one transfer routine and amend the monitor messages table rsmonmess for the same.
    Also,please check the tables starting with RSMO*
    regards
    Vishal

  • BW data loading to cube (delete data before load)

    I have a cube which load data from a SQL based source system.
    I would like to have a schedule job which first delete current year's data in the cube, then extract and load current year's data from SQL server to cube.
    How can it be done? Thanks.

    Hi,
    If you are using Process chains then use...
    Use DELETE_FACTS tcode and genarate the report name i.e. select Data target and Genarate Program then execute it will genarate the program, so create varient for that program and then use it in process chain. i.e.
    Start
      |
    Delete using that Program (with varient on year/month)
      |
    Load data
    Thanks
    Reddy

  • Transactional Data Load without Respective Master Data in Place

    Hi Folks,
    Could anyone let me know if i could load transactional data without loading the respective master data. If yes, well could you give me a scenario when you do it. Also could you let me know the option of how to load transactional data without master data in place.
    Best Regards,
    Prithvi

    Hi,
    You can load transaction data without loading master data.
    You have the option in the info package to load data even if no master data exists..
    You may use it in cases where your transfer rules/update rules or routines donot refer to master data tables to populate transaction data..
    Ashish

  • Full load - Every month data loading without deleting previous data

    Dear SDN,
    I am loading (Full load) Data from 001.2005 to 010.2007 using 0FI_GL_1 with the selection option Fiscal Year/Period (From : 001.2005 and To : 010.2007) ...
    From next month (011.2007) onwards, when loading data everytime it deletes entire data and again loading (Since i have kept an option in Infopackage -- Datatargets Tab -- Delete entire content of datatarget)...
    My doubt is wheather without deleting the whole data, can we load from next month data onwards...so, if that is the case shall i use Fiscal year/Period range ?? (or) should not use Fiscal year/Period range...
    But the client needs data only from 001.2005 onwards not beyond that
    Please help me to resolve this issue..
    Help will be greatly appreciated with points...
    Thanks....

    Hello Venkat,
    Unfortunately you can not do that way, because in period 11.2007 (assume this is the period that you want to newly load to cube.) there might be posting in previous period. And these postings made to previous period should also reflect to cube. Only way is delete and reload by using full.
    Other option you might use a ODS before sending to cube. you might every time load full data, ODS key fields might be checked by system and neccesary inserts and updates will be done by system, so there wont be duplications. And finally only new data will be sent to cube as delta.
    Sarhan.

  • Data load errors in Essbase

    I am attempting to load our data pulled from our HFM application into Essbase. I keep getting "data value encountered before all dimensions selected" error. I have checked the load file line by line and the rule everything lines up correctly from the file to the rule. We have recently migrated our Planning application to 3.3.1 of Planning and Essbase 6.5.4. We have loaded data before in this new system but for some reason I can't get this to load. Any ideas or help would be greatly appreciated. thanks,Eleasha

    Is every dimension in the OUTLINE represented in the data file? If not, do the unrepresented dimensions have a member specified in the headers?If every dimension is represented, do the data fields come AFTER the member fields?

  • Data load issue with export data source - BW 3.5

    Hi,
    We are facing issues in loading data with the help of export data source.
    We have created export data source of 0PCA_C01 cube. With the help of this export datasource,  we are loading data to other custom cube. Scenario is working fine in development server.
    But when we transported objects to quality server data is not getting loaded to custom target cube.
    It is extracting zero records.  All transports are ok and we have generated export datasource in quality before transports .Also regenerated export datasource after transport and activated infosource, update rule via RS* programs.  Every object is active but data is not getting extracted.
    RSA3 for 80PCA_C01 datasource isn't extracting any record in Quality. Records getting extracted in development.   We are in BW 3.5 with patch level 19.
    Please guide us to resolve the issue.
    Thanks,
    Aditya

    Hi
    Make sure that you have relevant Role & Authorization at Quality/PRS.
    You have to Transport the Source Cube first and then Create a Generate Export Data Source in QAS. Then, replicate data sources for BW QAS Soruce System. Make sure this replicated Data Source in QAS. Only then can transport new update rules for second cube.
    Hope it helps and clear

  • Data Load error in Essbase

    Hello all-
    I extracted some data on flat file from my application in production using report script and i am trying to push the data in another essbase cube using rule file and it comes up with following error:
    Reading Rule SQL Information For Database [Database name]
    Reading Rules From Rule Object For Database [Database name]
    Parallel dataload enabled: [1] block prepare threads, [1] block write threads.
    Loading Unicode-encoded data to non-Unicode mode application is not supported
    Unexpected Essbase error 1003048
    I have previously pushed the level 0 data export many times and this time i am trying to push it via rule file and it come up with this error. How can i look if my orginal application is a unicode application or not??? The rule file that i am using is also from the source application .. I tried loading the data in the same source application but different database and it came up up with same error... Anybody have come across this error before?
    Please advise!

    *Error “Loading Unicode Encoded Data to Non Unicode Mode Application is not Supported” When Importing An Export File. [ID 764447.1]*
    Symptoms
    When doing an export with a Report Script the export file is being saved as an UTF-8 document.
    Subsequently when importing the export file back into the database the following error occurs:
    “Loading Unicode encoded data to Non Unicode mode application is not supported”.
    The file has to be saved as an ANSI file for the import to work.
    Cause
    When exporting data from the EAS console, the export file is created as UTF-8 encoded.
    As such Essbase cannot read this file when importing it back in.
    This is Bug 6763776 and has been fixed in version 11.1.1.
    Solution
    This issue has been fixed in version 11.1.1.
    If you are unable to upgrade at this time then the suggested workaround is to use MaxL or Esscmd to run the report script, as running it in this manner will save the file in non-unicode mode and thus allow you to load it back in without errors.
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • How to..Show a wait dialog when data loads, then hide once data loads

    Description: This code will have the SAP default wait dialog become visible when the page loads and hidden after the data is returned for the table.
    1. Place this code within the head tags.
    <head>
    <SCRIPT Language="JavaScript">
    function show_wait_dialog () {
    document.body.style.cursor = "wait";
    document.getElementById("SAPBWProcessBoxSpan").style.setAttribute("display", "block", false);
    document.getElementById("SAPBWProcessBoxSpan").style.setAttribute("visibility", "visible", false);
    document.getElementById("SAPBWProcessBoxSpan").style.setAttribute("overflow", "auto", false);
    document.getElementById("SAPBWProcessBoxSpan").style.setAttribute("width", "255", false);
    document.getElementById("SAPBWProcessBoxSpan").style.setAttribute("left", document.body.offsetWidth/2-125, false); document.getElementById("SAPBWProcessBoxSpan").style.setAttribute("top",
    document.body.offsetHeight/2-38, false);
    function hide_wait_dialog () {
    document.body.style.cursor = "auto";
    document.getElementById("SAPBWProcessBoxSpan").style.setAttribute("display", "block", false);
    document.getElementById("SAPBWProcessBoxSpan").style.setAttribute("visibility", "hidden", false);
    document.getElementById("SAPBWProcessBoxSpan").style.setAttribute("overflow", "auto", false);
    document.getElementById("SAPBWProcessBoxSpan").style.setAttribute("width", "255", false);
    document.getElementById("SAPBWProcessBoxSpan").style.setAttribute("left", document.body.offsetWidth/2-125, false); document.getElementById("SAPBWProcessBoxSpan").style.setAttribute("top", document.body.offsetHeight/2-38, false);
    </script>
    </head>
    2. Place this code right after the <body> tag.
    <body>
    <!-- New Code -->
    <script>
    show_wait_dialog()
    </script>
    3. Lastly, place this code below right after the Table Item and before the </body> tag. (in this case GR1Table)
    <object>
             <param name="OWNER" value="SAP_BW"/>
             <param name="CMD" value="GET_ITEM"/>
             <param name="NAME" value="GR1Table"/>
             <param name="ITEM_CLASS" value="CL_RSR_WWW_ITEM_GRID"/>
             <param name="DATA_PROVIDER" value="DP"/>
             <param name="WIDTH" value="700"/>
             <param name="BLOCK_SIZE" value="3000"/>
             <param name="SHOW_PAGING_AREA_TOP" value="X"/>
             <param name="HELP_SERVICE" value="ZPRINTING"/>
             <param name="HELP_SERVICE_CLASS" value="Z_PRINTING_HELP_SERVICE"/>
             ITEM:            GR1Table
    </object></P>
    <P>
    <!—New code --&#61664;
    <script>
    hide_wait_dialog ()
    </script>
    </body>

    This is very helpful. Thanks.

  • Data Load using only 1 data package

    Hi,
    I am moving data from one cube to another manually using an infopackage. This system has recently been upgraded from 3.0 to 3.5.
    It is only using 1 datapackage (I am loading 19m records) and have done this successfully before and it used 4 datapackages.
    Settings in SBIW show that MAXPROC for myself system is set to 4. Entries in ROOSPRMS show MAXPROC as being 1.
    With these settings in the failover system (3.0B), it uses 3 datapackages. With these settings in teh production system (3.5), it uses 1 datapackage.
    Does anyone have any ideas as to how to get it to use 4 datapackages ?
    regards
    Sue

    Check Default Datasource data transefer in the scheduler menu item in the Infopacakge; here you can find max number of records and max size of datapackage.
    Hope it helps.
    Regards

Maybe you are looking for

  • How to reload purchased apps no longer in app store?

    I have an very old ipad 2. Over time, I purchased and loaded many apps from app store.  It have never been sync with an itunes.  Now I just got the new ipad air 2.  Both ipads runs iOS 8.1.  When I try to reload the apps from the "purchased" section

  • WRT300N dropped wireless completely

    I've had this WRT300N for several years and it has worked very well. It suddenly dropped all wireless connections and no one can re-connect.  My main pc is hard wired directly to the router and works perfectly as always.  3 wireless pc's and 3 handhe

  • When renting movies do you need the internet to rewatch?

    I want to rent a few movies to watch on a plane ride. Will I be able to watch the movies that I rent without an interent connection?

  • Where does Mountain Lion store iCloud files?

    Mountain Lion knows a bunch of iCloud-enabled apps. For back-up considerations, where on my HDD can I find the files, stored in iCloud, for programs like Pages, Preview and TextEdit ? I'm sure they must be cached locally somewhere.

  • Regarding Upgradation From 4.6 C To ECC 6.0-Urgent

    Hi all can anybody let me know what is procedure to upgrade the SAP From 4.6 C To ECC 6.0 from MM and CIN point of view --is Up gradation is just like an implementation or Enhancement process any help would be appreciable thanks in Advance Edited by: