EIS 6.2 Data load

We are upgrading to EIS 6.2. I've set up an outline that mirrors our production outline and can perform a member load without errors. The data load also runs without errors, the problem is no data gets loaded. I have cut and paste the SQL query the EIS generates from the log file and used it the query the Oracle database. It returns valid data. Has anyone had this data load issue with EIS 6.2. I checked the log file after the data load and I cannot find any errors.

have you tried putting that working SQL statement in the user defined sql and using it for the data load? what do the loadinfo/dataload.txt files say (if anything)

Similar Messages

  • EIS to Essbase data load process

    I have two user defined queries in a metaoutline. When i perform dataload operation from EIS, the Essbase application log says that dataload updated [] cells and Data Load Elapsed Time [] seconds but the EIS dataload process is still running. It seems that Essbase is loading the data only for the first query and completing dataload while EIS is still running second query.Has anyone else experienced this problem? Essbase and EIS both are 6.5.0.

    In EIS
    1) You have to define the Logical OLAP Model connecting to the relational source.It defines the joins between fact table and dimension tables.
    2) Based on the OLAP Model You have to create meta outline which defines the rules for loading members and data into essbase.

  • EIS Member and Data Load-Getting OS Error-Please help!

    Please help! I have created a OLAP model then created a Metaoutline.
    Then I went ahead to do the Member and Data load. I logged into my server and started the member and data load.
    Then it gives me the following error:
    SELECT /*+ */ .. FROM <my_view_name>
    OS Error No such file or directory IS Error Member load terminated with error.
    The load terminated with errors.
    Thanks in advance for any replies.

    thanks all! the error has been resolved.
    Jus had to create a directory in the Integration services folder: $ISHOME/loadinfo
    the loadinfo folder was missing.
    Prathap,
    Is that view available at that time? --the query is generated automatically.
    Which Data Source and which version of the Hyperion - datasource is Oracle10g and 9.3 is Hyp version.

  • EIS data load via rules file

    Hi,
    PLease let me know the process to control the EIS data load into essbase using a rules file. I did not find the option in EIS. please help me.
    Thanks,
    pr

    In EIS
    1) You have to define the Logical OLAP Model connecting to the relational source.It defines the joins between fact table and dimension tables.
    2) Based on the OLAP Model You have to create meta outline which defines the rules for loading members and data into essbase.

  • EIS Data load

    Hi all,
    I am bit confused about loading data via EIS. We have two process one is to load data via rule file hisdld created by EIS and second is using user defined sql query. Are both these methods independent of each other or not?
    When i say independent i mean if we make some changes in user defined sql and load data are those changes first refelected in the rule file and then data load process occurs using this rule file or data load occurs just via sql quey we run
    It would be great if somebody could clarify the same.
    Thanks in advance!!

    Thanks for the reply.
    Yes i am talking about sql query within EIS so if i understood it right by your post if i make changes to that user defined sql code(within EIS ) it will make the changes in the data load rule file created by EIS i.e "hisdld". So If we talk of a data load process within EIS the User defined sql is used to create data load rule "hisdld" and hence they are dependant on each other.
    Edited by: user10859098 on Jan 29, 2009 10:54 AM
    Edited by: user10859098 on Jan 29, 2009 10:54 AM
    Edited by: user10859098 on Jan 29, 2009 10:56 AM

  • EIS data load problem

    Hi,
    I use user defined sql to do a data load using EIS 7x
    When i run the sql in toad just to see the data I see 70000 rows
    But when I use the same sql in the userdefined sql and try to load data...I dont know why EIS says that it loaded around 69000 rows and also I see no rejections...
    I even made some sql changes to find out what are the records that are not being loaded..and I see some rows when I run a sql in toad and if i use the same in EIS its not loading those rows which I can see in toad for the same sql ...
    This is very strange.. Can any one help me with this,..??

    Glenn-
    I dont think there is anything unique about the missing rows..
    Actually the a part of code has been added to the previously working view(which i use to load data) to bring in some additional data...
    I took that part and tried to load data...and I see no data being loaded or rejected..
    It just says that records loaded "0"..rejected 0...
    but the same part brings in the required additional data... when exectued in toad..
    and about the excel sheet lock and send..I did that a week and as you said to my surprise it loads everything....
    This was the test bu which I figured out that EIS is not even able to load a part of data...and I found the exact part of data by going through it closely..
    So I think this is something to do with sql interface..
    And I did not understand the last lines in your post...
    /////I know when I do regular Sql interface some data types don;t load and I have to convert them to varchar tto get them to work. If the Excel file and Flat file loads work, look at that//////
    what did convert to varchar???
    Excel load works fine..So I think its something similar to what you mentioned in your post...
    Can you please explain it ...

  • In EIS at data loading Few Records are getting rejected

    Hello,
    I m loading the data in Essbase through EIS.I am using user defined SQL for data loading.
    In log this this is showing that 22000 records are loaded and 564 records rejected.
    Can any Expert let me know why these records are getting rejected.
    Thx
    Shivendra

    If you look on your EIS server, under hyperion\eis there is a logfiles directory. Under it are folders for every load. In these folders are the error files. If you look at the file, the first column shows you the member that got rejected. You can then look to see if it is in your outline. My guess is your custom sql is bringing back rows that are not in the cube

  • All data load records rejected by EIS

    I have a very strange behavior with EIS:
    We used to successfully build an Essbase cube (V6.5.3) from a SQL 2000 database using EIS (V6.5.4).
    A few weeks ago we doubled the size of a dimension (from about 13000 members to 27000) by adding a new branch exclusively based on shared members (no increase in the data).
    Since then, most of the time (but not all the time !) the data load rejected absolutely all the records.
    I can certify that the records are OK. When we try to reduce this dimension to smaller parts, each individual section can be built and loaded successfully. But when loading everything, EIS reject all the data records.
    We were able to do the full load only once or twice just after having restarted EIS...
    Did anybody already faced such problem ? Are you aware of a limit in the number of members EIS can handle in one dimension ?
    Thanks for any tips or workaround about this issue that becomes very critical for us.
    Best regards,
    Stephane

    i faced some similar problem AIS 9.3.0.1
    once check the rul file for data load in essbase.EIS automatically creates one rul file for data loading.

  • Automate EIS dim builds and data loads

    I want to automate the dimesion builds and data loads from my ETL tool (DTS). I have not been able to find anything about scripting EIS automation in the documentation. Is there any?

    what you can do is create go into EIS metadata outline and create a member load and data load script. Do this by selecting the Outline menu item, then select member load. click next, on this screen, select only save load script. Click the button "Save scripts" to give it a name. click finish. repeat for the dataload script. (If you are using ASO cubes, you must use separate scripts, you can't do both in one script)Then create a batch file to run the member load and data loads. In DTS, use an execute process task to run the batch file

  • EIS data load slow

    Greetings all,
    Windows Server 2003 SP2. SQL Server 2000 SP3. Hyperion 9.3.1.
    I am running an EIS data load, and the load time is progressively longer depending on the number of records loaded.
    For example, when doing a data load of 1,000,000 records, it takes 2 minutes to load the first 500,000 records but an additional 10 minutes to load the next 500,000 records.
    Is there any reason for this, and can I do anything about it?
    Regards,
    Paul.

    Since non-Streaming mode is the only mode that EIS leverages to load data, it's not surprising that- Data loads are slow.
    If you want to continue using EIS for data loads, I believe- All you can do is- To refine/tune your User-defined SQL.
    You've mentioned about just few million records. However, in the volume-intensive environments where records are in the order of hundreds/thousands of millions, the majority would think of using EIS only for Metadata load & deferring data load to the flat file data extracts, which is way faster. At the same time, you'd have to consider if the extraction time required is not too much that it defeats the very purpose.
    - Natesh

  • Incremental Data loading in ASO 7.1

    HI,
    As per the 7.1 essbase dbag
    "Data values are cleared each time the outline is changed structurally. Therefore, incremental data loads are supported
    only for outlines that do not change (for example, logistics analysis applications)."
    That means we can have the incremental loading for ASO in 7.1 for the outline which doesn't change structurally. Now what does it mean by the outline which changes structurally? If we add a level 0 member in any dimension, does it mean structrual change to that outline?
    It also syas that adding Accounts/Time member doesn't clear out the data. Only adding/deleting/moving standard dimension member will clear out the data. I'm totally confused here. Can anyone pls explain me?
    The following actions cause Analytic Services to restructure the outline and clear all data:
    ● Add, delete, or move a standard dimension member
    ● Add, delete, or move a standard dimension
    ● Add, delete, or move an attribute dimension
    ● Add a formula to a level 0 member
    ● Delete a formula from a level 0 member
    Edited by: user3934567 on Jan 14, 2009 10:47 PM

    Adding a Level 0 member is generally, if not always, considered to be a structural change to the outline. I'm not sure if I've tried to add a member to Accounts and see if the data is retained. This may be true because by definition, the Accounts dimension in an ASO cube is a dynamic (versus Stored) hierarchy. And perhaps since the Time dimension in ASO databases in 7.x is the "compression" dimension, there is some sort of special rule about being able to add to it -- although I can't say that I ever need to edit the Time dimension (I have a separate Years dimension). I have been able to modify formulas on ASO outlines without losing the data -- which seems consistent with your bullet points below. I have also been able to move around and change Attribute dimension members (which I would guess is generally considered a non-structural change), and change aliases without losing all my data.
    In general I just assume that I'm going to lose my ASO data. However, all of my ASO outlines are generated through EIS and I load to a test server first. If you're in doubt about losing the data -- try it in test/dev. And if you don't have test/dev, maybe that should be a priority. :) Hope this helps -- Jason.

  • Choosing Dimension Members / Creating Data Load Rule Files

    Hi
    Let me explain this problem using an example. Let's say I have a dimension table called "Product" and its has ProductId, ProductCode & ProductName fields, and a Fact table with ProductId and Value columns. As you can see ProductId has beed used to link to Fact table. I want to display ProductCode as dimension members (ProductName as Alias). I have no problem creating this scenario using EIS. But when I am not sure whether I could do this with EAS. I tried but I cannot get it working. Does anybody know to do this in EAS please? Also, how to create the data load rule for this problem using ProductId?
    Regards
    Chandra

    You can do this using a SQL load rule.
    One rule to build your dimension, simply select statement on the one Dimension table.
    Second rule to load data is going to require a Select statement with a join to pull the product name from the dimension table and the data from the fact table linked on the Product id.

  • Data load problem - BW and Source System on the same AS

    Hi experts,
    I’m starting with BW (7.0) in a sandbox environment where BW and the source system are installed on the same server (same AS). The source system is the SRM (Supplier Relationship Management) 5.0.
    BW is working on client 001 while SRM is on client 100 and I want to load data from the SRM into BW.
    I’ve configured the RFC connections and the BWREMOTE users with their corresponding profiles in both clients, added a SAP source system (named SRMCLNT100), installed SRM Business Content, replicated the data sources from this source system and everything worked fine.
    Now I want to load data from SRM (client 100) into BW (client 001) using standard data sources and extractors. To do this, I’ve created an  InfoPackage in one standard metadata data source (with data, checked through RSA3 on client 100 – source system). I’ve started the data load process, but the monitor says that “no Idocs arrived from the source system” and keeps the status yellow forever.
    Additional information:
    <u><b>BW Monitor Status:</b></u>
    Request still running
    Diagnosis
    No errors could be found. The current process has probably not finished yet.
    System Response
    The ALE inbox of the SAP BW is identical to the ALE outbox of the source system
    and/or
    the maximum wait time for this request has not yet run out
    and/or
    the batch job in the source system has not yet ended.
    Current status
    No Idocs arrived from the source system.
    <b><u>BW Monitor Details:</u></b>
    0 from 0 records
    – but there are 2 records on RSA3 for this data source
    Overall status: Missing messages or warnings
    -     Requests (messages): Everything OK
    o     Data request arranged
    o     Confirmed with: OK
    -     Extraction (messages): Missing messages
    o     Missing message: Request received
    o     Missing message: Number of sent records
    o     Missing message: Selection completed
    -     Transfer (IDocs and TRFC): Missing messages or warnings
    o     Request IDoc: sent, not arrived ; Data passed to port OK
    -     Processing (data packet): No data
    <b><u>Transactional RFC (sm58):</u></b>
    Function Module: IDOC_INBOUND_ASYNCHRONOUS
    Target System: SRMCLNT100
    Date Time: 08.03.2006 14:55:56
    Status text: No service for system SAPSRM, client 001 in Integration Directory
    Transaction ID: C8C415C718DC440F1AAC064E
    Host: srm
    Program: SAPMSSY1
    Client: 001
    Rpts: 0000
    <b><u>System Log (sm21):</u></b>
    14:55:56 DIA  000 100 BWREMOTE  D0  1 Transaction Canceled IDOC_ADAPTER 601 ( SAPSRM 001 )
    Documentation for system log message D0 1 :
    The transaction has been terminated.  This may be caused by a termination message from the application (MESSAGE Axxx) or by an error detected by the SAP System due to which it makes no sense to proceed with the transaction.  The actual reason for the termination is indicated by the T100 message and the parameters.
    Additional documentation for message IDOC_ADAPTER        601 No service for system &1, client &2 in Integration Directory No documentation exists for message ID601
    <b><u>RFC Destinations (sm59):</u></b>
    Both RFC destinations look fine, with connection and authorization tests successful.
    <b><u>RFC Users (su01):</u></b>
    BW: BWREMOTE with profile S_BI-WHM_RFC (plus SAP_ALL and SAP_NEW temporarily)
    Source System: BWREMOTE with profile S_BI-WX_RFCA (plus SAP_ALL and SAP_NEW temporarily)
    Someone could help ?
    Thanks,
    Guilherme

    Guilherme
    I didn't see any reason why it's not bringing. Are you doing full extraction or Delta. If delta extraction please check the extractor is delta enabled or not. Some times this may cause problems.
    Also check this weblog on data Load errors basic checks. it may help
    /people/siegfried.szameitat/blog/2005/07/28/data-load-errors--basic-checks
    Thanks
    Sat

  • BI 7.0 data load issue: InfoPackage can only load data to PSA?

    BI 7.0 backend extraction gurus,
    We created a generic datasource on R3 and replicated it to our BI system, created an InfoSource, the Transformation from the datasource to the InfoSource, an ODS, the transformation from the InfoSource to the ODS. 
    After the transformation creation between the InfoSource and the ODS is done on this BI system, a new folder called "Data Transfer Process" is also created under this ODS in the InfoProvider view.  In the Data Transfer Process, in the Extraction tab, picks 'Full' in the field Extraction Mode, in the Execute tab, there is a button 'Execute', click this button (note: so far we have not created InfoPackage yet) which sounds like to conduct the data load, but find there is no data available even if all the status show green (we do have a couple of records in the R3 table). 
    Then we tried to create an InfoPackage, in the Processing tab, find 'Only PSA' radio button is checked and all others like 'PSA and then into Data Targets (Package by Package)" are dimmed!  In the Data Target tab, find the ODS as a target can't be selected!  Also there are some new columns in this tab, 'Maintain Old Update Rule' is marked with red color 'X', under another column 'DTP(S) are active and load to this target', there is an inactive picture icon, that's weird since we have already activated the Data Transfer Process!  Anyway, we started the data load in the InfoPackage, and the monitor shows the records are brought in, but since in the Process tab in the InfoPackage, 'Only PSA' radio button is checked with all others dimmed that there is no any data going to this ODS!  Why in BI 7.0, 'Only PSA' radio button can be checked with others all dimmed?
    Many new features with BI 7.0!  Any one's idea/experience is greatly appreciate on how to load data in BI 7.0!

    You dont have to select anything..
    Once loaded to PSA in DTP you have the option of FULL or DELTA ,full loads all the data from PSA and DELTA loads only the last load of PSA.
    Go through the links for Lucid explainations
    Infopackage -
    http://help.sap.com/saphelp_nw2004s/helpdata/en/43/03808225cf5167e10000000a1553f6/content.htm
    DTP
    http://help.sap.com/saphelp_nw2004s/helpdata/en/42/f98e07cc483255e10000000a1553f7/frameset.htm
    Creating DTP
    http://help.sap.com/saphelp_nw2004s/helpdata/en/42/fa50e40f501a77e10000000a422035/content.htm
    <b>Pre-requisite-</b>
    You have used transformations to define the data flow between the source and target object.
    Creating transformations-
    http://help.sap.com/saphelp_nw2004s/helpdata/en/f8/7913426e48db2ce10000000a1550b0/content.htm
    Hope it Helps
    Chetan
    @CP..

  • Open HUB ( SAP BW ) to SAP HANA through DB Connection data loading , Delete data from table option is not working Please help any one from this forum

    Issue:
    I have SAP BW system and SAP HANA System
    SAP BW to SAP HANA connecting through a DB Connection (named HANA)
    Whenever I created any Open Hub as Destination like DB Table with the help of DB Connection, table will be created at HANA Schema level ( L_F50800_D )
    Executed the Open Hub service without checking DELETING Data from table option
    Data loaded with 16 Records from BW to HANA same
    Second time again executed from BW to HANA now 32 records came ( it is going to append )
    Executed the Open Hub service with checking DELETING Data from table option
    Now am getting short Dump DBIF_RSQL_TABLE_KNOWN getting
    If checking in SAP BW system tio SAP BW system it is working fine ..
    will this option supports through DB Connection or not ?
    Please follow the attachemnet along with this discussion and help me to resolve how ?
    From
    Santhosh Kumar

    Hi Ramanjaneyulu ,
    First of all thanks for the reply ,
    Here the issue is At OH level ( Definition Level - DESTINATION TAB and FIELD DEFINITION )
    in that there is check box i have selected already that is what my issue even though selected also
    not performing the deletion from target level .
    SAP BW - to SAP HANA via DBC connection
    1. first time from BW suppose 16 records - Dtp Executed -loaded up to HANA - 16 same
    2. second time again executed from BW - now hana side appaended means 16+16 = 32
    3. so that i used to select the check box at OH level like Deleting data from table
    4. Now excuted the DTP it throws an Short Dump - DBIF_RSQL_TABLE_KNOWN
    Now please tell me how to resolve this ? will this option is applicable for HANA mean to say like , deleting data from table option ...
    Thanks
    Santhosh Kumar

Maybe you are looking for