No data loaded to the cube

hello,
I have created four dimensions, all validated, deployed and loaded successfully.
I created a cube, and used a joiner to load the keys from the tables. It has just one measure, and I am loading it from a different table.
I ran a query with a join condition (where clause) in Oracle SQL developer, it returns the desired result.
So i use the same join condition, in the cube.
the cube and the corresponding map is validated and deployed successfully. When I start the map, it inserts no, records in the corresponding table of the cube.
Please, help me solve this problem.
best regards
P.S: I am using OWB 11g on SUSE 10.2

The loading process in a cube do a join with the dimensions.
"D_DIMENSION_X"."DIMENSION_CODE" = "INGRP1"."D_DIMENSION_CODE" )Thus, you can have no problem even if a dimension code is missing in the dimension table.
I had the problem with the time dimension (miss the year id for instance) and it's very painful to find.
First, you have to retrieve the SQL generated :
Check in this article how to have the SQL (intermediate sql generation)
http://gerardnico.com/wiki/dw/etl/owb/owb_mapping_debugger
Second, Take the sql and execute it by suppressing a dimension, see if you have data in your sql
and repeat the steps until you see the light.
Good luck
Nico

Similar Messages

  • Data Load into the Cube based on Fiscal Year

    Hi All,
    I was told to load the data into the cube coming from 3 different datasources, but based on fiscal year. I was told to load the data for 2010, 2009 and so on.
    Any suggestions please...

    Hi Dear,
                Write the following code in start Routine of Update Rule:-
    In BI 3.x
                   Delete DATA_Package where calday lt '20090101' and calday gt '20091231'.
               to load data onlyfor 2009.
              same thing for 2010.
    In BI 7.0
               Start Routine of Transformation:-
              delete source_package where calday lt  '20090101' and calday gt '20091231'.
    Regards
    Obaid

  • How is data loaded into the BCS cubes?

    We are on SEM-BW 4.0 package level 13. I'm totally new to BCS from the BW view point. I'm not the SEM person but I support the BW side.
    Can anyone explain to me or point me to documentation that explains how the data gets loaded into cube 0BCS_C11 Consolidation (Company/Cons Profit Center? I installed the delivered content and I can see various export data sources that were generated. However I do not see the traditional update rules, infosources etc.
    The SEM person has test loaded some data to this cube and I can see the request under 'manage' and even display the content. However the status light remains yellow and data is not available for reporting unless I manually set the status to green.
    Also, I see on the manage tab under Info-Package this note: Request loaded using the APO interface without monitor log.
    Any and all assistance is greatly appreciated.
    Thanks
    Denny

    Hi Dennis,
    For reporting the virtual cube 0BCS_VC11 which is fed by 0BCS_C11 is used.
    You don't need to concern about the yellow status. The request is closed automatically after reaching 50000 records.
    About datastream - you right - the BW cube is used.
    And if your BW has some cubes with information for BCS on a monthly basis, you may arrange a load from a data stream.
    This BW cube I make as much similar to 0BCS_C11 by structure as possible -- for a smooth data load. The cube might be fed by another cube which contains information in another format. In update rules of the first cube you may transform the data for compatibility of the cubes structure.
    Best regards,
    Eugene

  • Report does not show data , but data exists in the cube.

    Hi All,
    I have a situation where I could not show the data in the report. When I load data from an extractor 0CO_OM_WBS_1 into a Cube directly I am able to show the data in my report. When I load the same extractor to a DSO and from the DSO when I load it into the Cube, the data does not show up in the query. To check the data I use the same restriction and could see the data reside in the cube (LISTCUBE). I compressed the requests, still it is not showing up in the query. No aggregates create on the cube.
    It shows the data if I load directly from the extractor, but not when I load data thru DSO.
    Any ideas.
    Alex(Arthur Samson)

    Hi Alex,
    I am facing same problem, i have data in cube but in report i cant see....
    i have created a generic DS, i loaded the data to DSO  then CUBE.
    data is loaded succesfully and i can see data in CUBE ( Manage ), when i am running the repory i am not getting data.
    i think you solve this issue... plz help me to resolve this issue.
    Regards,
    SHAIK.

  • Performance Tuning Data Load for ASO cube

    Hi,
    Anyone can help how to fine tune data load on ASO cube.
    We have ASO cube which load around 110 million records from a total of 20 data files.
    18 of the data files has 4 million records each and the last two has around 18 million records.
    On average, to load 4 million records it took 130 seconds.
    The data file has 157 data column representing period dimension.
    With BSO cube, sorting the data file normally help. But with ASO, it does not seem to have
    any impact. Any suggestion how to improve the data load performance for ASO cube?
    Thanks,
    Lian

    Yes TimG it sure looks identical - except for the last BSO reference.
    Well nevermind as long as those that count remember where the words come from.
    To the Original Poster and to 960127 (come on create a profile already will you?):
    The sort order WILL matter IF you are using a compression dimension. In this case the compression dimension acts just like a BSO Dense dimension. If you load part of it in one record then when the next record comes along it has to be added to the already existing part. The ASO "load buffer" is really a file named <dbname.dat> that is built in your temp tablespace.
    The most recent x records that can fit in the ASO cache are still retained on the disk drive in the cache. So if the record is still there it will not have to be reread from the disk drive. So you could (instead of sorting) create an ASO cache as large as your final dat file. Then the record would already still be on the disk.
    BUT WAIT BEFORE YOU GO RAISING YOUR ASO CACHE. All operating systems use memory mapped IO therefore even if it is not in the cache it will likely still be in on the disk in "Standby" memory (the dark blue memory as seen in Resource Monitor) this will continue until the system runs out of "Free" memory (light blue in resource monitor).
    So in conclusion if your system still has Free memory there is no need (in a data load) to increase your ASO cache. And if you are out of Free memory then all you will do is slow down the other applications running on your system by increasing ASO Cache during a data load - so don't do it.
    Finally, if you have enough memory so that the entire data file fits in StandBY + Free memory then don't bother to sort it first. But if you do not have enough then sort it.
    Of course you have 20 data files so I hope that you do not have compression members spread out amongst these files!!!
    Finally, you did not say if you were using parallel load threads. If you need to have 20 files read up on having parrallel load buffers and parallel load scripts. that will make it faster.
    But if you do not really need 20 files and just broke them up to load parallel then create one single file and raise your DLTHREADSPREPARE and DLTHREADSWRITE settings. Heck these will help even if you do go parallel and really help if you don't but still keep 20 separate files.

  • HT1386 I have an older iPhone (3gs) and need to upgrade to a newer phone (4S).  I need to get my NOTES, CALENDAR, CONTACTS, PICTURES, etc backed up on iTunes so I can get that data loaded onto the new phone.  But not sure how to do that.

    I have an older iPhone (3gs) and need to upgrade to a newer phone (4S).  I need to get my NOTES, CALENDAR, CONTACTS, PICTURES, etc backed up on iTunes so I can get that data loaded onto the new phone.  But not sure how to do that.  When I open iTunes it has a button that say "Back Up iPhone", but I'm not sure what that does.  When I go into the sync options it say I have another user account and asks me if I want to merge of replace. I'm assuming it's trying to tell me I have an older iTunes Library, but don't know that.  Geez, maybe people over 60 shouldn't have iPhones, iTunes just bafles me.

    http://manuals.info.apple.com/en_US/iphone_user_guide.pdf

  • Looking for a specific data in all the cubes and ods

    Hi Gurus
    "i am looking for all the cubes/ods that contain a specific Controlling area(lets say 0123) and a specific 0plant (lets say plant 4567), now i can go down to every cube and ods and search for it in its contents but i have like hundereds of cubes it will take days, is there a simple way to look for some particular data in all the cubes/ods, and it tells me which cube/ods contains these plants and controlling area."
    <b>now based on this above post i got a reply that abaping can help.</b>
    "you could write an ABAP where you call for every InfoProvider function RSDRI_INFOPROV_READ_RFC like
    loop at <infoprov-table> assigning <wa>.
    call function 'RSDRI_INFOPROV_READ_RFC'
    exporting
    i_infoprov = <wa>
    tables
    i_t_sfc = i_t_rsdri_t_sfc
    i_t_range = l_t_rsdri_t_range
    e_t_rfcdata = l_t_rsdri_t_rfcdata
    exceptions
    illegal_input = 1
    illegal_input_sfc = 2
    illegal_input_sfk = 3
    illegal_input_range = 4
    illegal_input_tablesel = 5
    no_authorization = 6
    generation_error = 7
    illegal_download = 8
    illegal_tablename = 9
    illegal_resulttype = 10
    x_message = 11
    data_overflow = 12
    others = 13.
    endloop.
    i_t_sfc should contain 0PLANT and i_t_range the restriction on you plant value.
    with a describe table statement on l_t_rsdri_t_rfcdata you can get the hits.
    check test program RSDRI_INFOPROV_READ_DEMO for details
    best regards clemens "
    <b>now my question is how do  i use this code to check each and every cube in bw, it seems like it is meant to be for only one cube at a time. and what does he  mean by  "for every infoprovider function"</b>
    thanks

    THANKS

  • Data is not getting loaded in the Cube

    Hi Experts,
    I had a cube to which I created aggregates and then I deleted the data of the cube and made some changes in it and trying to load the data again from PSA.
    But I'm the data is not getting loaded in to it the cube.
    Can anyone tell me do I have to delete the aggregates before I start a fresh data in to my cube. If yes can you tell me how do i delete it and load it again.
    If it is something else please help me resolve my issue.
    Thanks

    Hi,
    Deactivate the aggregates and then load the data from PSA to Cube.
    And then reactivate the aggregates. While reactivating them you can ask to fill them also.
    Regards,
    Anil Kumar Sharma. P

  • Error while loading data on to the cube : Incompatible rule file.

    Hi,
    I am trying to load data on to essbase cube from a data file. I have a rule file on the cube already, and I am getting the following error while loading the data. Is there any problem with the rules file?
    SEVERE: Cannot begin data load. Essbase Error(1019058): Incompatible rule file. Duplicate member name rule file is used against unique name database.
    com.essbase.api.base.EssException: Cannot begin data load. Essbase Error(1019058): Incompatible rule file. Duplicate member name rule file is used against unique name database.
         at com.essbase.server.framework.EssOrbPluginDirect.ex_olap(Unknown Source)
         at com.essbase.server.framework.EssOrbPluginDirect.essMainBeginDataload(Unknown Source)
         at com.essbase.api.session.EssOrbPlugin._invokeMainMethod(Unknown Source)
         at com.essbase.api.session.EssOrbPlugin._invokeMethod2(Unknown Source)
         at com.essbase.api.session.EssOrbPlugin._invokeMethod(Unknown Source)
         at com.essbase.server.framework.EssOrbPluginDirect._invokeProtected(Unknown Source)
         at com.essbase.api.session.EssOrbPluginEmbedded.invokeMethod(Unknown Source)
         at com.essbase.api.session.EssOrbPluginEmbedded.invokeMethod(Unknown Source)
         at com.essbase.api.session.EssOrbPlugin.essMainBeginDataload(Unknown Source)
         at com.essbase.api.datasource.EssCube.beginDataload(Unknown Source)
         at grid.BudgetDataLoad.main(BudgetDataLoad.java:85)
    Error: Cannot begin data load. Essbase Error(1019058): Incompatible rule file. Duplicate member name rule file is used against unique name database.
    Feb 7, 2012 3:13:37 PM com.hyperion.dsf.server.framework.BaseLogger writeException
    SEVERE: Cannot Load buffer term. Essbase Error(1270040): Data load buffer [3] does not exist
    com.essbase.api.base.EssException: Cannot Load buffer term. Essbase Error(1270040): Data load buffer [3] does not exist
         at com.essbase.server.framework.EssOrbPluginDirect.ex_olap(Unknown Source)
         at com.essbase.server.framework.EssOrbPluginDirect.essMainLoadBufferTerm(Unknown Source)
         at com.essbase.api.session.EssOrbPlugin._invokeMainMethod(Unknown Source)
         at com.essbase.api.session.EssOrbPlugin._invokeMethod2(Unknown Source)
         at com.essbase.api.session.EssOrbPlugin._invokeMethod(Unknown Source)
         at com.essbase.server.framework.EssOrbPluginDirect._invokeProtected(Unknown Source)
         at com.essbase.api.session.EssOrbPluginEmbedded.invokeMethod(Unknown Source)
         at com.essbase.api.session.EssOrbPluginEmbedded.invokeMethod(Unknown Source)
         at com.essbase.api.session.EssOrbPlugin.essMainLoadBufferTerm(Unknown Source)
         at com.essbase.api.datasource.EssCube.loadBufferTerm(Unknown Source)
         at grid.BudgetDataLoad.main(BudgetDataLoad.java:114)
    Error: Cannot Load buffer term. Essbase Error(1270040): Data load buffer [3] does not exist
    Thanks,
    Santhosh

    " Incompatible rule file. Duplicate member name rule file is used against unique name database."
    I am just guessing here as I have never used the duplicate name functionality in Essbase, nor do I remember which versions it was in. However with that said I think your answer is in your error message.
    Just guessing again.... It appears that your rule file is set to allow duplicate member names while your database is not. With that information in hand (given to you in the error message) I would start to explore that.

  • Inventory data load from Inventory Cube to another Copy Cube

    Hello Experts,
    I am trying to load Inventory data from the Inventory cube(say YCINV) to a copy cube  say YCOPY_CINV(copy of Inventory cube), but the results appear inconsistant when I compare the reports on these 2 cubes. I am trying to populate a fiield in copy cube so that I can populate the same data back to the original cube with the new field in it, I am doing this reload back and forth for historical data purpose only.
    I have seen lot of posts as how to run the setups for Inventory data, but my case does not need to perform set up runs.
    Does the note 1426533 solve the issue of loading from one cube to another ? we are on SAP BI 7.01 with SP 06 ,but the note specifies SP 07 ?
    I have tried note 375098 to see if it works, but I do not see the options as mentioned from  step "Using DTP (BW 7.x)" in BI to perform this note
    Please advise on whether to go with implementing note 1426533 or is there any other way to load inventory data from one cube to other.
    Regards,
    JB

    Hi Luis,
    Thanks for your reply,
    I do not see any setting like Intial stock in DTP except "initial non-cumulative for non- cumulative". I did try using the option "initial non-cumulative for non- cumulative" ,but the results still do not match with the inventory cube data.I do not see the check box for marker in the copy cube (under roll up tab). Please let me know if we can really implement this solution ,i.e. copying from inventory cube to copy cube and then re-loading it back to inventory cube for historical data. Currenlty, I am comparing the queries on these 2 cubes, if the data matches then I can go  ahead and implement it in Production, other wise it would not be wise to do so .
    Regards,
    JB

  • Error regarding data load into Essbase cube for Measures using ODI

    Hi Experts,
    I am able to load metadata for dimensions into Essbase cube using ODI but when we are trying same for loading data for Measures encountring following errrors:
    Time,Item,Location,Quantity,Price,Error_Reason
    '07_JAN_97','0011500100','0000001001~1000~12~00','20','12200','Cannot end dataload. Essbase Error(1003014): Unknown Member [0011500100] in Data Load, [1] Records Completed'
    '28_JAN_97','0011500100','0000001300~1000~12~00','30','667700','Cannot end dataload. Essbase Error(1003014): Unknown Member [0011500100] in Data Load, [1] Records Completed'
    '28_JAN_97','0011500100','0000001300~1000~12~00','500','667700','Cannot end dataload. Essbase Error(1003014): Unknown Member [0011500100] in Data Load, [1] Records Completed'
    Can anyone look into this and reply quickly as it's urgent requiremet.
    Regards,
    Rohan

    We are having a similar problem. We're using the IKM SQL to Hyperion Essbase (DATA) knowledge module. We are mapping the actual data to the field called 'Data' in the model. But it kicks everything out saying 'Unknown Member [Data] in Data Load', as if it's trying to read that field as a dimension member. We can't see what we missed in building the interface. I would think the knowledge module would just know that the Data field is, um, data; not a dimension member. Has anyone else encountered this?
    Sabrina

  • Attributes not loaded in the cube.

    Hi All,
    I am loading master data, the load is successful.while i run the query i cannot able to see the value of the master data attribute instead i can see "NOT ASSIGNED".
    More to it, i cannot able to see the data in the cube for the loaded masters attribute.
    Attribute is converted into individual characteristics and Navgational is on.and variable is created.
    In the query for the master data variable "ready for input" is checked.
    But still i cannot see the values in the selection option as well as inthe query output.
    Can any one help to resolve this error...
    Ganesh Kumar

    Hello Ganesh,
    After loading master data attributes you will have to activate it then only it will be available in reporting part.
    Right click on that info object and activate master data option will be visible to you.
    Regards
    Kapadia

  • Data Loading Error for cube 0TCT_C22

    Dear Colleagues,
    I am working on BI 7.0 / SP09.
    I am loading technical content cube 0TCT_C22 from datasource 0TCT_DS22. Till PSA, there is no problem with data load. But from PSA to data target, data load fails. Went to monitor and it shows teh error "Error calling number range object 0TCTPRCSCHN for dimension D2 (  ). Message no: RSAU023".
    I tried to find the SAP notes fro this, but no success. Also checked the Dump and application logs, but nothing is there.
    Please advice ASAP.
    Regards
    PS

    Hi Pank,
    I just solved the very similar issue. Try what I did to see if it works for you.  For each dimension in each Infocube a number range is created. For some weird reason during the activation of the Infocube the number range for the dimension giving troubles was not created. Look for it in TCODE SNRO and you should find a number range per each dimension in the cube but the one giving you error.
    To solve it, (the easiest way I found) just add any characteristic to the problematic dimension. Activate the Infocube. After that, modify again you Infocube and remove the characteristic you just added to leave the dimension how you need it. Activate the Infocube again. By doing that you will force the regeneration of the dimension and with it the number range. You can chek in TCODE SNRO and the number range should be there. Try loading your data again and it should work.
    One thing I don't understand is why that number range sometimes is not created during activation
    Good luck, I hope you can solve it!!!
    Regards,
    Raimundo Alvarez

  • Data loaded in the target are not reportable.

    Dear all,
    In one of our cubes, data is loaded daily. A couple of days back a load failed with ABAP Dump " Raise exception" - subsequently the other days have data loaded but are found to be not reportable.
    How to handle this.
    Can anyone provide reasons for the error and subsequently making the requests reportable in steps.
    Experts help will be appreciated.
    Note:- Further details of the error gives details as "ADD_PARTITION_FAILED"
    Regards,
    M.M

    Hi,
      can you have look in ST22 for Short dumps on the day this load got failed, there you can get some more details of the error. it will help us to understood your error.
    have a look to the below SAP Note:539757
    Summary
    Symptom
    The process of loading transaction data fails because a new partition cannot be added to the F fact table. The loading process terminates with a short dump.
    Other terms
    RSDU_TABLE_ADD_PARTITION_ORA , RSDU_TABLE_ADD_PARTITION_FAILED , TABART_INCONSITENCY, TSORA , TAORA , CATALOG
    Reason and Prerequisites
    The possible causes are:
    SQL errors when creating the partition.
    Inconsistencies in the Data Dictionary control tables TAORA and TSORA. There must be valid entries for the data type of the table in these tables.
    Solution
    BW 3.0A & BW 3.0B
               If there are SQL errors: Analyze the SQL code in the system log or short dump and if possible, eliminate the cause. The cause is often a disk space problem or lock situations on the database catalog or, less frequently: The partitioning option in the ORACLE database is not installed.
               In most cases, there are inconsistencies in the Data Dictionary control tables TAORA and TSORA. As of Support Package 14 for BW 3.0B/Support Package 8 for BW 3.1C, the TABART_INCONSITENCY exception is issued in this case.
               1. ) Check whether there is an entry for the corresponding data type of the table in the TAORA table.
               The TAORA table contains the assignment of data classes to data tablespaces and their attributes, for example:
               Tablespace Data class
               DDIM      PSAPDIMD   . .......
               DFACT     PSAPFACTD  ........
               DODS      PSAPODSD   . ......
               2. ) Check whether the TSORA contains an entry for the tablespace from the TAORA table.
               The TSORA table contains the assignment of the data tablespace to the index tablespace.
               TABSPACE                      INDSPACE
               PSAPDIMD                      PSAPDIMD
               PSAPFACTD                      PSAPFACTD
               PSAPODSD                      PSAPODSD
               For more information, see Notes 502989 and 46272.

  • Can i get the data file from the cube

    Hi All,
    Previously I created and activated one cube using one data file, by mistake lost that data file. Is there any chance to get that data file from that Cube?
    Thanks
    Bhaskar

    Hi Paul,
    yes you can..
    1) If you have loaded through  PSA , Then goto that PSA table and from settings menu> change display variants-> view tab> view as Excel> then save this file
    2) If you have not used PSA.then use LISTCUBE to view the data and change settings to excel and download..
    Don't forget to assign points,If it is useful..
    Further queries always welocme..
    Regards.
    Pradeep choudhari

Maybe you are looking for

  • Curve won't sync with Outlook Calendar 2007

    First the specs: BB 8330m (CDMA) - Cellular South - running v 4.5.0.127 Win XP Pro SP3 Outlook 2007 DM is v  5.0.0 B28 Multilanguage - With Media Manager No BES The problem: Although the phone and Outlook Calendar had been able to sync (either throug

  • "could not initialize photoshop because the file could not be found"

    I get this error whenever I try and run Photoshop CS3. Any ideas?

  • Blurred image when picking highest available resolution

    Hi all, I do apologize if this issue was discussed somewhere else on this forum - I could not find any related post so I decided to start this topic. A few words of introduction - I am using 27'' dell ultra-sharp (U2711) with my old nvidia GeForce GT

  • SetCurrentRow feature or bug?

    When trying to set the current row of a RowSet to a Row that are obtained from the same Rowset (using findByKey) but with a rangeSize smaller than the total number of rows in the row set, the method setCurrentRow(Row) of the RowSet doesn't work: View

  • Share the SVG pie...

    HTMLDB's charting ability and the ease of it is one of my most favorite. I'd like to be able to display simple 2-D SVG pie chart in my app on 8i so I can get away from the nasty dependency on "M$ Web Component" ... i am wondering if you could be so k