How to check data in the cube

for integration purpose, how can you go and check data in the cube and validate it against the query data ? i know we can go to corresponded infocube and right click "display data" but what's on the query is not available when i execute "display data"

You can always use the similar set of Restrictions as used in RKF and get the value very close to what RKF is displaying.
But getting the same values in CKF is a bit tough..I can suggest may be you can take few/1-2 examples and do the calculations manually and compare the query result and Infocube data.

Similar Messages

  • How to check data in the cube ? just for a selected request .

    I am running 5 different DTP for a cube. and i can see 5 Succesful request ID available for reporting (In manage cube ).
    Now, how do i check recoards for a perticuler request, so that i can check , if the data is comming right or not !
    if i select the request id ( in the Request Tab ) and then switch to content tab, then i guess its showing me all the recoards , which are there in the cube by all the DTP.
    please help me out,

    Check if you have Loading date Field in your info cube. you can reconcile the records based on loading date.

  • How to load data in the cube

    dear all,
    I need some help: I'm new to OWB and I'm trying to load a cube with dimensions keys data. I have successfully deployed the cube and dimensions without any problems. Now when I deploy the mapping to load the cube I cannot get any data there...
    - I have tried mapping by using dimension operator and cube operator by mapping dimension business keys with keys in the cube...and still no data...
    -I have also tried joining all the dimensions with the joint operator and joint condition with the cube as input and have the output to the cube...but still no data. I have also disable the constraints on the cube for the purpose of loading the data, but without much success. What am I doing wrong?
    Any help and guidance will be greatly appreciated. Thanks much!

    It sounds like you are not getting any matching keys for loading into the cube.
    Do you have a time dimension created by OWB in the cube? The key used by the cube operator for an OWB time dimension is a formatted number. The time dimension keys are stored as follows;
    Day Level - YYYYMMDD
    Month Level - YYYYMM
    Week Level - YYYYWW
    Quarter - YYYYQ
    Year - YYYY
    If you have a source that has a SQL date datatype for example and want to construct the key for a cube's time dimension at the day level something like the following expression can be used to construct the time reference from a SQL date...
    to_number(to_char( time_key, 'YYYYMMDD'))
    It may not be this but just a thought.

  • How to reduce the run time of ABAP code (BADI) when it is reading huge data from BPC Cube and thus writing  back huge data to the cube.

    Hi All ,
    In Case of reading huge amount of record from BPC Cube  from BADI code , performing calculations and writing back huge amount of data into the cube , It takes lot of time . If there is any suggestion to read the data  from Cube  or writing data into the cube using some Parallel Processing  methods , Then Please suggest .

    Hi Gersh ,
    If we have a specific server say ( on which we are working, Then under RZ12 we make the following entry  as :
    parallel_generators         ( Lets assume : The instance number is 10 )
    Now in SM59 under ABAP Connections , I am giving the following technical settings:
    IP address        
    Instance number          10
    Now if we have a scenario of load balancing servers with following server details (with all servers on different instance numbers ) :
    In this case how can we make the RZ12 settings and SM59 settings such that we don't have to hardcode any IP Address.
    If the request is redirected to and not to , in that case how will the settings be.
    I have raised this question on the below thread :
    How to configure RZ12  and SM59 ABAP connection settings when we have work with Load Balancing servers rather than a specific server .

  • How to check data type of the field symbol at run time

    My code is as following:
          LOOP AT <fs> ASSIGNING <wa_covp_ext>.
            ASSIGN COMPONENT 86 OF STRUCTURE <wa_covp_ext> TO <f_zzname>.
            IF sy-subrc = 0.
              ASSIGN COMPONENT 158 OF STRUCTURE <wa_covp_ext> TO <f_pernr>.
              IF sy-subrc = 0.
                  SELECT SINGLE sname INTO <f_zzname> FROM pa0001
                                WHERE pernr = <f_pernr>
                                AND endda GE sy-datum
                                AND begda LE sy-datum.
    This query is giving dump when <f_zzname> is type P length 8 and decimals 2, because it tries to put PA0001-sname into it which is type C length 30. So I want to check the type of <f_zzname> before the select statement. If it is character 30, then I will write the select statement else not.
    How to check data type of the field symbol at run time? If it's not possible, then can somebody suggest a workaround? Thanks.

    check this ...
    write describe statement  ...
    field-symbols : <f_zzname> .
    data : sname like pa0001-sname,
    assign sname to  <f_zzname>.
    describe  field <f_zzname> type typ.
    write : typ. <-- typ contains character type in this case ..
    U can check if typ is of character(C) if so .. write the select statement ...

  • How to find the date when the cube was created

    Hi gurus
    I just want to find out the date when the cube was created,

    Go to table RSDCUBE, you can see date there.
    Hope it helps,

  • Error while loading data into the cube

    I loaded data on to PSA and when I am loading the data to the cube through DataTransferProcess, I get an error (red color).
    Through "Manage", I could see teh request in red. How do I get to knoe the exact error? Also what could be the possibel reason for this?
    Also can some one explain the Datatransfer process(not in process chain)?

    Hi Sam
        after you load the data  through DTP(after click on execute button..) > just go to monitor screen.. in that press  the refresh button..> in that it self.. you can find the  logs..
       otherwise.. in the request  screen also..  beside of the request number... you can see the logs icon.. you can click on this..
    DTP  means..
    DTP-used for data transfer process from psa to data target..
    check thi link..for DTP:
    to load data in to the datatargets or infoproviders like DSO, cube....
    in your  case.. the problem may be.. check the date formats.. special cherectrs... and

  • Input ready query is not showing loaded data in the cube

    Dear Experts,
    In Input ready query we have problem that it is not showing the values which was not entered by throught hat query. Is any settings in input ready query that we can do to populate the loaded data on the cube as well as data entered through Input ready query itself.
    Gopi R

    input ready queries always should display most recent data (i.e. all green and the yellow request). So you can check the status of the requests in the real-time InfoCube. There should exist only green requests and maybe at most one yellow request.
    In addition you can try to delete the OLAP cache for the plan buffer query: Use RSRCACHE to do this. The technical names of the plan buffer query can be found as follows:
    1. InfoCube\!!1InfoCube, e.g. ZTSC0T003/!!1ZTSC0T003 if ZTSC0T003 is the technical name of the InfoCube
    2. MPRO\!!1MPRO, e.g. ZTSC0M002/!!1ZTSC0M002 if ZTSC0M002 is the technical name of the multiprovider
    If the input ready query is defined on an aggregation level using a real-time InfoCube, the first case is relevant; if the aggregation level is defined on a multiprovider the second case is relevant. If the input-ready query is defined on a multiprovider containing aggregation levels again the first case is relevant (find the real-time InfoCubes used in the aggregation level).

  • How to load data to a cube from multiple infosources ?

    Hi friends,
    How to load data to a cube from multiple infosources ? could u please answer this question .
    thanks in advance......

    Hi ,
    say for example you need to load data to 1 cube from 3 info sources:
    1) You need to create 3 update rules for the Cube.
    2) Each time you create the update rules. Mention the name the name of the Info source. and create update rules correspondingly.
    Message was edited by:
            satish murthy

  • How to write data in the Infocube using BI IP input ready functionality

    Hi All,
    Please help me in the following scenario...
    Input given to me is-- ZCUBSDID cube with associated role ZSP_BW_SAI_BUDGETS-DID_ALL.
    Query is ZR_V7_CO_PA_CH_PB_DID_BUDGET_2 based on agregation level ZALBSDID (& ZALBVDID) in infoarea ZDACOPA.
    I have to write data in the cube by using Input-Ready functionality of BI IP. Can you please provide me step by step procedure for writing the data in the cube as i am very new in IP side.
    Also if any further input required just let me know.
    My Analysis-> I checked that cube is in Planning mode and in Query Designer>Property Pane-->Planning Pane, Query is in Change mode and also associated Keyfigures are in Input-Ready enabled.
    Also i am able to access query in Analyzer.
    This is very critical for me and need to provide solution asap. Please update asap. It really be great help.
    Thanks & Regards
    Rupali Singh

      To enter data in to planning cube through manual planning, you can do the following,
    1. In a query that is used for manual planning, a cell is only input ready if each characteristic value of all the characteristics included in the aggregation level is unique. None of the aggregated values on the aggregation level are therefore input ready: Totals, subtotals and inner hierarchy nodes are not input ready.
    2. To be able to change values for calculated key figures (like Average Price as a quotient of Amount and Quantity), these must be based on input-ready formulas, and at least one operand must be input ready. More information: Defining Inverse Formulas (Designtime) and Inverse Formulas (Runtime).
    3. In order to change aggregated values (with respect to the aggregation level), these values must be disaggregated on all the data records that contribute to the aggregated value of the cell. More information: Disaggregation (Top-Down-Distribution).
    4. If a query used for manual planning includes a navigation attribute that is restricted using a fixed or dynamic filter or a restricted key figure, the system treats the navigation attribute as a normal characteristic. The rule under point 1 applies here. The system only reacts as though the navigation attribute were not part of the query if the navigation attribute is not restricted.
    5. In a query defined on a MultiProvider or a complex aggregation level that you want to use manual planning for, a cell is not input ready if the InfoProvider that is defined by this cell is:
                                a.      Not a real-time InfoCube
                                b.      A real-time InfoCube that has been switched to load mode
    6. If an input-ready query is executed in change mode, but the requested data is locked by another user, the query starts in display mode.
    For more details, you can refer to the following link,
    Hope this helps you to solve your issue.

  • How to extract data from info cube into an internal table using ABAP code

    Can Anyone plz suggest me
    How to extract data from info cube into an internal table using ABAP code like BAPI's or function modules.
    Thankx in advance

    HI Dinesh,
    Thankq for ur reply
    but i ahve already tried to use the function module.
    When I try to Use the function module RSDRI_INFOPOV_READ
    I get an information message "ERROR GENERATION TEST FRAME".
    can U plz tell me what could be the problem

  • How to copy data in ASO cube across two scenario's?

    How to copy data in ASO cube across two scenario's?

    There are multiple ways to do it.
    How is the source scenario sourced? If it is populated using load rules, then why don't change it in the rule itself?

  • Wish to keep always 2 years of data on the cube

    Hi Experts,
                 we wish to keep always 2 years of data on the cube and archive the data after it is 2 years old.
    Any thought on this is appreciated.

    hi Varun,
    you can use tcode SARA and 0calyear as selection to archive the cube to flat file or another cube,
    for more detail please follow this how to archive in bw doc
    hope this helps.

  • Infospke not pulling data from the cube

          Though there is data in the cube ,when I say start extraction the infospoke is pulling zero records .But I see a file created but I am not able to see the data.In the monitor it says 0 record extracted.
        I have done this first time in the new upgrade system where I changed the server name.

    Check out the layout definition(Data column) for the corresponding planning level and also you can check the <b>planning area</b> value whether it corresponds to your cube name ..
    Also make a cross check with the <b>listcube</b> for your corresponding restriction which you have made in planning level..

  • How to extract data from BW cube

    Hello friends,
    Kindly advice me how to extract data from BW cube.
    My requirement is to write an ABAP program to retrive the data from the BW cube.
    Kindly provide me any sample code for my requirement.
    Thanks in advance.
    S.Jaffer Ali

    Hi Jaffer,
    If you want to load the data from the BW cube into any other Cube/ODS in the same of another BW system you can use the export datasource.
    If you want to load the data from the BW cube into any other system then you can use any of these options:
    1. Infospoke - Open hub service and is widely used.
    2. APD
    3. RFC enabled fn modules.

Maybe you are looking for