Regarding the cube

hi,
      any one can help regarding i want to know the details of  cubes and ods  which are having the dimensions, character,key fig,lineitem dimension,queries,in how many multiproviders it is using , no of navigational attributes, non cumulative key fig,records it is stored, no of aggregates, no of high cardinality dimensions.
to know above things is there any table or tools  instead of doing manually.
thanks in advance
regards
praveen

hi,
Goto RSA1-> Business content .
or
Transport connection.
Regards,
Senthil Kumar.P

Similar Messages

  • Regarding the tracking the live progress of the cube processing.

    Hello,
             Good Morning.  As we are having multiple cube applicatons as part of our project, there are multiple cubes for which we need to evaluate the completion time for the cube processing.
    By referencing one cube, the cube processing is taking one hour one day and some other day, It is taking an hour for completion. Beacuse of this reason, we could not be able to asses the time for completion of cube processing.
    Is there any chance to track the live status of the cube processing?. Which means, the expected time for completion? How many measure group processing are completed and how many left?
    Please provide your inputs on this.
    Thanks in advance.
    Regards,
    Pradeep.

    Hi Kumar,
    According to your description, you want to monitor the progress of cube processing. Right?
    In Analysis Services, there are several tools can help us monitor the processing performance, like SQL Profiler, AS Trace, Performance Monitor, etc. We can also use XMLA command or DMV to get the information. Please see:
    Monitoring processing performance
    Monitoring and Tuning Analysis Services with SQL Profiler
    However, if you want to get the exact live data of the cube processing, Olaf's script can be an effective solution to get the current processing status for some measures or dimensions. But some information still can't be traced, like expected time for completion,
    etc. So for your requirement, it can't be fully achieved.
    If you have any question, please feel free to ask.
    Simon Hou
    TechNet Community Support

  • Regarding Reloading of Data Into The Cube

    Hi All,
    I have a situation where I have to reload the data into the InfoCube. The InfoCube is already having Full Load and Deltas for a period of 1 month. Now if I do a full load will I get the data as of today and can I do deltas tomorrow onwards after doing an Init???
    Should I have to delete the Data from Delta Queue in Source System (RSA7)???
    Please clarify me.
    Thanks in advance.
    Regards
    Jay

    Hi,
      if you already have intialisations on the cube you can do a repair ful load. which will bring all the data and you can continue using the delta functionality from tomorrow.
    hope this helps.
    Ravi.

  • Problem regarding request not coming to the cube,

    Hello,
    I created infocube then in source of transformation , select object type( data source) and specifed the name and source system u2022 Activate created transformation u2022 Create Data transfer process (DTP) by right clicking the master data attributes u2022 In extraction tab I specified extraction mode ( full) u2022 In update tab I specified  error handling (requestgreen).I activated DTP and executed tab click execute button to load data in data targets.
    Then I right clicked data targets and selected manage and in contents tab select contents to view the loaded data. But the duplicate records were there.So i deletd the requets from cube and again ran the infopackage but the request didn't come again.
    So please tell me what to do in this point of time to bring the request again into the cube and get the selected contents.
    Thanks
    JINI

    This is a known issue and the workaround is to do ctrl+alt+bksp+bksp(kill session) or disable USB redirection for that pool
    if you are not using USB-R.
    Thanks
    P.S.M.Swamiji
    Note: These are my personal opinions, nothing to do with my employer

  • Can not find master data in the cube

    ItemNo is an navigation attribute of 0Material. We are using 0material in a cube. When I display the cube data include ItemNo, I don't find the data I am looking for. For example:
    In 0Material:
    0Material      ItemNo    Description
    AAA                001           Test
    BBB                002           Test
    In the cube I expect to see:
    0Material             ItemNo            On Hand Qty
    AAA                      001                      1000
    BBB                      002                       2000
    Instead I saw:
    0Material             ItemNo            On Hand Qty
    AAA                                                1000
    BBB                                                2000
    0Material got refreshed after cube data got loaded. Is that why I don't find the refreshed Master data?
    Thank you!

    Hi,
    When you laod the cube, the Values for Master are automatically get created in the master data (If they are not present earlier and with "Data Update Type in the Data Tragets" setting in InfoPackage).
    Now when you are loading your cube with 0Material and Item Number, the values for Item Number seems blank. Thats the reason why you are not able to see the Item Number in Cube though it is present in Master. Because now these are two different keys, one with Item Number value and other without Item Number value.
    Regards,
    Yogesh.

  • Amount not getting displayed in the cube

    Hi,
    I have a ZTable which has got an Amount field declared like this.
    AMOUNT     WAERS     CUKY     5     0     Currency Key
    And I declared a ZAMOUNT as Amount in the cube and provided 0CURRENCY.
    Initially there was no value in the Amount column in the Ztable and hence could not see any value of Amount in the cube.But after entering values in the ZTable, I deleted the replicated the DS, create TR, UR and then create an InfoPckg and then created the Data Transfer Process. But still I am not able to see the Amount Value in the cube.
    Kindly Help
    Sam..

    > hi Sam,
    >
    > did you assign AMOUNT field from ztable to infoobject
    > ZAMOUNT in transfer rules ? move AMOUNT from left to
    > right in tab 'datasource./transf.rules ?
    > and ZAMOUNT mapped ? (with ZAMOUNT)
    >
    > hope this helps.
    Hi Edwin,
    Assigning and mapping which you mentioned is just pulling an arrow from the left Amount to right ZAmount, right? I did that. But there is an additional 0Currency on the right side without getting mapped? Should I map that also to Amount? Any issue with Rules.
    Regards,
    Sam...

  • Fields in the cube

    Hi Experts
    Here client wants 200 fields in cube
    whether that is suggestable help me in this
    Regards
    Anand

    Hi,
    Its possible to have that many fields in the cube, but you have to consider what the reporting requirements are and how the data is connected, etc. You also have to consider how much data volume is going to be generated and at what level you are going to store the data in the cube. If your client is suggesting you to have line item level or granular data in your cube , suggest to them to have that in an ODS and provide jump queries from the cube to the ODS if necessary.
    If many of the fields are not going to be required for reporting at present or in the future, suggest to the client to have them in the ODS and a smaller subset of the fields in the cubes.
    Cheers,
    Kedar

  • How to Delete the dimension from the cube ?

    Hi ,
    how to Delete the dimension from the cube ?
    i have added the new dimension by assiging one characteristic to that dimension .
    now i  want to delete it ,
    but system saying that   Dimension ZXXX  contains InfoObjects; deletion not possible .
    how to delete it ? any help .
    Thanks

    Make sure you donot have any data in cube. If you have some data in cube, then you will not be able to see Delete option.
    Right click on the cube-->Delete data.
    Then double click on the cube>Goto Edit mode> select the IO under that Dimension> right click> now you will be able to see the Delete option (provided you have deleted all the date from Cube)
    Then right click on Dimension-->Delete
    Regards,
    Pavan

  • Error while transporting the cube

    Hi all,
    While I was transporting my cube from Development box to Acceptance box, following error occured.
       Execution of programs after import (XPRA)
       Transport request   : GCDK905220
       System              : NCA
       tp path             : /usr/sap/NCA/DVEBMGS01/exe/tp
       Version and release: 372.04.10 700
       Post-import methods for change/transport request: GCDK905220
          on the application server: bdhp4684
       Post-import method RS_AFTER_IMPORT started for CUBE L, date and time: 20080802045222
       Program terminated (job: RDDEXECL, no.: 04522200)
          See job log
       Execution of programs after import (XPRA)
       End date and time : 20080802045237
       Ended with return code:  ===> 12 <===
    Actually transport failed but the cube was present in the Acceptance box in inactive state
    Please help me for the same.
    With thanks and regards
    Mukul Singhal

    Hi Mukul,
    This is an RC=12 issue.Contact the basis team or raise ticket basis team stating that your specified transport request is got terminated.Basis team will resolve the issue and then you can Import the transport into your acceptance system.
    I hope it will help you.
    Regards
    Suresh B.G.

  • I am not getting the data in Bex from the cube

    Hi Experts,
    i am not getting the data from the cube to Bex,
    cube having the data, but i am not able to see this data in the Bex result area.
    Can  any one help me.
    Best regards,
    Bhaskar

    Hi
    In the manage cube, is reporting symbol is there for your requests
    If you have any red requests delete them and refresh
    If all requests having reporting symbols you may check the BEx without any filters or selection criterions
    Regards
    N Ganesh

  • Data in the cube is showing multiple entries when compared with ODS

    Hello BW Gurus,
    We have a waste report in production planning on Cube and ODS separately. The same info package loads both targets (which means same infosource) but when we run a report on Cube, the records are showing multiple entries (i.e. Key Figures are not matching when compared to ODS) where as the ODS records are showing correctly as it was in R/3. There are totally 6 key figures out of which 4 pulled from R/3 and 2 are populated in BW.
    An Example:
    Waste report in PP run for plant 1000 for 12/2005 and process order 123456. The operational scrap should be 2.46% and the component scrap should be 3.00% for material 10000000. The report is showing 7.87% for planned operational waste % and 9.6% for planned component waste %. These values are not correct. The ODS values for order 123456 matched the data in R/3 for component and operational scrap.
    There is a Start routine to the ODS and also to the cube. I am not good at ABAP so requesting your Help.
    Here is the ODS Code:
    tables: /BI0/PPRODORDER.
    loop at data_package.
    select single COORD_TYPE
    PRODVERS
    into (/BI0/PPRODORDER-COORD_TYPE,
    /BI0/PPRODORDER-PRODVERS)
    from /BI0/PPRODORDER
    where PRODORDER = data_package-PRODORDER
    and OBJVERS = 'A'.
    if sy-subrc = 0.
    if /BI0/PPRODORDER-COORD_TYPE = 'XXXX'
    or /BI0/PPRODORDER-COORD_TYPE = 'YYYY'.
    data_package-PRODVERS = space.
    else.
    data_package-PRODVERS = /BI0/PPRODORDER-PRODVERS.
    endif.
    endif.
    if data_package-calday = space
    or data_package-calday = '00000000'.
    if data_package-TGTCONSQTY NE 0.
    data_package-calday = data_package-ACTRELDATE.
    endif.
    endif.
    modify data_package.
    endloop.
    Here is Cube Code:
    tables: /BI0/PPRODORDER,
    /BIC/ODS.
    TYPES:
    BEGIN OF ys_mat_unit,
    material TYPE /bi0/oimaterial,
    mat_unit TYPE /bi0/oimat_unit,
    numerator TYPE /bi0/oinumerator,
    denomintr TYPE /bi0/oidenomintr,
    END OF ys_mat_unit.
    DATA:
    l_s_mat_unit TYPE ys_mat_unit,
    e_factor type p decimals 5.
    loop at data_package.
    select single COORD_TYPE
    PRODVERS
    into (/BI0/PPRODORDER-COORD_TYPE,
    /BI0/PPRODORDER-PRODVERS)
    from /BI0/PPRODORDER
    where PRODORDER = data_package-PRODORDER
    and OBJVERS = 'A'.
    if sy-subrc = 0.
    if /BI0/PPRODORDER-COORD_TYPE = 'XXX'
    or /BI0/PPRODORDER-COORD_TYPE = 'YYY'.
    data_package-PRODVERS = space.
    else.
    data_package-PRODVERS = /BI0/PPRODORDER-PRODVERS.
    endif.
    endif.
    if data_package-calday = space
    or data_package-calday = '00000000'.
    if data_package-TGTCONSQTY NE 0.
    data_package-calday = data_package-ACTRELDATE.
    endif.
    endif.
    data_package-agsu = 'GSU'.
    data_package-agsu_qty = 0.
    select single gr_qty
    base_uom
    into (/BIC/ODS-gr_qty,
    /BIC/ODS-base_uom)
    from /BIC/ODS
    where prodorder = data_package-prodorder
    and material = data_package-material.
    if sy-subrc = 0.
    if /BIC/ODS-base_uom = 'GSU'.
    data_package-agsu_qty = /BIC/ODS-gr_qty.
    else.
    SELECT SINGLE * FROM /bi0/pmat_unit
    INTO CORRESPONDING FIELDS OF l_s_mat_unit
    WHERE material = data_package-material
    AND mat_unit = 'GSU'
    AND objvers = 'A'.
    IF sy-subrc = 0.
    IF l_s_mat_unit-denomintr <> 0.
    e_factor = l_s_mat_unit-denomintr /
    l_s_mat_unit-numerator.
    multiply /BIC/ODS-gr_qty by e_factor.
    data_package-agsu_qty = /BIC/ODS-gr_qty.
    ENDIF.
    else.
    CALL FUNCTION 'UNIT_CONVERSION_SIMPLE'
    EXPORTING
    INPUT = /BIC/ODS-gr_qty
    NO_TYPE_CHECK = 'X'
    ROUND_SIGN = ' '
    UNIT_IN = /BIC/ODS-base_uom
    UNIT_OUT = 'GSU'
    IMPORTING
    OUTPUT = DATA_PACKAGE-gsu_qty
    EXCEPTIONS
    CONVERSION_NOT_FOUND = 1
    DIVISION_BY_ZERO = 2
    INPUT_INVALID = 3
    OUTPUT_INVALID = 4
    OVERFLOW = 5
    TYPE_INVALID = 6
    UNITS_MISSING = 7
    UNIT_IN_NOT_FOUND = 8
    UNIT_OUT_NOT_FOUND = 9
    OTHERS = 10.
    endif.
    endif.
    endif.
    modify data_package.
    endloop.
    some how the AGSU qyt is not populating in the cube and when I dbug the code, I could see a clean record in the internal table but not in the cube.
    your suggestion and solutions would be highly appreciated.
    thanks,
    Swathi.

    Hi Swathi
    In ODs we have option of overwriting and addition however in Cube we have only adition.Thats why you are getting multiple enteries.
    If you are running daily full load on the cube then please delete the earlier requests.
    So at one point of time there should be only one full load request in cube. Hope this will solve your problem.
    Regards,
    Monika

  • Data in the Cube not getting aggregated

    Hi Friends
    We have Cube 1 and Cube 2.
    The data flow is represented below:
    R/3 DataSource>Cube1>Cube2
    In Cube1 data is Stored according to the Calender Day.
    Cube2 has Calweek.
    In Transformations of Cube 1 and Cube 2 Calday of Cube 1 is mapped to Calweek of Cube 2.
    In the CUBE2 when i upload data from Cube1.Keyfigure Values are not getting summed.
    EXAMPLE: Data in Cube 1
    MatNo CustNo qty calday
    10001  xyz     100  01.01.2010
    10001  xyz      100  02.01.2010
    10001  xyz      100   03.01.2010
    10001  xyz     100  04.01.2010
    10001  xyz      100  05.01.2010
    10001  xyz      100   06.01.2010
    10001  xyz      100   07.01.2010
    Data in Cube 2:
    MatNo CustNo qty calweek
    10001  xyz     100  01.2010
    10001  xyz      100  01.2010
    10001  xyz      100   01.2010
    10001  xyz     100   01.2010
    10001  xyz      100   01.2010
    10001  xyz      100   01.2010
    10001  xyz      100   01.2010
    But Expected Output Should be:
    MatNo CustNo qty calweek
    10001  xyz     700  01.2010
    How to acheive this?
    I checked in the transformations all keyfigures are maintained in aggregation summation
    regards
    Preetam

    Just now i performed consisyency check for the cube:
    I a getting following warnings:
    Time characteristic 0CALWEEK value 200915 does not fit with time char 0CALMONTH val 0
    Consistency of time dimension of InfoCube &1
    Description
    This test checks whether or not the time characteristics of the InfoCube used in the time dimension are consistent. The consistency of time characteristics is extremely important for non-cumulative Cubes and partitioned InfoCubes.
    Values that do not fit together in the time dimension of an InfoCube result in incorrect results for non-cumulative cubes and InfoCubes that are partitioned according to time characteristics.
    For InfoCubes that have been partitioned according to time characteristics, conditions for the partitioning characteristic are derived from restrictions for the time characteristic.
    Errors
    When an error arises the InfoCube is marked as a Cube with an non-consistent time dimension. This has the following consequences:
    The derivation of conditions for partitioning criteria is deactivated on account of the non-fitting time characteristics. This usually has a negative effect on performance.
    When the InfoCube contains non-cumulatives, the system generates a warning for each query indicating that the displayed data may be incorrect.
    Repair Options
    Caution
    No action is required if the InfoCube does not contain non-cumulatives or is not partitioned.
    If the Infocube is partitioned, an action is only required if the read performance has gotten worse.
    You cannot automatically repair the entries of the time dimension table. However, you are able to delete entries that are no longer in use from the time dimension table.
    The system displays whether the incorrect dimension entries are still being used in the fact table.
    If these entries are no longer being used, you can carry out an automatic repair. In this case, all time dimension entries not being used in the fact table are removed.
    After the repair, the system checks whether or not the dimension is correct. If the time dimension is correct again, the InfoCube is marked as an InfoCube with a correct time dimension once again.
    If the entries are still being used, use transaction Listcube to check which data packages are affected.  You may be able to delete the data packages and then use the repair to remove the time dimension entries no longer being used. You can then reload the deleted data packages. Otherwise the InfoCube has to be built again.

  • Input ready query is not showing loaded data in the cube

    Dear Experts,
    In Input ready query we have problem that it is not showing the values which was not entered by throught hat query. Is any settings in input ready query that we can do to populate the loaded data on the cube as well as data entered through Input ready query itself.
    Thanks,
    Gopi R

    Hi,
    input ready queries always should display most recent data (i.e. all green and the yellow request). So you can check the status of the requests in the real-time InfoCube. There should exist only green requests and maybe at most one yellow request.
    In addition you can try to delete the OLAP cache for the plan buffer query: Use RSRCACHE to do this. The technical names of the plan buffer query can be found as follows:
    1. InfoCube\!!1InfoCube, e.g. ZTSC0T003/!!1ZTSC0T003 if ZTSC0T003 is the technical name of the InfoCube
    2. MPRO\!!1MPRO, e.g. ZTSC0M002/!!1ZTSC0M002 if ZTSC0M002 is the technical name of the multiprovider
    If the input ready query is defined on an aggregation level using a real-time InfoCube, the first case is relevant; if the aggregation level is defined on a multiprovider the second case is relevant. If the input-ready query is defined on a multiprovider containing aggregation levels again the first case is relevant (find the real-time InfoCubes used in the aggregation level).
    Regards,
    Gregor

  • Data is not getting loaded in the Cube

    Hi Experts,
    I had a cube to which I created aggregates and then I deleted the data of the cube and made some changes in it and trying to load the data again from PSA.
    But I'm the data is not getting loaded in to it the cube.
    Can anyone tell me do I have to delete the aggregates before I start a fresh data in to my cube. If yes can you tell me how do i delete it and load it again.
    If it is something else please help me resolve my issue.
    Thanks

    Hi,
    Deactivate the aggregates and then load the data from PSA to Cube.
    And then reactivate the aggregates. While reactivating them you can ask to fill them also.
    Regards,
    Anil Kumar Sharma. P

  • Value displayed in the cube but not in BEx query

    Hi gurus,
    I have the field for no. of orders  (ANZAU) in the datasource 2LIS_11_VAHDR. It is displaying all the values in the column as '1' in the RSA3 and also in the cube but when the same field is selected in the query there is no display of value '1' .In the query all the values shown in the column is '0' .what is the reason.
    How to display the correct value '1' which is in RSA3?
    regards,
    man

    hi,
    Check out the following
    1.Check the report at the detailed level(line item level)
    2.Check if the report is from multiprovider, if so chk the field you needed is selceted from all the cubes you require.
    3.Run a full repair and chk upto PSA.
    Hope it helps you
    Assign points if so
    Ramesh

Maybe you are looking for

  • Help my daughter did something really bad to my Mactop while playing a game.

    My 3yr old was playing a game and when she was done it appears that she was also playing with my Mactop. My spellcheck is not working right, it underlines the word but when I right click nothing happens. The folders on my desktop are all shrunk and I

  • 2 Service Orders created on save

    Experts, We are trying to create service order process type by directly using the logical link "SRV-ORD-CR" from the navigation bar of IC webclient. However its creating 2 orders on save. This happens only once per session. I have debugged through EH

  • Hosting an IWeb site... and Other Q's

    Hey all! I'm about to start a small business in my town and i've played around with iweb before, and like many of the things you can do with it. I have a couple questions that i hope you guys can help me with... 1) * Biggest one - Does I web allow yo

  • Top-of-page in ALV  AND table

    hy folks, i read many topics about the top of page in alv but none answered to my problem, i have an alv_grid with top of page, but i need to put before the alv grid another table. is it possible to use CALL FUNCTION 'REUSE_ALV_COMMENTARY_WRITE'    

  • IP address assisgnment w/ Secure ACS

    I have been trying to set up a switch to do the following: 1) Use 802.1x port level authentication 2) Assign VLAN to the client based on username/group 3) Assign IP address to the client I am using 3750 switches w/ IOS ver 12.2.25 and ACS server ver