BI Program SAP_INFOCUBE_DESIGNS.

Hello,
Through the BI Program SAP_INFOCUBE_DESIGNS i have got the big dimension in my cube. This dimension has two chac in it and now i want to move one of them to the other small dimension. how do i take the decision on which char to be moved? has SAP provided any method to do that?
Thanks,
Pallavi

It depends on the number of unique values you get for that IO.
Edited by: Ananda Theerthan on Apr 22, 2010 4:32 PM

Similar Messages

  • Redline in program SAP_INFOCUBE_DESIGNS indicates what?

    Hi,
    I am running the SE38 and the program "SAP_INFOCUBE_DESIGNSu201D. In some places it is showing the result for some dimensions as a red line. What does tht mean? and I can see that the relationship with the fact table is of high precentage, but I can also see that other cubes with even higher relationship is not showing red.Can someone plz help with this?
    Rishi

    1.Use the program SAP_INFOCUBE_DESIGNS (Performance of BW infocubes) to see the aggregation ratio for the cube.
    2.If the cube does not appear in the list of this report, try to run RSRV checks on the cube and aggregates.
    Go to SE38 > Run the program SAP_INFOCUBE_DESIGNS
    It will show dimension Vs Fact tables Size in percent.
    Note: unless you are sure  do not select a dimension having high cardinality.
    If your dim table size exceeds the 20% of your fact table then you can say it as high cardinality, for ex: your fact table contains 1000 records and your customer dimension contains 201 records means this dim is with high cardinality. Please check with  client the expected records for those dimensions or for the infoobjs defined under single dimension.Run the program SAP_INFOCUBE_DESIGNS to know the sizes of the dim tables and fact tables, this displays all your info cube's fact and dimension tables with sizes, if any dimension exceeds the more than 10% to 20% it will be highlighted in RED.
    It means that the dimension is to have a large number of instances (that is, a high cardinality). use this inforamtion for optimzation.
    A general rule is that a dimension has a high cardinality when the number of dimension entries is at least 20% of the fact table entries.
    This thumb rule is being used since very long, help.sap.com will give you good understanding of this.
    Rgds,
    Ray
    Edited by: rayapa reddy vaddireddi on Mar 23, 2009 3:48 PM

  • SAP_INFOCUBE_DESIGNS do not shown the current data

    Hi guru's,
    I have deleted all data from an InfoCube with the dimensions and the data from all the master data around the InfoCube.
    But when I execute SAP_INFOCUBE_DESIGNS with the transaction SE38 I see that there are rows in the F-table and the dimensions.
    How is this posible??
    Greetings,
    Raymond

    Hi,
    I think the data which you are looking at is not the actual data from Infocube but the statistics prepared by report SAP_INFOCUBE_DESIGNS.
    We can use this report in many scenarios for e.g.
    BW program SAP_INFOCUBE_DESIGNS can be used to see whether the dimensions in our cubes were over / under the recommended percentage of the fact table as recommended by SAP.
    Regards,
    Durgesh.

  • Cube not appearing in SAP_INFOCUBE_DESIGNS list

    HI Experts,
    I could able to see a cube in a Quality ssytem through SAP_INFOCUBE_DESIGN  program to find the dimension to Fact ratio.
    Now i have done som echanges to the cube(dropped certain fields from cube)  and sent it to Quality system. Now my cube doesn't appear in the Infocube list to trace the Dimension to Fact ratio( I have loaded data).
    Now i would like to Know.
    1. When does this information get updated in a BW system and how( which SAP program or job does this work)?
    2. Is there any possibility that after changes to a cube it would not appear in the list of Cubes?
    3. does the progarm SAP_INFOCUBE_DESIGN gives the list of all the cubes of a system or any limitation or any critteris to deliver that data?
    Your quick responses are really appreciated.
    Thanks in advance.
    Ganesh Thota

    Hello Ganesh,
    This is because STATS has not been collected for the infocube. The program SAP_INFOCUBE_DESIGN read the stats to obtain fact dimension ration, unless the stats are avaiable and up-to-date this would not show accurate information.
    Goto RSA1 --> find the cube --> right click and go to "manage" -->  under the "Performance" tab click on "Refresh Statistics" button (use batch button, if cube is large.
    Upon completion of above steps you should be able see the cube in the output of the program SAP_INFOCUBE_DESIGN .
    Thanks
    Vasu

  • SAP_INFOCUBE_DESIGNS does not show all cubes

    Hello,
    I run standard program SAP_INFOCUBE_DESIGNS (shows the list opf all cubes) and realized that this program does not show cube which was recently created.
    Can anybody knows why?

    Hi Aleksandrs,
    You will need to update the statistics for the new cubes before they will turn up in this program.
    Also see this thread:
    SAP_INFOCUBE_DESIGNS Not Working
    Hope this helps...

  • "density" in SAP_INFOCUBE_DESIGNS

    Can someone tell me what is the meaning of "density" that is displayed in white text for each cube from program SAP_INFOCUBE_DESIGNS ?

    Never really noticed the Density measure - with my color pallet, that line of text is barely visible.
    I think Patrick is talking about the size of a dimension relative the the size of the fact table.  Don't believe that is what the InfoCube density is trying measure.
    Here's my interpretation ( could be full of you know what, but got to take a stab at it ) -
    As Bhanu points out, it's the number of fact table rows divided by the product of the multiplication of all of the row counts all of the dimension tables. ( hope that creates a clear mental image... yikes!)
    This product of the multiplication of the dimensions provides the number of cells in the OLAP cube, assuming that every dimension is being considered.  Now in reality, it would be really rare that every possible combination dimension values exists in your data.  If it did, you would have a density of 100% - that is every cell in the OLAP cube would be populated with data.
    The density measure then, is trying to tell you what percentage of the cells of your OLAP cube are populated - or put another way - of all of the theoretically possible combinations of dimension values, how many actually exist.   
    What do you do with that info? Could perhaps be used to indicate that there are certain relationships among your characteristics that you might want to explore (which is something you would typically look at when defining dimensions during the modelling phase. Have to think a little more about it's use. 
    That's my best guess for what it's worth.  Hey, if I'm wrong, I at least hope it sounded good!
    Pizzaman
    Message was edited by: Pizzaman
    Thinking some more (always a dangerous thing) Not sure of SAP's cube design with respect to the amount of storage a cell in the OLAP cube consumes if it has values vs cells without values.  Maybe there is no difference, but if there is, the density measure might give you an idea of how memory the cube would consume, e.g. a cube with 10,000 cells with 9,000 populated with a value might consume more memory than if the cube only had 1,000 cells populated.  But, I really don't know - maybe a null cell consumes just as much storage - sounds like a question for the Cube gurus at SAP.

  • Report taking too much time in the portal

    Hi freiends,
    we have developed a report on the ods,and we publish the same on the portal.
    the problem is when the users are executing the report at the same time it is taking too much time.because of this the perfoemance is very poor.
    is there any way to sort out this issue,like can we send the report to the individual user's mail id
    so that they can not log in to the portal
    or can we create the same report on the cube.
    what could be the main difference if the report made on the cube or ods?
    please help me
    thanks in advance
    sridath

    Hi
    Try this to improve performance of query
    Find the query Run-time
    where to find the query Run-time ?
    557870 'FAQ BW Query Performance'
    130696 - Performance trace in BW
    This info may be helpful.
    General tips
    Using aggregates and compression.
    Using less and complex cell definitions if possible.
    1. Avoid using too many nav. attr
    2. Avoid RKF and CKF
    3. Many chars in row.
    By using T-codes ST03 or ST03N
    Go to transaction ST03 > switch to expert mode > from left side menu > and there in system load history and distribution for a particular day > check query execution time.
    /people/andreas.vogel/blog/2007/04/08/statistical-records-part-4-how-to-read-st03n-datasets-from-db-in-nw2004
    /people/andreas.vogel/blog/2007/03/16/how-to-read-st03n-datasets-from-db
    Try table rsddstats to get the statistics
    Using cache memory will decrease the loading time of the report.
    Run reporting agent at night and sending results to email. This will ensure use of OLAP cache. So later report execution will retrieve the result faster from the OLAP cache.
    Also try
    1. Use different parameters in ST03 to see the two important parameters aggregation ratio and records transferred to F/E to DB selected.
    2. Use the program SAP_INFOCUBE_DESIGNS (Performance of BW infocubes) to see the aggregation ratio for the cube. If the cube does not appear in the list of this report, try to run RSRV checks on the cube and aggregates.
    Go to SE38 > Run the program SAP_INFOCUBE_DESIGNS
    It will shown dimension Vs Fact tables Size in percent.If you mean speed of queries on a cube as performance metric of cube,measure query runtime.
    3. To check the performance of the aggregates,see the columns valuation and usage in aggregates.
    Open the Aggregates...and observe VALUATION and USAGE columns.
    "---" sign is the valuation of the aggregate. You can say -3 is the valuation of the aggregate design and usage. ++ means that its compression is good and access is also more (in effect, performance is good). If you check its compression ratio, it must be good. -- means the compression ratio is not so good and access is also not so good (performance is not so good).The more is the positives...more is useful the aggregate and more it satisfies the number of queries. The greater the number of minus signs, the worse the evaluation of the aggregate. The larger the number of plus signs, the better the evaluation of the aggregate.
    if "-----" then it means it just an overhead. Aggregate can potentially be deleted and "+++++" means Aggregate is potentially very useful.
    In valuation column,if there are more positive sign it means that the aggregate performance is good and it is useful to have this aggregate.But if it has more negative sign it means we need not better use that aggregate.
    In usage column,we will come to know how far the aggregate has been used in query.
    Thus we can check the performance of the aggregate.
    Refer.
    http://help.sap.com/saphelp_nw70/helpdata/en/b8/23813b310c4a0ee10000000a114084/content.htm
    http://help.sap.com/saphelp_nw70/helpdata/en/60/f0fb411e255f24e10000000a1550b0/frameset.htm
    performance ISSUE related to AGGREGATE
    Note 356732 - Performance Tuning for Queries with Aggregates
    Note 166433 - Options for finding aggregates (find optimal aggregates for an InfoCube)
    4. Run your query in RSRT and run the query in the debug mode. Select "Display Aggregates Found" and "Do not use cache" in the debug mode. This will tell you if it hit any aggregates while running. If it does not show any aggregates, you might want to redesign your aggregates for the query.
    Also your query performance can depend upon criteria and since you have given selection only on one infoprovider...just check if you are selecting huge amount of data in the report
    Check for the query read mode in RSRT.(whether its A,X or H)..advisable read mode is X.
    5. In BI 7 statistics need to be activated for ST03 and BI admin cockpit to work.
    By implementing BW Statistics Business Content - you need to install, feed data and through ready made reports which for analysis.
    http://help.sap.com/saphelp_nw70/helpdata/en/26/4bc0417951d117e10000000a155106/frameset.htm
    /people/vikash.agrawal/blog/2006/04/17/query-performance-150-is-aggregates-the-way-out-for-me
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/1955ba90-0201-0010-d3aa-8b2a4ef6bbb2
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/ce7fb368-0601-0010-64ba-fadc985a1f94
    http://help.sap.com/saphelp_nw04/helpdata/en/c1/0dbf65e04311d286d6006008b32e84/frameset.htm
    You can go to T-Code DB20 which gives you all the performance related information like
    Partitions
    Databases
    Schemas
    Buffer Pools
    Tablespaces etc
    use tool RSDDK_CHECK_AGGREGATE in se38 to check for the corrupt aggregates
    If aggregates contain incorrect data, you must regenerate them.
    202469 - Using aggregate check tool
    Note 646402 - Programs for checking aggregates (as of BW 3.0B SP15)
    You can find out whether an aggregate is usefull or useless you can find out through a proccess of checking the tables RSDDSTATAGGRDEF*
    Run the query in RSRT with statistics execute and come back you will get STATUID... copy this and check in the table...
    This gives you exactly which infoobjects it's hitting, if any one of the object is missing it's useless aggregate.
    6
    Check SE11 > table RSDDAGGRDIR . You can find the last callup in the table.
    Generate Report in RSRT
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/cccad390-0201-0010-5093-fd9ec8157802
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/4c0ab590-0201-0010-bd9a-8332d8b4f09c
    Business Intelligence Journal Improving Query Performance in Data Warehouses
    http://www.tdwi.org/Publications/BIJournal/display.aspx?ID=7891
    Achieving BI Query Performance Building Business Intelligence
    http://www.dmreview.com/issues/20051001/1038109-1.html
    Assign points if useful
    Cheers
    SM

  • Logical grouping of dimensions

    Hi,
    What is criteria to enter dimension into the cube and what does mean by logical grouping of dimensions
    Thanks in advance
    krish

    Hi,
    While designing the dimensions, you must ask to yourself that "What will be the size of dimension table if I add this characteristic to this dimension?".The smaller ratio of size of dimension table/ fact table means  better performance.For this ratio, 20% is dangerous and 10% is preffered.
    If the characteristics included in one dimension have mxn relationship, then your dimension table will be too big(specially if those charactersitics are already big) and performance will be decreased.Therefore, 2 characteristics that have mxn relationship(e.g sales document and sales document item) must be in seperate dimensions i.e two characteristics that have nx1 relationship(e.g material and material group) may be put into one dimension(Not logically related(groupped) characteristics, this is not a criteria)Two characteristics that have nx1 relationship in one dimension will add only m+n rows to dimension table instead of mxn.To sum up, you must create the dimensions by considering the relationships of characteristics that will be included in the cube.I propose you to activate some SAP cubes(e.g 0FIAR_C03) from BI content to view their dimensions.You can see the logic there.
    You can see the size of dimension table and fact table by executing the program SAP_INFOCUBE_DESIGNS in se38.
    Useful threads:
    Regards,
    Güneş

  • Line item dimension issue

    Hi,
          I am on development system and want to know the fact and dimension % as I need to create line item dimesion.
    One solution is executing program SAP_INFOCUBE_DESIGN.
          But if I want to use this same program in Production system, it will take a lot of time and will engage resources......
          Kindly suggest me as to how to find out which dimension should be made a line item dimension?
    Thanks,
    Sonu

    Hi,
    Use RSRV.
    Execute RSTRV and here you have the option of single cubes.
    This takes the ourput from the same tables where the SAP_INFOCUBE_DESIGNS takes.
    All Elementary Tests -> Database -> Database Information about InfoProvider Tables.
    Give your infocube name here and execute.
    It gives you a good detailed output.
    Thanks
    Ajeet

  • Query Is Running Slow

    Hi all,
    I am runnig the query first time but it is taking very much time to run query . are there any parameter which we need to set from BIW when we run the query first time ?
    when i try to dril down any char. or when i try to move any free characteristic to rows it is taking very much time.
    please tell me the steps to improve the performance.
    Regards,
    Komik Shah

    Hi,
    This info may be helpful.
    General tips
    Using aggregates and compression.
    Using  less and complex cell definitions if possible.
    1. Avoid using too many nav. attr
    2. Avoid RKF and CKF
    3. Many chars in row.
    By using T-codes ST03 or ST03N
    Go to transaction ST03 > switch to expert mode > from left side menu > and there in system load history and distribution for a particual day > check query execution time.
    /people/andreas.vogel/blog/2007/04/08/statistical-records-part-4-how-to-read-st03n-datasets-from-db-in-nw2004
    /people/andreas.vogel/blog/2007/03/16/how-to-read-st03n-datasets-from-db
    Try table rsddstats to get the statistics
    Using cache memoery will decrease the loading time of the report.
    Run reporting agent at night and sending results to email.This will ensure use of OLAP cache. So later report execution will retrieve the result faster from the OLAP cache.
    4. Go to SE38 > Run the program SAP_INFOCUBE_DESIGNS
    It will shown dimension Vs Fact tables Size in percent.If you mean speed of queries on a cube as performance metric of cube,measure query runtime.
    3. To check the performance of the aggregates,see the columns valuation and usage in aggregates.
    Open the Aggregates...and observe VALUATION and USAGE columns.
    "---" sign is the valuation of the aggregate. You can say -3 is the valuation of the aggregate design and usage. ++ means that its compression is good and access is also more (in effect, performance is good). If you check its compression ratio, it must be good. -- means the compression ratio is not so good and access is also not so good (performance is not so good).The more is the positives...more is useful the aggregate and more it satisfies the number of queries. The greater the number of minus signs, the worse the evaluation of the aggregate. The larger the number of plus signs, the better the evaluation of the aggregate.
    if "-----" then it means it just an overhead. Aggregate can potentially be deleted and "+++++" means Aggregate is potentially very useful.
    Refer.
    http://help.sap.com/saphelp_nw70/helpdata/en/b8/23813b310c4a0ee10000000a114084/content.htm
    http://help.sap.com/saphelp_nw70/helpdata/en/60/f0fb411e255f24e10000000a1550b0/frameset.htm
    Note 356732 - Performance Tuning for Queries with Aggregates
    5. Run your query in RSRT and run the query in the debug mode. Select "Display Aggregates Found" and "Do not use cache" in the debug mode. This will tell you if it hit any aggregates while running. If it does not show any aggregates, you might want to redesign your aggregates for the query.
    Also your query performance can depend upon criteria and since you have given selection only on one infoprovider...just check if you are selecting huge amount of data in the report
    Check for the query read mode in RSRT.(whether its A,X or H)..advisable read mode is X.
    Generate Report in RSRT  
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/cccad390-0201-0010-5093-fd9ec8157802
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/4c0ab590-0201-0010-bd9a-8332d8b4f09c
    Achieving BI Query Performance Building Business Intelligence
    http://www.dmreview.com/issues/20051001/1038109-1.html
    Performance Tuning with the OLAP Cache
    http://www.sapadvisors.com/resources/Howto...PerformanceTuningwiththeOLAPCache$28pdf$29.pdf
    Business Intelligence Journal Improving Query Performance in Data Warehouses
    http://www.tdwi.org/Publications/BIJournal/display.aspx?ID=7891
    Achieving BI Query Performance Building Business Intelligence
    http://www.dmreview.com/issues/20051001/1038109-1.html
    Hope this helps.
    Thanks,
    JituK

  • Reg: Process Chain, query performance tuning steps

    Hi All,
    I come across a question like,  There is a process chain of 20 processes.out of which 5 processes are completed at the 6th step error occured and it cannot be rectified. I should start the chain again from the 7th step.If i go to a prticular step i can do that particular step, How can i start the entair chain again from step 7.i know that i need to use a function module but i dont know the name of FM. Please somebody help me out.
    Please let me know the steps involved in query performance tuning and aggregate tuning.
    Thanks & Regards
    Omkar.K

    Hi,
    Process Chain
    Method 1 (when it fails in a step/request)
    /people/siegfried.szameitat/blog/2006/02/26/restarting-processchains
    How is it possible to restart a process chain at a failed step/request?
    Sometimes, it doesn't help to just set a request to green status in order to run the process chain from that step on to the end.
    You need to set the failed request/step to green in the database as well as you need to raise the event that will force the process chain to run to the end from the next request/step on.
    Therefore you need to open the messages of a failed step by right clicking on it and selecting 'display messages'.
    In the opened popup click on the tab 'Chain'.
    In a parallel session goto transaction se16 for table rspcprocesslog and display the entries with the following selections:
    1. copy the variant from the popup to the variante of table rspcprocesslog
    2. copy the instance from the popup to the instance of table rspcprocesslog
    3. copy the start date from the popup to the batchdate of table rspcprocesslog
    Press F8 to display the entries of table rspcprocesslog.
    Now open another session and goto transaction se37. Enter RSPC_PROCESS_FINISH as the name of the function module and run the fm in test mode.
    Now copy the entries of table rspcprocesslog to the input parameters of the function module like described as follows:
    1. rspcprocesslog-log_id -> i_logid
    2. rspcprocesslog-type -> i_type
    3. rspcprocesslog-variante -> i_variant
    4. rspcprocesslog-instance -> i_instance
    5. enter 'G' for parameter i_state (sets the status to green).
    Now press F8 to run the fm.
    Now the actual process will be set to green and the following process in the chain will be started and the chain can run to the end.
    Of course you can also set the state of a specific step in the chain to any other possible value like 'R' = ended with errors, 'F' = finished, 'X' = cancelled ....
    Check out the value help on field rspcprocesslog-state in transaction se16 for the possible values.
    Query performance tuning
    General tips
    Using aggregates and compression.
    Using  less and complex cell definitions if possible.
    1. Avoid using too many nav. attr
    2. Avoid RKF and CKF
    3. Many chars in row.
    By using T-codes ST03 or ST03N
    Go to transaction ST03 > switch to expert mode > from left side menu > and there in system load history and distribution for a particual day > check query execution time.
    /people/andreas.vogel/blog/2007/04/08/statistical-records-part-4-how-to-read-st03n-datasets-from-db-in-nw2004
    /people/andreas.vogel/blog/2007/03/16/how-to-read-st03n-datasets-from-db
    Try table rsddstats to get the statistics
    Using cache memoery will decrease the loading time of the report.
    Run reporting agent at night and sending results to email.This will ensure use of OLAP cache. So later report execution will retrieve the result faster from the OLAP cache.
    Also try
    1.  Use different parameters in ST03 to see the two important parameters aggregation ratio and records transferred to F/E to DB selected.
    2. Use the program SAP_INFOCUBE_DESIGNS (Performance of BW infocubes) to see the aggregation ratio for the cube. If the cube does not appear in the list of this report, try to run RSRV checks on the cube and aggregates.
    Go to SE38 > Run the program SAP_INFOCUBE_DESIGNS
    It will shown dimension Vs Fact tables Size in percent.If you mean speed of queries on a cube as performance metric of cube,measure query runtime.
    3. --- sign is the valuation of the aggregate. You can say -3 is the valuation of the aggregate design and usage. ++ means that its compression is good and access is also more (in effect, performance is good). If you check its compression ratio, it must be good. -- means the compression ratio is not so good and access is also not so good (performance is not so good).The more is the positives...more is useful the aggregate and more it satisfies the number of queries. The greater the number of minus signs, the worse the evaluation of the aggregate. The larger the number of plus signs, the better the evaluation of the aggregate.
    if "-----" then it means it just an overhead. Aggregate can potentially be deleted and "+++++" means Aggregate is potentially very useful.
    Refer.
    http://help.sap.com/saphelp_nw70/helpdata/en/b8/23813b310c4a0ee10000000a114084/content.htm
    http://help.sap.com/saphelp_nw70/helpdata/en/60/f0fb411e255f24e10000000a1550b0/frameset.htm
    4. Run your query in RSRT and run the query in the debug mode. Select "Display Aggregates Found" and "Do not use cache" in the debug mode. This will tell you if it hit any aggregates while running. If it does not show any aggregates, you might want to redesign your aggregates for the query.
    Also your query performance can depend upon criteria and since you have given selection only on one infoprovider...just check if you are selecting huge amount of data in the report
    Check for the query read mode in RSRT.(whether its A,X or H)..advisable read mode is X.
    5. In BI 7 statistics need to be activated for ST03 and BI admin cockpit to work.
    By implementing BW Statistics Business Content - you need to install, feed data and through ready made reports which for analysis.
    http://help.sap.com/saphelp_nw70/helpdata/en/26/4bc0417951d117e10000000a155106/frameset.htm
    /people/vikash.agrawal/blog/2006/04/17/query-performance-150-is-aggregates-the-way-out-for-me
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/1955ba90-0201-0010-d3aa-8b2a4ef6bbb2
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/ce7fb368-0601-0010-64ba-fadc985a1f94
    http://help.sap.com/saphelp_nw04/helpdata/en/c1/0dbf65e04311d286d6006008b32e84/frameset.htm
    You can go to T-Code DB20 which gives you all the performance related information like
    Partitions
    Databases
    Schemas
    Buffer Pools
    Tablespaces etc
    use tool RSDDK_CHECK_AGGREGATE in se38 to check for the corrupt aggregates
    If aggregates contain incorrect data, you must regenerate them.
    Note 646402 - Programs for checking aggregates (as of BW 3.0B SP15)
    Thanks,
    JituK

  • Line item dim

    hi,
    could u tell me about line item dim and give me some realtime scenario.
    thanks
    rashmi..

    Hi,
      A typical scenario when you would use line Item dimension would be:
    1. When the number of master data is very high - like for example - customer master running into millions.
    2. When you add this to a dimension - the DIM table size increases dramatically and the cube performance is affected. Since a DIM ID is created for every combination of characteristics possible- where if you put customer master ( 1 mill records ) and the material master ( 50000 records ) you get 1 mill * 50,000 records in DIM Table.
    3. TO avoid this you would separate the customer master into a separate dimension but even then if the size is high you can declare it as a line item dimension which would make the SID part of the fact table and hence improve cube performance. However you can have only one characteristic in a line item dimension.
    4. If you are familiar with the star schema and the extended star schema - a cube with only line item dimension would become a star schema cube.
    Point: If the master table is 20% of the fact table, then its recommended to use line item dimension. It will remove the dimension table and SId table will directly to fact table. During the blue print desing you need to ask the user the amount of data that char is going to be fethced(Approx) and the amount of data the cube is going to be.
    You should decide at what level cube is going to hold the data (summerised or detail). Mostly cusotmer, vendor and material are created as line item dimension.
    when you run stadard program "sap_infocube_designs" it will diplay all the sizes of dimention tables vs fact tables.
    it will display in red those dimentions have exceeded the standard sizes. normally dim tables should have 10 - 20 %      of      the      fact      table.      so based on this criteria you will change the char. as line item dimentions. ex:document      numbers..etc... and you should check high cardinality for the line item dimention tables. if you check high cardinality b-tree indexes will be used instead of bitmap indexes.
    Cardinality: With Line Item Dimension, you can assign ONE Dimension to only one characteristic.
    With Cardinality, you can assign a Dimension containing Multiple Characteristics.
    The scenario totally depends on what is the size of the Dimension Table w.r.t Fact Table. If it is greater than 10% of Fact Table then you need to consider about these.
    Cardinality creates indexes on the Dimension table entries and there by you would see an improvement in performance.
    Point: Some of the options on index type that get created are DB dependent - that is - not available in all the DBs that      BW      runs      on.
    With Oracle DB, setting the Cardinality option causes a b-tree index to be built instead of a bitmap index even if it is not      a      line      item      dim.
    Setting it as a Line Item dim also causes a b-tree index to be built instead of a bitmap index, but also embeds SID directly      in      the      fact      table,      eliminating      the      dimension      table.
    Added mention of embedding SID in fact table for Line Item dimension.
    b-tree indices are more efficient than bitmap in case of high cardinality dimensions.

  • Tuning, short dumps

    What are techniques used for tuning query performance? Please explain in detail?
    What are short dumps how can we check?
    What methodologies we follow for the PSA cleansing? Where exactly is the PSA change log where we can delete the request based on timestamp details am using BW 3.1c?
    plz help points assured....

    Hi,
    This info may be helpful.
    Tuning query performance
    General tips
    Using aggregates and compression.
    Using  less and complex cell definitions if possible.
    1. Avoid using too many nav. attr
    2. Avoid RKF and CKF
    3. Many chars in row.
    By using T-codes ST03 or ST03N
    Go to transaction ST03 > switch to expert mode > from left side menu > and there in system load history and distribution for a particual day > check query execution time.
    /people/andreas.vogel/blog/2007/04/08/statistical-records-part-4-how-to-read-st03n-datasets-from-db-in-nw2004
    /people/andreas.vogel/blog/2007/03/16/how-to-read-st03n-datasets-from-db
    Try table rsddstats to get the statistics
    Using cache memoery will decrease the loading time of the report.
    Run reporting agent at night and sending results to email.This will ensure use of OLAP cache. So later report execution will retrieve the result faster from the OLAP cache.
    Also try
    1.  Use different parameters in ST03 to see the two important parameters aggregation ratio and records transferred to F/E to DB selected.
    2. Use the program SAP_INFOCUBE_DESIGNS (Performance of BW infocubes) to see the aggregation ratio for the cube. If the cube does not appear in the list of this report, try to run RSRV checks on the cube and aggregates.
    Go to SE38 > Run the program SAP_INFOCUBE_DESIGNS
    It will shown dimension Vs Fact tables Size in percent.If you mean speed of queries on a cube as performance metric of cube,measure query runtime.
    3. --- sign is the valuation of the aggregate. You can say -3 is the valuation of the aggregate design and usage. ++ means that its compression is good and access is also more (in effect, performance is good). If you check its compression ratio, it must be good. -- means the compression ratio is not so good and access is also not so good (performance is not so good).The more is the positives...more is useful the aggregate and more it satisfies the number of queries. The greater the number of minus signs, the worse the evaluation of the aggregate. The larger the number of plus signs, the better the evaluation of the aggregate.
    if "-----" then it means it just an overhead. Aggregate can potentially be deleted and "+++++" means Aggregate is potentially very useful.
    Refer.
    http://help.sap.com/saphelp_nw70/helpdata/en/b8/23813b310c4a0ee10000000a114084/content.htm
    http://help.sap.com/saphelp_nw70/helpdata/en/60/f0fb411e255f24e10000000a1550b0/frameset.htm
    4. Run your query in RSRT and run the query in the debug mode. Select "Display Aggregates Found" and "Do not use cache" in the debug mode. This will tell you if it hit any aggregates while running. If it does not show any aggregates, you might want to redesign your aggregates for the query.
    Also your query performance can depend upon criteria and since you have given selection only on one infoprovider...just check if you are selecting huge amount of data in the report
    Check for the query read mode in RSRT.(whether its A,X or H)..advisable read mode is X.
    5. In BI 7 statistics need to be activated for ST03 and BI admin cockpit to work.
    By implementing BW Statistics Business Content - you need to install, feed data and through ready made reports which for analysis.
    http://help.sap.com/saphelp_nw70/helpdata/en/26/4bc0417951d117e10000000a155106/frameset.htm
    /people/vikash.agrawal/blog/2006/04/17/query-performance-150-is-aggregates-the-way-out-for-me
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/1955ba90-0201-0010-d3aa-8b2a4ef6bbb2
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/ce7fb368-0601-0010-64ba-fadc985a1f94
    http://help.sap.com/saphelp_nw04/helpdata/en/c1/0dbf65e04311d286d6006008b32e84/frameset.htm
    Achieving BI Query Performance Building Business Intelligence
    http://www.dmreview.com/issues/20051001/1038109-1.html
    ODS Query Performance  
    Refer.
    /people/juergen.noe/blog/2007/12/13/overview-important-bi-performance-transactions
    /people/prakash.darji/blog/2006/01/26/query-optimization
    Thanks,
    JituK

  • How to improve query performance using infoset

    I create one infoset that including 4 char.and 3 DSO which all are time-dependent.When query run, system show very poor perfomance, sometimes no data show in BEX anayzer. In this case I have to close BEX analyzer at first and then open it again, after that it show real results. It seems very strange. Does anybody has experience on infoset performance improvement. pls info, thanks!

    Hi
    As info set itself doesn't have any data so it improves Performance
    also go through the below tips.
    Find the query Run-time
    where to find the query Run-time ?
    557870 'FAQ BW Query Performance'
    130696 - Performance trace in BW
    This info may be helpful.
    General tips
    Using aggregates and compression.
    Using less and complex cell definitions if possible.
    1. Avoid using too many nav. attr
    2. Avoid RKF and CKF
    3. Many chars in row.
    By using T-codes ST03 or ST03N
    Go to transaction ST03 > switch to expert mode > from left side menu > and there in system load history and distribution for a particular day > check query execution time.
    Statistical Records Part 4: How to read ST03N datasets from DB in NW2004
    How to read ST03N datasets from DB
    Try table rsddstats to get the statistics
    Using cache memory will decrease the loading time of the report.
    Run reporting agent at night and sending results to email. This will ensure use of OLAP cache. So later report execution will retrieve the result faster from the OLAP cache.
    Also try
    1. Use different parameters in ST03 to see the two important parameters aggregation ratio and records transferred to F/E to DB selected.
    2. Use the program SAP_INFOCUBE_DESIGNS (Performance of BW infocubes) to see the aggregation ratio for the cube. If the cube does not appear in the list of this report, try to run RSRV checks on the cube and aggregates.
    Go to SE38 > Run the program SAP_INFOCUBE_DESIGNS
    It will shown dimension Vs Fact tables Size in percent.If you mean speed of queries on a cube as performance metric of cube,measure query runtime.
    3. To check the performance of the aggregates,see the columns valuation and usage in aggregates.
    Open the Aggregates...and observe VALUATION and USAGE columns.
    "---" sign is the valuation of the aggregate. You can say -3 is the valuation of the aggregate design and usage. ++ means that its compression is good and access is also more (in effect, performance is good). If you check its compression ratio, it must be good. -- means the compression ratio is not so good and access is also not so good (performance is not so good).The more is the positives...more is useful the aggregate and more it satisfies the number of queries. The greater the number of minus signs, the worse the evaluation of the aggregate. The larger the number of plus signs, the better the evaluation of the aggregate.
    if "-----" then it means it just an overhead. Aggregate can potentially be deleted and "+++++" means Aggregate is potentially very useful.
    In valuation column,if there are more positive sign it means that the aggregate performance is good and it is useful to have this aggregate.But if it has more negative sign it means we need not better use that aggregate.
    In usage column,we will come to know how far the aggregate has been used in query.
    Thus we can check the performance of the aggregate.
    Refer.
    http://help.sap.com/saphelp_nw70/helpdata/en/b8/23813b310c4a0ee10000000a114084/content.htm
    http://help.sap.com/saphelp_nw70/helpdata/en/60/f0fb411e255f24e10000000a1550b0/frameset.htm
    performance ISSUE related to AGGREGATE
    Note 356732 - Performance Tuning for Queries with Aggregates
    Note 166433 - Options for finding aggregates (find optimal aggregates for an InfoCube)
    4. Run your query in RSRT and run the query in the debug mode. Select "Display Aggregates Found" and "Do not use cache" in the debug mode. This will tell you if it hit any aggregates while running. If it does not show any aggregates, you might want to redesign your aggregates for the query.
    Also your query performance can depend upon criteria and since you have given selection only on one infoprovider...just check if you are selecting huge amount of data in the report
    Check for the query read mode in RSRT.(whether its A,X or H)..advisable read mode is X.
    5. In BI 7 statistics need to be activated for ST03 and BI admin cockpit to work.
    By implementing BW Statistics Business Content - you need to install, feed data and through ready made reports which for analysis.
    http://help.sap.com/saphelp_nw70/helpdata/en/26/4bc0417951d117e10000000a155106/frameset.htm
    /people/vikash.agrawal/blog/2006/04/17/query-performance-150-is-aggregates-the-way-out-for-me
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/1955ba90-0201-0010-d3aa-8b2a4ef6bbb2
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/ce7fb368-0601-0010-64ba-fadc985a1f94
    http://help.sap.com/saphelp_nw04/helpdata/en/c1/0dbf65e04311d286d6006008b32e84/frameset.htm
    You can go to T-Code DB20 which gives you all the performance related information like
    Partitions
    Databases
    Schemas
    Buffer Pools
    Tablespaces etc
    use tool RSDDK_CHECK_AGGREGATE in se38 to check for the corrupt aggregates
    If aggregates contain incorrect data, you must regenerate them.
    202469 - Using aggregate check tool
    Note 646402 - Programs for checking aggregates (as of BW 3.0B SP15)
    You can find out whether an aggregate is usefull or useless you can find out through a proccess of checking the tables RSDDSTATAGGRDEF*
    Run the query in RSRT with statistics execute and come back you will get STATUID... copy this and check in the table...
    This gives you exactly which infoobjects it's hitting, if any one of the object is missing it's useless aggregate.
    6
    Check SE11 > table RSDDAGGRDIR . You can find the last callup in the table.
    Generate Report in RSRT
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/cccad390-0201-0010-5093-fd9ec8157802
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/4c0ab590-0201-0010-bd9a-8332d8b4f09c
    Business Intelligence Journal Improving Query Performance in Data Warehouses
    http://www.tdwi.org/Publications/BIJournal/display.aspx?ID=7891
    Achieving BI Query Performance Building Business Intelligence
    http://www.dmreview.com/issues/20051001/1038109-1.html
    Assign points if useful
    Cheers
    SM

  • Basic Questions on CUBE & DSO size

    Hi Experts
    I had a few basic questions with Regards to Cube and DSO.
    I had a CUBE  & ODS in Live with Historical Data from 2000 to tell date.
    1) How can i check the No of Data Records avaliable in Cube (2000 to till date)
    2) How can i check the No of Data Records avaliable in ODS (2000 to till date)
    AND
    3)When i check the Data Load request in Cube (Context Menu Manage) in request tab i can see
    Request ID,Avaliable for reporting...Transfered Records & Added Records
    What is the difference between Transfered Records & Added Records in a request
    Thanks in advance

    Hi,
    1 & 2. to check ODS/Cube data goto SE11/SE16-->give the name of the ODS/Cube as /BIC/A(ODS name)
    and for the cube as /BI0/F(cube name) (before compression) and /BI0/E(cube name) (after compression).
    3. Transferred records--> are the records loaded from Source.
        Added records--> these are the records which are added to the cube/ods. This added records will depends on the Transfer/update rules of cube and ODS and also key fields in the ODS. This added records may be less or greater than transferred records.
    Added to FCI the program: SAP_INFOCUBE_DESIGNS is to display how many rows per dimension table.
    Hope you understood ......
    Edited by: Ravi kanth on May 19, 2009 5:25 PM

Maybe you are looking for

  • Is it possible for double view in iCal?

    It would be great if I could have my week view, and month view side by side. Or one on top of the other. I'm not talking about the month calender in the corner, I mean the full on moth view. Is this possible? Thanks

  • Blue Ray Project done - time for SD

    Ok I have fininshed my Blue Ray projct and made my iso. Stupid question time. Can I go to project settings and change the project to DVD to make my SD or should I start a brand new project? Thank you.

  • What patterns for this application???

    We have a large standalone C++ application which we are planning to port into java. Right now we are planning to port only the GUI (front end) part and retain the backend in C++ (Though, in the long long run we might even port the backend part to jav

  • Fetching from LDB PNP - Performance Issue

    Hi, My apologies if my question sounds too basic, but i have not worked in HR earlier. So just needed to clarify once. I've researched it well on SDN , then created a solution but just wanted to check with you guys once, if this the the correct way i

  • Safari can't find plugin

    I'm running Safari 5.1.2 on a Windows XP computer at work. Everything seems to work fine except when I do an Edit/Reset Safari. It appears to do the reset, but then an error message comes up saying it can't find a particular plug-in from some program