To determine the Queries which uses Berfore Aggregation

Hi,
Is there any best way to determine the Queries which uses Berfore Aggregation properies like any table which has information of all the calculated key figures which uses Before aggregation properties.
Otherwise how can we determine the same using RSRT.
Please help.
Thanks
Bhanukiran

Hi,
Below listed are the tables related to queries. See if you get this information in any one of those...
RSZELTDIR     Directory of the reporting component elements
RSZELTTXT     Texts of reporting component elements
RSZELTXREF     Directory of query element references
RSRREPDIR     Directory of all reports (Query GENUNIID)
RSZCOMPDIR     Directory of reporting components
RSZRANGE     Selection specification for an element
RSZSELECT     Selection properties of an element
RSZELTDIR     Directory of the reporting component elements
RSZCOMPIC     Assignment reuseable component <-> InfoCube
RSZELTPRIO     Priorities with element collisions
RSZELTPROP     Element properties (settings)
RSZELTATTR     Attribute selection per dimension element
RSZCALC     Definition of a formula element
RSZCEL     Query Designer: Directory of Cells
RSZGLOBV     Global Variables in Reporting
Regards,
Yogesh

Similar Messages

  • How to determine the process alias using condition technique in Transportation & Shipment Scheudling?

    Dear All,
    I am trying to use the functionality Transportation and Shipment Scheduling in GATP.
    I would like to know that how the sytem determines the process alias using condition technique in Transportation & Shipment Scheudling?
    As the Transportation & Shipment Scheduling functionality can be extended further by using the configurable process to overcome the complex scenario. But it needs that process alias should not be determined by using the condition technique.
    Moreover, is it possible to use both functionality simultaneously in the system. i.e. Transportation and Shipment Scheduling using condition technique and Transportation and Shipment Scheduling using configurable process?
    Thanks & Regards
    Piyush Ranpura

    Hi Piyush,
    I have added a small overview in the SCN WIKI which shows you the two ways for the process alias determination and also here as attachment:
    http://wiki.sdn.sap.com/wiki/display/SCM/Time+and+Scheduling+Functions
    Yes, depending on your customizing you can use both scheduling methods in parallel. e.g. you activate CPS just for a specific ITEM category or product and/or customer and/or....what ever you want.
    best regards,
    Michael

  • MORE THAN 1 OF THE DETAIL FOLDERS USES NON-AGGREGATED ITEMS

    What does this join configuration message mean? I am a user not an administrator but I notice that sometimes I get this message when I am trying to create reports in Discoverer.
    Currently I have a report that is using a control table and a detail table. I need to link to another table (not a detail table). The join did not exist but I was able to take a transaction and join to the table with sql. I asked that this join be created in Discoverer so I can link to the other table in the report and it was created, but when I go to pull in the field I need I get this message:
    Join configuration cannot be resolved. Reason: more than 1 of the detail folders uses non-aggregated items.
    Can this be resolved and if so how? We seem to get this message fairly often.

    I am always leery of disabling fan traps, as it can cause errors with aggregation (and by errors, I mean including values multiple times). Make sure you check the documentation (the Disco user guide) for an explanation of fan traps, and the types of problems that can occur, and the check the results of your queries (or have the data owners do it) to make sure the results are valid.
    I use the fan trap warning as an indication that there is something amiss in the BA, and then try to find the problem there. Sometimes the solution is changing a join (you can check the box that states there is a 1 to 1 relationship and that will also eliminate the fan traps), other times it is creating a view or custom/complex folder to specify how the data is to be joined. Rarely is the only solution to disable fan traps (and while I am on my soapbox, I'll add that I use it as a last resort).
    Sometimes removing fan traps is the correct solution, but you need to test your reports to make certain the results are valid.

  • Need to find out the queries which did not executed in 2014

    Hi All,
    I need to find out the queries which start from YY  and did not executed in 2014, earlier we used to find out by looking into the technical content cube 0TCT_C02 but now its giving dump as number of queries increases. I tried to find out using table RSZCOMPDIR but not able to trace it exactly. Please suggest some alternative solution to find out the queries which did not execute in 2014 so that we can delete it.
    Thanks,
    Asim.

    Hi,
    RSDDSTAT_OLAP : This is the table which contains query execution details
    CALDAY field in that table gives you query execution date
    OBJNAM gives the query name
    As you said RSZCOMPDIR gives list of all queries. doing a lookup between these tables will identify the queris which are not executed in particular period.
    -Sriram

  • How to determine the purchasing group using BADI BBP_PGRP_FIND

    Hello Friends
    We use SRM Server 5.5 (Classic scenario).
    As you know that, purchasing group has the responsible product categories in PPOMA_BBP, when user creates the shopping cart to determine the purchasing group. 
    In our case, we should determine the purchasing group using other criteria.
    MM side is OK to determine the purchasing group using specific criteria.
    Because Material Master has the purch. Group information. 
    In case of SRM, we should develop the logic using BBP_PGRP_FIND BADI.
    But I’m not sure how to use this badi.
    We want to determine the purchasing group based on the following logic.
    If described shopping cart without product ID
    Do Standard purchasing group determination logic.
    Else
    Assign the purch group from MARC- EKGRP (Purchasing Group)
    Is it possible to determine above the logic?
    If yes, Could you give me some information?
    Thank you,
    Best regards,
    SH
    Message was edited by:
            So Hee Lee

    Thank you for your answer.
    I found the way to develop the our logic using BADI: bbp_pgrp_find.
    But I’m not sure the following logic is correct.
    To differentiate described req. or product id req, We check the E_ITEM – PRODUCT using BBP_PD_SC_GETDETAIL.
    IF E_ITEM – PRODUCT is blank, Read the category ID to find the purchasing group.
    To find the purch. Group, we can use FM:  BBP_OM_DETERMINE_RESP_PGRP.
    Import parameter: IS_RESP_ITEM_DATA – Category _ID
    Export parameter: ES_PURCH_DATA
    ELSE
    CALL the customer function to read the R/3 purchasing group from MARC- EKGRP
    Search the HRT5500-TABNR using purchasing group from R/3.
    Search the HRP5500 – OBJID using HRT5500-TABNR
    Call the FM:  BBP_OM_DETERMINE_RESP_PGRP
    Import parameter: IS_RESP_ITEM_DATA – PROC_GROUP – OBJID
    Export parameter: ES_PURCH_DATA
    Thank you
    Best Regards
    SH

  • I just updated Firefox to 6.0, and suddenly certain parts of the text which used to be in English now display in Devanagari, the font used in Hindi. How do I get it back to English?

    I just updated Firefox to 6.0, and suddenly certain parts of the text which used to be in English now display in Devanagari, the font used in Hindi. How do I get it back to English?

    You can do a check for corrupted fonts and other font issues:
    * http://www.creativetechs.com/iq/garbled_fonts_troubleshooting_guide.html - CreativeTechs Tips: Garbled Fonts Troubleshooting Guide
    * http://www.thexlab.com/faqs/multipleappsquit.html - Font Book 2.0 Help: Checking for damaged fonts

  • Map the user exit variables with the queries/cubes using them

    Hello Friends,
    What are the post unicode conversion tests that you can perform on front end/existing queries?
    is there any added advantage for queries due to unicode?
    is testing the working of variable  enough? is there any table to map the user exit variables with the queries/cubes using them?
    Thanks
    Tanya

    Guys, any clue about this? Answer are appreciated.
    Thanks
    Tanya

  • HT1349 i lost my iphone... if somebody use it my i phone with different apple ID, Do apple company can determine the apple ID used on my iphone serial number?? If i gave my serial number, can it retrieve the apple ID user???

    i lost my iphone... if somebody use it my i phone with different apple ID, Do apple company can determine the apple ID used on my iphone serial number?? If i gave my serial number, can it retrieve the apple ID user???

    No! Apple is not law enforcement and therefore does not have personal investigating lost or stolen property other than Apple's own.
    Even if they had the information data protection laws would prevent them from giving it to you or anybody else including law enforcement.

  • I have an ipod I want to sell, but I don't know how to determine the specs, which one it is, how much memory, etc. Where can I find this information?

    I have an ipod I want to sell, but I don't know how to determine the specs, which one it is, how much memory, etc. Where can I find this information?

    See this Apple support document.
    http://support.apple.com/kb/ht1353
    B-rock

  • I lost my phone 2 days ago, can I get the phoneNumber which use my phone by IMEI or Serifal NO  .  I am in China.

    I lost my phone 2 days ago, can I get the phoneNumber which use my phone now by IMEI or Serifal NO  .  I am in China.

    nothing unless you had "find my iPhone" installed before it was lost.

  • How to determine the "Last Value" in exception aggregation

    Hi gurus,
    I have a KeyFigure with an exception aggregation -Last value- by time.
    I report on a multiprovider ("containing" a infocube and a realtime infocube with BPS-InputLayout). I can correct separate Items with this BPS-Layout. So, at the end I have 2 values for one item.
    So, by these facts, my report shows the the last value for the specific KeyFigure, but I expected that both values (the value in the infocube and the correction via BPS)are sumed up.
    As this is due to the exception aggregation not the case, I could "overwrite" the normal value in the infocube with the BPS.
    But SAP determines the normal value in the Infocube as "Last Value", not the one in the correction cube, which is entered via BPS.
    How does SAP determine the "last value" as I can´t see any date...
    Thanks in advance!

    Thank you for your answer...
    I got the soltion by myself.
    I thought, Last Value means the last Value, depending on the time, but in my opinion the function last value should be called "Greatest Value".
    All figures with the same value for the reference characteristc are aggregated anyway.

  • Fine tunning the query which uses /*+ USE_CONCAT */ table hint

    Some of the queries in our application uses the /*+ USE_CONCAT */ table hint which takes huge time for the execution in Oracle database.when we fire the below query in 250Million production database it takes approx.3min for the execution.
    Because of this we are facing a performance issue in the application.
    Below is the sample query :
    SELECT /*+ USE_CONCAT */ * FROM DI_MATCH_KEY WHERE NORM_COUNTRY_CD = 'US'
    AND ((( NORM_CONAME_KEY1 ='WILM I' OR NORM_CONAME_KEY2 = 'WILM I' OR NORM_CONAME_KEY23 = 'WILM I'
    OR NORM_CONAME_KEYFIRST ='WILLIAM' ) AND NORM_STATE_PROVINCE = 'CA' ) OR NORM_ADDR_KEY2 = 'CALMN 12 3 OSAI')
    Regarding the indexes almost for all the columns combination for the above sql index already created in table.
    Your suggestions will be appreciated
    Thanks,
    Regards,
    Krishna kumar

    Hi,
    Thanks for our valuable inputs.
    As suggested find attached explain plan and trace file details in the excel sheet.
    TRACE FILE:
    <pre>
    TKPROF: Release 10.2.0.1.0 - Production on Thu Feb 14 11:11:59 2008
    Copyright (c) 1982, 2005, Oracle. All rights reserved.
    Trace file: implhr01_ora_26457.trc
    Sort options: prsela exeela fchela
    count = number of times OCI procedure was executed
    cpu = cpu time in seconds executing
    elapsed = elapsed time in seconds executing
    disk = number of physical reads of buffers from disk
    query = number of buffers gotten for consistent read
    current = number of buffers gotten in current mode (usually for update)
    rows = number of rows processed by the fetch or execute call
    SELECT /*+ USE_CONCAT */ COUNT(*) FROM DI_MATCH_KEY WHERE NORM_COUNTRY_CD = 'US'
    AND ((( NORM_CONAME_KEY1 ='WILM I' OR NORM_CONAME_KEY2 = 'WILM I' OR NORM_CONAME_KEY23 = 'WILM I'
    OR NORM_CONAME_KEYFIRST ='WILLIAM' ) AND NORM_STATE_PROVINCE = 'CA' ) OR NORM_ADDR_KEY2 = 'CALMN 12 3 OSAI')
    call count cpu elapsed disk query current rows
    Parse 1 0.03 0.05 0 0 0 0
    Execute 1 0.00 0.00 0 0 0 0
    Fetch 2 3.97 65.43 30633 64053 0 1
    total 4 4.00 65.49 30633 64053 0 1
    Misses in library cache during parse: 1
    Optimizer mode: ALL_ROWS
    Parsing user id: 76
    Rows Row Source Operation
    1 SORT AGGREGATE (cr=64053 pr=30633 pw=0 time=65436584 us)
    73914 CONCATENATION (cr=64053 pr=30633 pw=0 time=61623673 us)
    0 TABLE ACCESS BY INDEX ROWID DI_MATCH_KEY (cr=4 pr=2 pw=0 time=32617 us)
    0 INDEX RANGE SCAN MKADDR_KEY2 (cr=4 pr=2 pw=0 time=32599 us)(object id 122583)
    75 TABLE ACCESS BY INDEX ROWID DI_MATCH_KEY (cr=80 pr=20 pw=0 time=1369427 us)
    75 INDEX RANGE SCAN MK_CY_KEY1_ST_PROV (cr=5 pr=0 pw=0 time=2119 us)(object id 122666)
    2 TABLE ACCESS BY INDEX ROWID DI_MATCH_KEY (cr=81 pr=2 pw=0 time=26723 us)
    77 INDEX RANGE SCAN MK_CY_KEY2_ST_PROV (cr=4 pr=0 pw=0 time=3641 us)(object id 122667)
    2 TABLE ACCESS BY INDEX ROWID DI_MATCH_KEY (cr=6 pr=0 pw=0 time=47 us)
    2 INDEX RANGE SCAN MK_CY_KEY1_AGN (cr=4 pr=0 pw=0 time=28 us)(object id 122670)
    73835 TABLE ACCESS BY INDEX ROWID DI_MATCH_KEY (cr=63882 pr=30609 pw=0 time=61503773 us)
    73905 INDEX RANGE SCAN MK_CY_KEY2_AGN (cr=266 pr=176 pw=0 time=148198 us)(object id 122745)
    alter session set sql_trace true
    call count cpu elapsed disk query current rows
    Parse 0 0.00 0.00 0 0 0 0
    Execute 1 0.00 0.00 0 0 0 0
    Fetch 0 0.00 0.00 0 0 0 0
    total 1 0.00 0.00 0 0 0 0
    Misses in library cache during parse: 0
    Misses in library cache during execute: 1
    Optimizer mode: ALL_ROWS
    Parsing user id: 76
    OVERALL TOTALS FOR ALL NON-RECURSIVE STATEMENTS
    call count cpu elapsed disk query current rows
    Parse 1 0.03 0.05 0 0 0 0
    Execute 2 0.00 0.00 0 0 0 0
    Fetch 2 3.97 65.43 30633 64053 0 1
    total 5 4.00 65.50 30633 64053 0 1
    Misses in library cache during parse: 1
    Misses in library cache during execute: 1
    OVERALL TOTALS FOR ALL RECURSIVE STATEMENTS
    call count cpu elapsed disk query current rows
    Parse 9 0.05 0.03 0 0 0 0
    Execute 23 0.05 0.08 0 0 0 0
    Fetch 98 0.01 0.01 0 137 0 76
    total 130 0.11 0.13 0 137 0 76
    Misses in library cache during parse: 9
    Misses in library cache during execute: 9
    2 user SQL statements in session.
    23 internal SQL statements in session.
    25 SQL statements in session.
    Trace file: implhr01_ora_26457.trc
    Trace file compatibility: 10.01.00
    Sort options: prsela exeela fchela
    1 session in tracefile.
    2 user SQL statements in trace file.
    23 internal SQL statements in trace file.
    25 SQL statements in trace file.
    11 unique SQL statements in trace file.
    561 lines in trace file.
    91 elapsed seconds in trace file.
    </pre>
    EXPLAIN PLAN:
    <pre>
    ID     OPERATION     OBJECT_NAME     CARDINALITY     BYTES     COST     CPU_COST
    0     SELECT STATEMENT          1     53     1886     15753172
    1     SORT          1     53          
    2     CONCATENATION                         
    3     TABLE ACCESS     DI_MATCH_KEY     8     424     10     78334
    4     INDEX     MKADDR_KEY2     8          4     30086
    5     TABLE ACCESS     DI_MATCH_KEY     3     159     7     53574
    6     INDEX     MK_CY_KEY1_ST_PROV     3          4     29286
    7     TABLE ACCESS     DI_MATCH_KEY     3     159     7     53740
    8     INDEX     MK_CY_KEY2_ST_PROV     3          4     29286
    9     TABLE ACCESS     DI_MATCH_KEY     5     265     7     55941
    10     INDEX     MK_CY_KEY1_AGN     5          4     29686
    11     TABLE ACCESS     DI_MATCH_KEY     2192     116176     1855     15511582
    12     INDEX     MK_CY_KEY2_AGN     2228          12     524136
    </pre>
    Kindly help us regarding this.
    Thanks,
    Krishna Kumar.

  • How to determine the IPS throughput using Cisco ASA 5500 IPS Solution?

    Hello there!
    I´ve been desinging a solution to protect de Server Farm and I intend to use the ASA 5500 series with AIP-SSM module. There´s any tool to determine the real throughput that I need? I mean, how to determine the performance (Firewall + IPS  throughput), what main points I should consinder?

    If the server farm is running production levels of traffic today you can get statistics off a variety of networking devices passing the existing traffic. Switches, routers and firewalls all count every byte of traffic they pass. There are plenty of tools that can gather this traffic into tables via SNMP too, such as MRTG.
    Do not average your traffic over too great a time peroid, you will miss busy hour peaks. At most, use 5 min averages.
    - Bob

  • How would you determine the total ram used by applets?

    hi guys;
    I have 2 applets.
    One applet runs in a browswer. - in the microsoft jvm.
    Another applet on another browser (or another page) - in the sun plugin.
    How would you guys determine the total RAM (not just heap memory) used by these applets.
    my strategy would be to open up the browser. and look at the memory next to that internet explorer process. Then I would open up another browser with applet and see total memory taken. Then I would subtract to determine the difference.
    Does this sound correct?
    Each applet has multiple jar files associated with it.
    Stev

    What you really should do is test your applet on a bunch of machines, like a 32 meg ram win 98, 64 meg ram win NT, and a 256 meg ram win 2000, and then give them a performance report under the different configurations. That's really what they're after (probably)

  • How to activate the queries which are inactive

    Hello All,
    1) After BW upgradation from 3.5 to 7.0 we found some of the queries are inactive. Can anyone let me know how to activate those inactive queries.
    2) How to find what all roles are inactive nd how to activate them.

    Hi,
    After upgrade or client copy you'll need to set the release indicator for queries (prog. RSAQUM40), in general it is only possible to transport those queries that are created in the global area. For conversion of queries refer OSS notes 92124/109094/119655.
    However if you wish to move queries from the standard area use program RSAQR3TR, execute the program via TX SA38 then click in the blue "i" (information button) for full details of the program.
    Best Regards,
    Chandrasekhar

Maybe you are looking for