Query execution not traced in ST03N

Dear performance gurus,
we want to perform some query performance checks in our BW 3.5 system. To measure the impact of aggregates and eliminate the OLAP cache effect on runtime we execute the queries from the query monitor (rsrt). With "execute and debug" we can decide for each execution if we want to use agrregates or not and we can switch of OLAP cache usage.
For what ever reason these executions are not shown in ST03N (Expert mode). If we run the same queries from BEx Analyzer, runtime can be analyzed. But in BEx we don't have the necessary options regarding aggregates and OLAP cache.
Are you aware of any settings in RSRT or ST03N to get the executions out of the query monitor also displayed in ST03N??
Thanks in advance
Thomas

while running the report user is facing the below error.
"query execution was not successful"This is a generic error which is reported in few docs.
Discoverer Viewer fails Moving Pivot Table Columns: Query Execution was not Successful [ID 948027.1]
Query Execution Not Successful Error In Viewer, Runs In Desktop And Plus [ID 404974.1]
Discoverer 10g (10.1.2.3) Plus/Viewer Cumulative Patch 7 (9112482) Readme For Linux/Unix [ID 821844.1]
Launching A Worksheet In Discoverer 10g / 11g Plus/Viewer On Linux Fails With 'Contact with backend server lost' [ID 871012.1]
Discoverer 10g (10.1.2.3) Plus/Viewer Cumulative Patch 4 (7595032) Readme For Windows [ID 822183.1]
Discoverer 10g (10.1.2.3) Plus/Viewer Cumulative Patch 3 (7319096) Readme For Linux/Unix [ID 761997.1]
Running Some Reports In Discoverer Plus/Viewer 10.1.2.2 Are Failing With "An error occurred while attempting to perform the operation. The operation did not complete successfully." [ID 733603.1]
Discoverer Viewer 10g (10.1.2.3) Passes A Blank Parameter As 'NULL' When Using A 'Drill to Link' [ID 820003.1]
Query Execution was not Successful Error When Running a Workbook [ID 550684.1]
-ORA-01722 Invalid number.Please check the data type of the column or the data type of the records in the table/view and make sure it match the data type of the column.
Thanks,
Hussein

Similar Messages

  • LIKE causing delay in query execution

    Hi,
    I have this query that takes about 20 mins to execute. Is there a way to tune this, especially for the LIKE where clauses?
    SELECT product_id, territory_code, global_target_id,
    DECODE (service_type, 'SERVICE', 'Y', 'N') service_type,
    equipment_deployment, created_date, updated_date,
    fiscal_period_id, customer_id,
    SUM (bookings_amount) bookings_amount, ship_to_country,
    alliance_id
    FROM sp_refresh_incre_vw2@cdw
    WHERE fiscal_period_id = in_chr_fiscal_period_id
    AND ( ec_name LIKE in_chr_search_string --- culprit
    AND NOT st_name LIKE in_chr_search_string --- culprit
    AND NOT bt_name LIKE in_chr_search_string --- culprit
    GROUP BY fiscal_period_id,
    product_id,
    territory_code,
    ship_to_country,
    customer_id,
    global_target_id,
    service_type,
    equipment_deployment
    HAVING SUM (bookings_amount) <> 0
    UNION
         SELECT product_id, territory_code, global_target_id,
    DECODE (service_type, 'SERVICE', 'Y', 'N') service_type,
    equipment_deployment, created_date, updated_date,
    fiscal_period_id, customer_id,
    SUM (bookings_amount) bookings_amount, ship_to_country,
    alliance_id
    FROM sp_refresh_incre_vw2@cdw
    WHERE fiscal_period_id = in_chr_fiscal_period_id
    AND bt_name LIKE in_chr_search_string --------- culprit
    GROUP BY fiscal_period_id,
    product_id,
    territory_code,
    ship_to_country,
    customer_id,
    global_target_id,
    service_type,
    equipment_deployment
    HAVING SUM (bookings_amount) <> 0
    UNION
    SELECT product_id, territory_code, global_target_id,
    DECODE (service_type, 'SERVICE', 'Y', 'N') service_type,
    equipment_deployment, created_date, updated_date,
    fiscal_period_id, customer_id,
    SUM (bookings_amount) bookings_amount, ship_to_country,
    alliance_id
    FROM sp_refresh_incre_vw2@cdw
    WHERE fiscal_period_id = in_chr_fiscal_period_id
    AND ( st_name LIKE in_chr_search_string -------- culprit
    AND NOT bt_name LIKE in_chr_search_string -------- culprit
    GROUP BY fiscal_period_id,
    product_id,
    territory_code,
    ship_to_country,
    customer_id,
    global_target_id,
    service_type,
    equipment_deployment
    HAVING SUM (bookings_amount) <> 0;
    After some investigation, I have noted that it is the LIKE clause that is causing the delay. The three queries are exactly similar EXCEPT for the LIKE clause, which I have commented on with "--- culprit"
    Any alternative ideas to achieve a similar result by avoiding the LIKE clause will be very much appreciated.
    Thanks and regards,
    Ambili

    Hi,
    .I Accelerator BI accelerator is a separate hardware which when attached to BW server creates an index on data of infocube and stores in compressed format.Different technologies are used for compression and horizontal pratitioning of data. BI Accelerator can benefit businesses that have high volumes of data.
    This link explains you better on BI Accelerator
    What is BI Accelerator
    Comming to the query performance, Using BIAccelerator, query performance is highly improved.
    There is o ideal time for the query execution time.
    The query execution time depends on various factors, to list some of them,
    1)Complexity of the query(more complex, more execution time)
    2)Object on which the query is built (If it is built on CUbe, Performance will be more, on Dso performance is poor)
    3)Building indexes on cube increases performance....
    4)Data volume in the report.
    But yes, We can look at the stats of query execution using ST03 and ST03N .
    Hope the above answer helps you.
    Please let me know if you need any more clarifications

  • Query was not successful in Discoverer viewer

    Hi All,
    can you please guide me on below issue..
    while running the report user is facing the below error.
    "query execution was not successful"
    -ORA-01722 Invalid number.
    Ebiz:12.0.6
    DB:11.2.0.2
    Disco:10.1.2.55.26
    Thanks in Advance.
    Regards
    Kumar

    while running the report user is facing the below error.
    "query execution was not successful"This is a generic error which is reported in few docs.
    Discoverer Viewer fails Moving Pivot Table Columns: Query Execution was not Successful [ID 948027.1]
    Query Execution Not Successful Error In Viewer, Runs In Desktop And Plus [ID 404974.1]
    Discoverer 10g (10.1.2.3) Plus/Viewer Cumulative Patch 7 (9112482) Readme For Linux/Unix [ID 821844.1]
    Launching A Worksheet In Discoverer 10g / 11g Plus/Viewer On Linux Fails With 'Contact with backend server lost' [ID 871012.1]
    Discoverer 10g (10.1.2.3) Plus/Viewer Cumulative Patch 4 (7595032) Readme For Windows [ID 822183.1]
    Discoverer 10g (10.1.2.3) Plus/Viewer Cumulative Patch 3 (7319096) Readme For Linux/Unix [ID 761997.1]
    Running Some Reports In Discoverer Plus/Viewer 10.1.2.2 Are Failing With "An error occurred while attempting to perform the operation. The operation did not complete successfully." [ID 733603.1]
    Discoverer Viewer 10g (10.1.2.3) Passes A Blank Parameter As 'NULL' When Using A 'Drill to Link' [ID 820003.1]
    Query Execution was not Successful Error When Running a Workbook [ID 550684.1]
    -ORA-01722 Invalid number.Please check the data type of the column or the data type of the records in the table/view and make sure it match the data type of the column.
    Thanks,
    Hussein

  • An error has occurred during report processing. (rsProcessingAborted). Query execution failed for dataset 'DimUserWorkCentre'. (rsErrorExecutingCommand). The Cube either does not exists or has not been processed

    Hi,
    I'm having issues with the report created using SSAS cube.
    An error has occurred during report processing. (rsProcessingAborted)
    Query execution failed for dataset 'DimUserWorkCentre'. (rsErrorExecutingCommand)
    The Operator_Performance cube either does not exist or has not been processed.
    I have searched through internet and tried all the solutions, but didn't worked for me.
    SSRS services running as NETEWORK SERVICE user.
    SSRS Execution running as a Different user, which is the login is used to logon ot that server. I have also verified this user has access to database. I'm using Shared DataSource(SSAS Source) for this report.
    Can any one please help me.
    Thank You,
    Praveen.
    Praveen

    Hello,
    Have you tried it to execute on report manager , Is your data source properly configured in Report Manager  and your report is mapped with Datset correctly?
    Have you executed the Dataset query MDX editor  now?
    What is the volume of data you are fetching in the report. Try it to execute in other than IE , I don't know the exact reason but some of our report with large volume of data  are failing on IE , on the other hand these reports are running fine Google
    Chrome
    blog:My Blog/
    Hope this will help you !!!
    Sanjeewan

  • In-Memory Query execution mode rendering results but "both" as execution mode do not.

    Hi,
    I am facing this strange issue of fetching some rows based on my view criteria.
    I delete some rows and make them as inactive. Then I search for those rows by using filter as "inactive" rows. I get the result as expected. But then, if I enter any other filed apart from inactive condition, I do not get that record even when that record is present in the list.
    I looked out for the query execution mode for my VC based on which my rows are fetched. I found that if I keep the Query execution mode as "In-Memory" I am able to get the result but if I give the query execution mode as "Both", I do not get results even though the other factors are still same for it.
    Due to model settings, I can't use "In-Memory" as the query execution mode.
    I get the results properly after I commit my changes to DB,however.
    How can I proceed here?

    Hi,
    apparently the JDeveloper version you use does not work with "Both", which I suspect to be a bug. If you are on a latest version of JDeveloper 11g R1, JDeveloper 11g R2 or on JDeveloper 12c, please file a bug if you have access to customer support. If you are not yet on a latest version of JDeveloper, please try a newer version of your JDeveloper release.
    Frank

  • Data Driven Subscriptions Error - the query processor could not start the necessary thread resources for parallel query execution

    Hi,
    We are getting the following error when certain data driven subscriptions are fired off: "the query processor could not start the necessary thread resources for parallel query execution".  I've read other posts that have the same error, and
    the solution usually involves adjusting MaxDOP to limit the number of queries that are fired off in parallel.  
    Unfortunately, we cannot change this setting on our server for performance reasons (outside of data driven subscriptions, it negatively impacts our ETL processing times).  We tried putting query hints like "OPTION (MAXDOP 2);" in the reports
    that are causing the error, but it did not resolve the problem.
    Are there any settings within Reporting Services that can be adjusted to limit the number of subscriptions that get fired off in parallel?
    Any help is appreciated - thanks!

    Yes, that is correct.  It's a painful problem, because you don't know which specific subscription failed. For example, we have a data driven subscription that sends out about 800 emails. Lately, we've been having a handful of them fail. You don't know
    which ones out of the 800 failed though, even from the RS log files - all it tells you is that "the
    query processor could not start the necessary thread resources for parallel query execution".
    Thanks, I'll try changing <MaxQueueThreads> and will let you know if it works.
    On a side note: I've noticed that it is only reports with cascading parameters (ex. where parameter 2 is dependent on the selection from parameter 1) that get this error message...

  • Query execution was not successful in discoverer viewer

    hi all,
    disco viewer is showing an error like "Query execution was not successful. -Failed to connect to database" when i ran a report(in which rank function is included), where the same report is running fine is disco desktop. am just wondering are there any settings to be changed in desktop to run like these reports in disco viewer?
    ur help in this regard wil gr8ly appreciated.
    thanks very much in advance.
    _Rakesh Reddy.                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                           

    hi,
    thanks very much for your reply.
    the problem is with rank function it seems, its opening in disco viewer & running in disco plus also when i remove that rank calculation. but its not even running is disco plus with that rank function.
    am using 10.1.2.48.18 version of disco desktop with 10g DB.
    thanks.

  • How can I reduce BEx Query execution time

    Hi,
    I have a question regarding query execution time in BEx.
    I have a query that takes 45 mins to 1 hour to execute in BEx analyser. This query is run on a daily basis and hence I am keen to reduce the execution time.  Are there any programs or function modules that can help in reducing query execution time?
    Thanks and Regards!

    Hi Sriprakash,
    1.Check if your cube is performance tuned: in the manage cube from RSA1 / performance tab: check if all indexes and statistics are green. Aggregate IDx should as well be.
    2.Condense your cubes regularly
    3. Evaluate the creation of an aggregate with all characteristic used in the query (RSDDV).
    4.Evaluate the creation of a "change run aggregate": based on a standalone NavAttr (without its basic char in the aggr.) but pay attention to the consequent change run when loading master data.
    5. Partition (physically) your cubes systematically when possible (RSDCUBE, menu, partitioning)
    6. Consider logical partitioning (by year or comp_code or ...) and make use of multiproviders in order to keep targets not too big...
    7.Consider creating secondary indexes when reporting on ODS (RSDODS)
    8.Check if the query runtime is due the master data read or the infoprovider itself, or the OLAP processor and/or any other cause in tx ST03N
    9.Consider improving your master reads by creating customized IDX (BITMAP if possible and depending on your data) on master data table and/or attribute SIDs when using NAvs.
    10.Check that your basis team did a good job and have applied the proper DB parameters
    11.Last but not least: fine tune your datamodel precisely.
    hope this will give you an idea.
    Cheers
    Sunil

  • BW statistics: Query Execution Time per User

    I am struggling to get something meaningful out of the BW statistics. The only information I really need is:
    The time from the first call of a query until the first selection screen and
    The time from the execution request until the data is displayed (i.e. after filling selection screen)
    With the possibility to see this information by user.
    So with other words:  KF Duration of query call + KF Duration of query execution, with Characteristics: user, period, query
    This could sound like a very simple universal query, but indeed absolutely not straightforward when looking at statistics contents. Any hints on the KF calculations are most welcome.
    The BW version in 7.0, with Bex as well as WAD reporting.

    I was looking for the same KFs, but did not find anything meaningful in ST03N as well.
    But I would recommend to execute the Web Report (from WAD or from Portal, if you are using Enterprise Portal) and then look at table view RSDDSTAT_OLAP. In this table view, BW Statistic data for each step is saved on event level. So if you execute the Web Report, you will find entries for the step corresponding to displaying the selection screen and for the step of generating the Web Report after the selection screen. The column RUNTIME displays the runtime of each step, which is a good indicator for the average runtime KFs you want to have.
    Maybe there are more meaningful KFs in technical BI Content, but I did not find them yet.

  • How to make an index use in a query execution

    Hi,
    I have the below query for which ename column has an index. As of my knowledge below queries 1st and 2st will not use index. Hence i used the 3rd statement and that too its not using the index. Finally i used the 4th query, but even the 4th query is not using the index. Then how do i make this query to use my index??? Do i need to create a function based index for this??? Is that the final option????
    1. select * from emp where ename !='BH' ;
    2. select * from emp where ename <> 'BH';
    3. select * from emp where ename not in ('BH');
    4. select * from emp where ename < 'BH' or ename > 'BH';
    Regards,
    007
    Edited by: 007 on Jun 6, 2013 7:56 AM
    Edited by: 007 on Jun 6, 2013 8:06 AM
    Edited by: 007 on Jun 6, 2013 8:06 AM
    Edited by: 007 on Jun 6, 2013 8:06 AM
    Edited by: 007 on Jun 6, 2013 8:12 AM

    Sorry 007, I really thought you were posting a trick question as on the OCP tests.
    Anyway, as Justin mentioned, if you have an index on ename, it may be used when doing a comparison predicate statement with the ename value.
    What it depends on are several other things: stats, how many rows in the table, use of an index hint, etc.
    Rather than questioning the group on this, why not just turn on autotrace and run the query for the different scenarios.
    The output will show you if it used the index, number of rows returned, blocks read, etc.
    SQL> create table emp (ename  varchar2(40));
    Table created.
    SQL> insert into emp select username from sys.dba_users;
    25 rows created.
    SQL> commit;
    Commit complete.
    SQL> set autotrace on
    SQL> select * from emp where ename != 'SYSTEM';
    Execution Plan
    Plan hash value: 2951343571
    | Id  | Operation        | Name      | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | SELECT STATEMENT |           |    24 |   528 |     1   (0)| 00:00:01 |
    |*  1 |  INDEX FULL SCAN | ENAME_IDX |    24 |   528 |     1   (0)| 00:00:01 |
    Predicate Information (identified by operation id):
       1 - filter("ENAME"<>'SYSTEM')As you can see, the above used an index, even though there were only 25 rows in the table.
    You can test each of your scenarios, one by one, including use of a hint.

  • SSRS Error : Query execution failed for dataset 'dataSet'. (rsErrorExecutingCommand) Semantic query execution failed. Invalid object name 'RPT.*********'. (rsSemanticQueryEngineError)

    I am new to SSRS and I am trying to migrate reports from 2008 to 2012. As I have so many reports to migrate, I simply got the back up of ReportServer,
    ReportServerTempDB, and Encryption Key and restored them to test environment. I made necessary configuration from RS configuration tool. I am able to see the reports now when I browse //hostname/reports. But when I open any particular report I am getting some
    error.
    · An error has occurred during report processing.
    (rsProcessingAborted)
    Query execution       failed for dataset 'dataSet'.
          (rsErrorExecutingCommand
    Semantic query        execution failed. Invalid object name
           'RPT. ******'. (rsSemanticQueryEngineError)
    ****** - I am assuming this is a custom data class.
    Does anyone have insight on this? or any better way that I can migrate the reports to new server with less efforts.
    I don’t have the reports solution file to deploy the reports, so I have followed backup and restore process.

    Hi Kishore237,
    According to your description, you migrated some reports from Reporting Services (SSRS) 2008 to 2012. Now you get error when accessing the reports on SSRS 2012. Right?
    In this scenario, did you modify the report data source in database after migration? You can try to open the report in Report Builder or Report designer and check the report dataset. If you can preview the report in Report builder or Report designer,
    please try to redeploy the report to Report Server. If it is still not working, please try to restore the database from backup. And for migrating reports, please follow the "Content-Only Migration" in the link below:
    http://msdn.microsoft.com/en-us/library/ms143724(v=sql.110).aspx
    If you have any question, please feel free to ask.
    Best Regards,
    Simon Hou

  • Unable to select the filter value after query execution

    hi,
    I am unable to drill down my keyfigures.
    1.i have a lC and GC values out of which GC is hidden so after the query execution i would like to filter my value between GC and LC but i do get only LC value in the filter and i do not get GC value to select .
    Could any one tell me how this can be done.this is very urgent.
    Thx
    Subha

    resolved on my own,so i am closing this.
    Thx
    Subha

  • Asset query execution performance after upgrade from 4.6C to ECC 6.0+EHP4

    Hi,guys
    I am encounted a weird problems about asset query execution performance after upgrade to ECC 6.0.
    Our client had migrated sap system from 4.6c to ECC 6.0. We test all transaction code and related stand report and query.
    Everything is working normally except this asset depreciation query report. It is created based on ANLP, ANLZ, ANLA, ANLB, ANLC table; there is also some ABAP code for additional field.
    This report execution costed about 6 minutes in 4.6C system; however it will take 25 minutes in ECC 6.0 with same selection parameter.
    At first, I am trying to find some difference in table index ,structure between 4.6c and ECC 6.0,but there is no difference about it.
    i am wondering why the other query reports is running normally but only this report running with too long time execution dump messages even though we do not make any changes for it.
    your reply is very appreciated
    Regards
    Brian

    Thanks for your replies.
    I check these notes, unfortunately it is different our situation.
    Our situation is all standard asset report and query (sq01) is running normally except this query report.
    I executed se30 for this query (SQ01) at both 4.6C and ECC 6.0.
    I find there is some difference in select sequence logic even though same query without any changes.
    I list there for your reference.
    4.6C
    AQA0FI==========S2============
    Open Cursor ANLP                                    38,702  39,329,356  = 39,329,356      34.6     AQA0FI==========S2============   DB     Opens
    Fetch ANLP                                         292,177  30,378,351  = 30,378,351      26.7    26.7  AQA0FI==========S2============   DB     OpenS
    Select Single ANLC                                  15,012  19,965,172  = 19,965,172      17.5    17.5  AQA0FI==========S2============   DB     OpenS
    Select Single ANLA                                  13,721  11,754,305  = 11,754,305      10.3    10.3  AQA0FI==========S2============   DB     OpenS
    Select Single ANLZ                                   3,753   3,259,308  =  3,259,308       2.9     2.9  AQA0FI==========S2============   DB     OpenS
    Select Single ANLB                                   3,753   3,069,119  =  3,069,119       2.7     2.7  AQA0FI==========S2============   DB     OpenS
    ECC 6.0
    Perform FUNKTION_AUSFUEHREN     2     358,620,931          355
    Perform COMMAND_QSUB     1     358,620,062          68
    Call Func. RSAQ_SUBMIT_QUERY_REPORT     1     358,569,656          88
    Program AQIWFI==========S2============     2     358,558,488          1,350
    Select Single ANLA     160,306     75,576,052     =     75,576,052
    Open Cursor ANLP     71,136     42,096,314     =     42,096,314
    Select Single ANLC     71,134     38,799,393     =     38,799,393
    Select Single ANLB     61,888     26,007,721     =     26,007,721
    Select Single ANLZ     61,888     24,072,111     =     24,072,111
    Fetch ANLP     234,524     13,510,646     =     13,510,646
    Close Cursor ANLP     71,136     2,017,654     =     2,017,654
    We can see first open cursor ANLP ,fetch ANLP then select ANLC,ANLA,ANLZ,ANLB at 4.C.
    But it changed to first select ANLA,and open cursor ANLP,then select  ANLC,ANLB,ANLZ,at last fetch ANLP.
    Probably,it is the real reason why it is running long time in ECC 6.0.
    Is there any changes for query selcection logic(table join function) in ECC 6.0.

  • Select query in not working for Count(*)

    Hi,
    Our batch team running one query that is selecting one table TSFHEAD and this query is hanging. Below are diffrent shenario in which the select query is running on this table.
    select * from tsfhead where create_id = 'BATCH' and create_date = '26-OCT-12';
    --not returning any rows and hanging
    select * from tsfhead where create_id = 'BATCH' ---returning rows
    select * from tsfhead where create_date = '26-OCT-12'; --- returning rows
    select count(*) from tsfhead ----not returning rows and hanging
    This table TSFHEAD has 59000 rows.
    SQL> explain plan for select * from tsfhead where create_id = 'BATCH' and create_date = '26-OCT-12';
    Explained.
    SQL> SELECT * FROM TABLE(dbms_xplan.display);
    PLAN_TABLE_OUTPUT
    Plan hash value: 415503093
    | Id | Operation | Name | Rows | Bytes | Cost (%CPU)| Time |
    | 0 | SELECT STATEMENT | | 1 | 67 | 215 (1)| 00:00:03 |
    |* 1 | TABLE ACCESS FULL| TSFHEAD | 1 | 67 | 215 (1)| 00:00:03 |
    Predicate Information (identified by operation id):
    PLAN_TABLE_OUTPUT
    1 - filter("CREATE_ID"='BATCH' AND "CREATE_DATE"='26-OCT-12')
    I don't know why this query behaving in this manner.Please help.

    RanVijai_dba wrote:
    Thanks for your reply. But +"select count(*) from tsfhead"+ is hanging and +"select * from tsfhead where create_date = '26-OCT-12' "+ is showing records.It might be showing records, but not necessarily the correct records.
    As sb points out, you are treating dates as strings rather than the DATE datatype. That means you could be querying the wrong data, and your query is also not safe in different environments. It could also effect the query execution plan. Also, as pointed out, you should be ideally using 4 digit years. There were many headaches caused by the use of 2 digit years that most companies corrected as part of the millenium bug fixes prior to the year 2000... well over a decade ago, and most good designs now ensure that 4 digit years are used as standard.
    So your query would be better written as:
    select * from tsfhead where create_date = TO_DATE('26-OCT-2012','DD-MON-YYYY')When you say:
    select count(*) from tsfhead... is hanging, you say the table has around 59000 rows in it, and in reality that's a small amount of records, so a count(*) shouldn't take long at all, even doing a full table scan.
    Post the explain plan for that simple count(*) query for us. (and ensure you use {noformat}{noformat} tags to keep the formatting on the forum, as described in {message:id=9360002})                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                   

  • Characteristic "Master Data" setting in Query definition not working?

    Hi,
    I am having a problem with a query definition. (BI 7.0)
    The query comprises the following definition which relate to displaying Milestone dates for respective projects in a time series so the dates are populated in the correct column in a time series:
    Filter
    Project Definition = Fixed Single Project Definition
    Fiscal Year Variant = K4
    Project Profit Centre = Hierarchy Node
    Project Plant = Fixed Single Value
    Free Chars
    Network
    Network Activity
    Activity Element
    Rows
    Project Plant
    Project Profit Centre (With hierarchy)
    Project Definition
    Milestone Type  (Setting in characteristic to pick up Master Data)
    CSR Relevant    (Setting in characteristic to pick up Master Data)
          Structure
               Actual
               Scheduled
        (Actual:  Value type = Actual,  Origin = Manual ,Event = Start)
        (Scheduled:  Value type = Actual,  Origin = Scheduled ,Event = Start)
    Columns
    Key Figure structure of 12 key figures in a time series
    Selection
    Key Figure = Date
    Fiscal Year/Period = Current Fiscal Year / Period (SAP EXIT) (With Offset 1 to 12)
    FYI
    When I run the above it comes back correctly. However when I add "MILESTONE" (Characteristic setting (Posted Values) to the Query Definition immediately after the "Milestone Type" and before CSR it does not work it just is hanging for a considerable time.
    I am not using "Master Data" for the "Access Type for Result Vales" for the additional characteristic "MILESTONE" just "Posted Values".
    I was expecting that it would just show all Milestone types as per Master Data but just the Milestones that are posted for this result.
    I would appreciate a solution to this issue and an explanation to why it hangs when I add the "MILESTONE" characteristic to the definition as explained above.
    Thanks in advance..
    Stevo

    Hello there,
    Putting Milestone in the rows in the query definition is not being hang there (i.e., while creating and saving the query) right?
    It is stucked while executing the query isn't it so?
    That's because you have too many records to be desaggregated by milestone characteristic, and this is indepently of posted values in the characteristic settings.
    Putting milestone characteristic in the free characteristics will allow you to run the query, but desaggregate it (add it to the rows) will be to much information to read from your dataprovider (this is just a guess).
    If it is being stucked in query execution whike milestone is in the rows. Try this for example:
    Add milestone to the free chars. Execute the query. Filter milestone characteristic while still in the free rows and choose a value that you are pretty sure has the smallest records in the Dataprovider (could be for example value #). Than drill down milestone to the rows. You'll see that it can be executed without any problem.
    Diogo.

Maybe you are looking for