Copied query running indefinitely.

Dear All,
I am facing a strange issue. I have a query say A. On execution of the query it works fine and displays the results. When I save this query as B and execute it, it does not display any result. The query executes indefinitely.
In SM50 the status is displayed as running with report CL_SQL_RESULT_SET=============CP. The action is Sequential Read.
Does anyone have faced simillar issue?
Regards,
Niraj

Hi Niraj,
You can change it in the t-code RSRT.
Open RSRT and select your query, you will see query properties. you can change the cache mode of query from this screen. you can try different modes and observe the performance.
check the "Update cache objects in delta process" option from the properties of query in RSRT.
For more info..
cache is a temporary storage for recently access data.
Used to enhance query performance.
Use t-code RSRT to vew more on cache.
Cache helps to improve query performance. As it can save data in memory, flatfile, cluster table or BLOB.
You can remove cache per query or inactivate it for perticular info provider or inactivate it overall.
But which is not recommended. Yes if u know that certain queries you dont use often and its not accessing large number of records then you can inactivate it for those queries. you can manage cache via three t-codes RSRT or RSRCACHE or SPRO>SAP Reference IMG>SAP Busines Wearhouse--> Reporting relevent setting -->General Reporting Setting in BEX -->Global Cache Setting.
Hope this helps,
Sri...

Similar Messages

  • Query on partitioned table runs indefinitely

    Hi,
    Customer's Oracle GL implementation has tables GL_BALANCES and GL_JE_LINES partitioned on PERIOD_NAME i.e MON-YY.
    Both GL_BALANCES and GL_JE_LINES have LOCAL INDEXES.
    For GL_JE_LINES it is (PERIOD_NAME, CODE_COMBINATION_ID)
    For GL_BALANCES it is (PERIOD_NAME, CODE_COMBINATION_ID)
    We have had more than one occasion is our Production Instance when queries run indefinitely for hours when accessing rows via the above local indexes. We then gather stats using fnd_stats.gather_table_stats for the specific partition e.g. NOV-09 and immediately the queries complete once the stats gathering is in progress.
    I am really puzzled how gather stats helps in this case, when the query is already running and from tracing the SQL session shows usage of the LOCAL INDEX.
    Any pointers on what Gather Stats is doing would be great.
    We do gather stats regularly in our Production Instance.
    Also when the SQL query for example against GL_BALANCES for specific period is running for long, I checked the stats and the number of rows in the dba_tab_partitions is not much different from the physical count of records for that period.
    Sometimes we have noticed the TEMP space run out say even upto 48 GB.
    In one instance we saw huge number of direct path write temp wait events. After gather stats was done, we no more see this wait event. Note we never used SORT or GROUP BY when we saw the above wait event.
    Regards,
    Prince
    Edited by: prc_richard on 10/11/2009 15:16

    Presumably the optimizer was choosing a poor execution plan because the statistics (data volumes and distribution) were out of date, and gathering the stats enabled it to make some better decisions and arrive at a better execution plan.
    For example, it may have thought the newest partitions were empty (see num_rows in all_tab_partitions) when in fact they had been loaded with millions of rows since you last gathered stats.
    Edited by: William Robertson on Nov 11, 2009 12:16 AM

  • I have downloaded the latest stable release a few times now and Firefox code variants, all when I close the browser it's process is running indefinitely. Help?

    I have downloaded the latest stable release of Firefox and a couple of variants using Firefox code a number of times, the browser crashes totally on certain pages mostly on the yahoo home page.
    The main issue is when I manage to avoid the troublesome pages and close down the browser using the red, top right hand corner "X" the browser process is still running and runs indefinitely, this stops the user from reopening the browser unless they manually stop the process through a process manager (task manager). Please help as I "loved" using Firefox.
    Thanks,
    Karl

    Hello Karl, we're sorry to hear that Firefox is crashing. In order to assist you better, please follow the steps below to provide us crash ID's to help us learn more about your crash.
    #Enter about:crashes in the address bar (that's where you enter your website URL) and press Enter. You should now see a list of submitted crash reports.
    #Copy the 5 most recent crash ID's that you see in the crash report window and paste them into your response here.
    More information and further troubleshooting steps can be found in the [[Firefox crashes]] article.
    also you can try to boot the computer in Windows Safe mode with network support (press F8 on the boot screen) and check firefox again.
    (If works in Windows Safe mode then you have a problem with other software, like security software or maybe a system driver, that is running on your computer).
    thank you

  • Query running slow after 1000 rows in oracle

    Hi.
    I have one query which is fetching miln of rows.. the query runs very fast till 1000 to 1500 records after that it run very slow. Can you please help ,what could be the reason?
    Thanks

    831269 wrote:
    I have one query which is fetching miln of rows.. Why are you fetching that many rows? What is your client code going to do with a million rows? And why do you expect this to be fast? A million rows worth of I/O has to be done by Oracle (that will likely be mostly from disk and not buffer cache). That million rows has to be copied from the Oracle's SGA to client memory. If your client is PL/SQL code, that will be copied into the PGA. If your client is external, then that copy has to happen across platform boundaries and the network too.
    Then your code churns away on processing a million rows... doing what exactly? That "+what+" will need to be done once per row, for a million times. If it takes 10ms per row, that means almost 3h of client processing time.
    Fetching that many rows..? Often a design and coding mistake. Always an exception to the rule. Will never be "fast".
    And scalability and performance need to be addressed by re-examining the requirements, optimising the design that necessitates fetching that many rows, and using techniques such as parallel processing and thread safe code design.

  • Remote client copy (SCC9) runs a very long time!

    remote client copy (SCC9) runs a very long time!
    how to do it quickly process?
    (eg use imp and exp-oracle tool, as it can be done to understand what the SAP data has been copied and are now in a different location, for Developers)

    scn001 wrote:
    remote client copy (SCC9) runs a very long time!
    > how to do it quickly process?
    > (eg use imp and exp-oracle tool, as it can be done to understand what the SAP data has been copied and are now in a different location, for Developers)
    Hi,
    You can export the client, as well but it will take long time too, depended to your client size. Please note that client copy operation should be performed by standard SAP client management tools, such as client export/import or remote copy.
    Ask this question to SAP support first. Technically, you can choose many ways to copy a SAP client, but as far as I know that SAP will not support you (such as errors you faced during the client export/import or the problems related by the copy operation), if you use any other 3rd party tool while for the client copy purposes.
    Best regards,
    Orkun Gedik
    Edited by: Orkun Gedik on Jun 30, 2011 10:57 AM

  • Copy Query Variants from InfoCube to MultiProvider

    Hi,
    Is it possible to copy query variants from an infocube to a multiprovider?
    Basically, we used RSZC to copy the queries from the infocube to the multiprovider, but it doesn't look like the variants were copied over. Is there an ABAP program or some other workaround that can expedite/facilitate in recreating the variants on the multiprovider queries?
    Your help is much appreciated.
    Thanks,
    KK
    Edited by: Kristin Kalalau on Jan 29, 2009 10:48 PM

    It is possible to copy but the only requirement is your target cube structure should match with the source cube structure.
    Thanks...
    Shambhu
    Edited by: Shambhu Kumar Gupta on Jan 30, 2009 12:01 PM

  • Query runs forever when using selection on line item dim

    we have a cube zcube which has po number as a line item dimension .
    everymonth users run queries this cube using po number as selection
    criterai and using a wild card search on this field . every time
    queries have run fine . however , this month when we try to do so
    the query runs forever and no results are returned . i also tried
    listcubing with similar selection but it also did not return any results .
    our production system has lot of data . i tried in test system it worked fine
    i checked the cube for compression and indices in production , they look fine
    can anyone think of anything that could have gone wrong ? also did rsrv tests ..
    but all come green
    we have not done any developments on this cube , however we have shifted to
    a new hardware in the past month . can anyone think of any reasons ?
    anything that can help me catch the issue ? all suggestions welcome

    Not sure,
    Are you saying that you need both the counts seperately? or a combined count?
    how are you joining alpha and beta tables? what is the join condition?
    you need to do something like this,
    SELECT COUNT(CASE WHEN
                        (A.col1 = 'Pete' AND SUBSTR(A.col2,,1,12)=SUBSTR(B.col2,,1,13))
                     OR ( A.col1 != 'Pete' AND SUBSTR(A.col2,,1,15)=SUBSTR(B.col2,,1,15))
                    THEN 1 ELSE 0
                 END)
    FROM alpha A, beta b
    WHERE alpha.join_cloumn= beta.join_columnG.

  • SSRS 2008 R2 is extremely slow. The query runs in less than a second in the dataset designer but if you try to view the report it takes over 10 minutes. I have read this is a bug in SSRS 2008 R2. We installed the most recent patches and service packs.

    SSRS 2008 R2 is extremely slow.  The query runs in less than a second in the dataset designer but if you try to view the report it takes over 10 minutes.  I have read this is a bug in SSRS 2008 R2.  We installed the most recent patches and
    service packs.  Nothing we've done so far has fixed it and I see that I'm not the only person with this problem.  However I don't see any answers either.

    Hi Kim Sharp,
    According to your description that when you view the report it is extremely slow in SSRS 2008 R2 but it is very fast when execute the query in dataset designer, right?
    I have tested on my local environment and can‘t reproduce the issue. Obviously, it is the performance issue, rendering performance can be affected by a combination of factors that include hardware, number of concurrent users accessing reports, the amount
    of data in a report, design of the report, and output format. If you have parameters in your report which contains many values in the list, the bad performance as you mentioned is an known issue on 2008 R2 and already have the hotfix:
    http://support.microsoft.com/kb/2276203
    Any issue after applying the update, I recommend you that submit a feedback at https://connect.microsoft.com/SQLServer/ 
    If you don’t have, you can do some action to improve the performance when designing the report. Because how you create and update reports affects how fast the report renders.
    Actually, the Report Server ExecutionLog2  view contains reports performance data. You could make use of below query to see where the report processing time is being spent:
    After you determine whether the delay time is in data retrieval, report processing, or report rendering:
    use ReportServer
    SELECT TOP 10 ReportPath,parameters,
    TimeDataRetrieval + TimeProcessing + TimeRendering as [total time],
    TimeDataRetrieval, TimeProcessing, TimeRendering,
    ByteCount, [RowCount],Source, AdditionalInfo
    FROM ExecutionLog2
    ORDER BY Timestart DESC
    Use below methods to help troubleshoot issues according to the above query result :
    Troubleshooting Reports: Report Performance
    Besides this, you could also follow these articles for more information about this issue:
    Report Server Catalog Best Practices
    Performance, Snapshots, Caching (Reporting Services)
    Similar thread for your reference:
    SSRS slow
    Any problem, please feel free to ask
    Regards
    Vicky Liu

  • What are the ways to make Query run fast?

    Hi Experts,
    When a query runs slow, we generally go for creating an aggregate. My doubt is - what other things can be done to make a query run faster before creating an aggregate? What is the thumb rule to be carried out for creating an aggregate?
    Regards,
    Shreeem

    Hi Shreem,
    If you keep Query simple not complicate it with runtime calculations , it would be smooth. However as per business requirements we will have to go for it anyways mostly.
    regarding aggregates:
    Please do not use the standard proposal , it will give you hundreds based on std. rules , which consumes lots of space and adds up to load times. If you have users already using the Query and you are planning to tune it  then go for the statistics tables:
    1.RSDDSTAT_OLAP    find the query with long runtimes  get the Stepuid
    2. RSDDSTAT_DM
    3. RSDDSTATAGGRDEF  - use the stepuid above to see which aggregate is necessary for which cube.
    Another way to check ; check the users as in 1 to find the highest runtime users and find the last used bookmarks by user thru RSZWBOOKMARK for this query and check if the time matches and create the aggregates as in 3 above.
    You can also Use Transaction RSRT > execute & debug  (display stats ) - to create generic aggregates to support navigations for New queries and later refine as above.
    Hope it helps .
    Thnks
    Ram

  • Select query running long time

    Hi,
    DB version : 10g
    platform : sunos
    My select sql query running long time (more than 20hrs) .Still running .
    Is there any way to find sql query completion time approximately. (Pending time)
    Also is there any possibilities to increase the speed of sql query (already running) like adding hints.
    Please help me on this .
    Thanks

    Hi Sathish thanks for your reply,
    I have already checked in V$SESSION_LONGOPS .But it's showing TIME_REMAINING -->0
    select TOTALWORK,SOFAR,START_TIME,TIME_REMAINING from V$SESSION_LONGOPS where SID='10'
    TOTALWORK      SOFAR START_TIME      TIME_REMAINING
         1099759    1099759 27-JAN-11                    0Any idea ?
    Thanks.

  • Insert or copy query into workbook 7.0

    How Can I copy a query into a workbook with new Analyzer 7.0 ?
    In the old version (Analyzer 3.x), I can choose Tools->Insert (or copy) query.
    How Can I do the same with Analyzer 7.0 ?
    Thanks

    Hi,
    .) Open the analyser and connect to the BI system,
    .) open an empty excel
    .) switch to design mode (first icon on second task menu)
    .) insert a table (second icon on second task menu)
    .) select the query
    and so on...
    BR
    /C

  • Query runs fine in 9i but results to ORA-01652 unable to extend temp in 10g

    Hi,
    We are having issues in running a SQL query in 10g. In 9i, it runs fine with no problems but when run in 10g, It takes forever and the temp tablespace grows very large upto 60GB until we get ORA-01652 error due to lack of disk space. This does not occur in 9i, where query runs in only 20 mins and does not take up temp that big. 9i version is 9.2.0.8. 10g is 10.2.0.3

    Heres the SQL query:
    SELECT
    J2.EMPLID,
    TO_CHAR(J2.EFFDT,'YYYY-MM-DD'),
    J2.EFFSEQ,
    J2."ACTION",
    J2.ACTION_REASON,
    TO_CHAR(J2.GRADE_ENTRY_DT,'YYYY-MM-DD'),
    J2.COMPRATE,
    J2.CHANGE_AMT,
    J2.COMP_FREQUENCY,
    J2.STD_HOURS,
    J2.JOBCODE,
    J2.GRADE,
    J2.PAYGROUP,
    PN2.NATIONAL_ID,
    TO_CHAR(PC.CHECK_DT,'YYYY-MM-DD'),
    SUM(PO.OTH_EARNS),
    To_CHAR(SUM(PO.OTH_EARNS)),
    PO.ERNCD,
    '3',
    TO_CHAR(PC.PAY_END_DT,'YYYY-MM-DD'),
    PC.PAYCHECK_NBR
    FROM PS_JOB J2,
    PS_PERS_NID PN2,
    PS_PAY_OTH_EARNS PO,
    PS_PAY_CHECK PC
    WHERE J2.EMPL_RCD = 0
    AND PN2.EMPLID = J2.EMPLID
    AND PN2.COUNTRY = 'USA'
    AND PN2.NATIONAL_ID_TYPE = 'PR'
    AND J2.COMPANY <> '900'
    AND J2.EFFDT <= SYSDATE
    AND PC.EMPLID = J2.EMPLID
    AND PC.COMPANY = PO.COMPANY
    AND PC.PAYGROUP = PO.PAYGROUP
    AND PC.PAY_END_DT = PO.PAY_END_DT
    AND PC.OFF_CYCLE = PO.OFF_CYCLE
    AND PC.PAGE_NUM = PO.PAGE_NUM
    AND PC.LINE_NUM = PO.LINE_NUM
    AND PC.SEPCHK = PO.SEPCHK
    AND EXISTS (SELECT ERNCD
    FROM PS_P1_CMP_ERNCD P1_CMP
    WHERE P1_CMP.ERNCD = PO.ERNCD AND EFF_STATUS = 'A')
    GROUP BY J2.EMPLID,
    J2.EFFDT,
    J2.EFFSEQ,
    J2.ACTION,
    J2.ACTION_REASON,
    J2.GRADE_ENTRY_DT,
    J2.COMPRATE,
    J2.CHANGE_AMT,
    J2.COMP_FREQUENCY,
    J2.STD_HOURS,
    J2.JOBCODE,
    J2.GRADE,
    J2.PAYGROUP,
    PN2.NATIONAL_ID,
    PC.CHECK_DT,
    PO.ERNCD,
    '3',
    PC.PAY_END_DT,
    PC.PAYCHECK_NBR

  • Is index range scan the reason for query running long time

    I would like to know whether index range scan is the reason for the query running long time. Below is the explain plan. If so, how to optimise it? Please help
    Operation     Object     COST     CARDINALITY     BYTES
    SELECT STATEMENT ()          413     1000     265000
    COUNT (STOPKEY)                    
    FILTER ()                    
    TABLE ACCESS (BY INDEX ROWID)     ORDERS     413     58720     15560800
    INDEX (RANGE SCAN)     IDX_SERV_PROV_ID     13     411709     
    TABLE ACCESS (BY INDEX ROWID)     ADDRESSES     2     1     14
    INDEX (UNIQUE SCAN)     SYS_C004605     1     1     
    TABLE ACCESS (BY INDEX ROWID)     ADDRESSES     2     1     14
    INDEX (UNIQUE SCAN)     SYS_C004605     1     1     
    TABLE ACCESS (BY INDEX ROWID)     ADDRESSES     2     1     14
    INDEX (UNIQUE SCAN)     SYS_C004605     1     1

    The index range scan means that the optimiser has determined that it is better to read the index rather than perform a full table scan. So in answer to your question - quite possibly but the alternative might take even longer!
    The best thing to do is to review your query and check that you need every table included in the query and that you are accessing the tables via the best route. For example if you can access a table via primary key index that would be better than using a non-unique index. But the best way of reducing the time the query takes to run is to give it less tables (and indexes) to read.
    John Seaman
    http://www.asktheoracle.net

  • Query running long time

    hi
    I'm having a query running for long time, Im new to dba can any one suggest me methods to make it faster it's running now and i have to make it execute it faster
    parallel servers=4, and there are no inactive sessions.
    thanks in advance

    Make a habit of putting the database version in the post
    As i told u before i depends on lot of things not only merge(cartisian ) joins,
    1)It depends on the load the database is having,Was this query running fastly before?if it was running fastly then was the workload same as today?
    2)Any changes done to database recently or the server?
    3)only this query is slow all the queris are slow?
    4)When was database last restarted?
    5)Are u using bind variable in the query?
    6)Is you library cache properly sized?If the query is doing lots of sorts then is your PGA properly sized?
    7)Database buffer cache is properly sized?
    8)How much memory your database is having?
    9)Is your SGA properly fits in your memory or its getting swaped?
    Etc...Etc
    Check all these things
    Regards
    Kaunain

  • Initial execution of a workbook/query runs without ending

    Hi, Experts,
    I encountered a very strange behavior of Bex tool. Sometimes, Initial execution of a workbook/query runs without ending at u2018Waiting for reply from BW Serveru2019 state, then I cancel the report and run it again and it would return results within minutes. Itu2019s like it forgot to return results.
    The basis team looked at the queries in the database, and said that there were actually a number of different queries being executed by my Bex query, but each individual query runs fairly quickly. The basis person pulled the statements out and ran them directly against the database and they returned result in sub seconds.
    Any thoughts? Possible solution?
    Thanks
    Chimei

    Hi,
    Could u let us know what is the GUI frontend patch version you are on. Also, are there any AddOn's Installed.
    What version of Excel are you using. Also just for a check, run sapbexc.xla on your local machine and check the output.
    Regards,
    Sree.

Maybe you are looking for

  • F-20 bills of exchange No Due items F5345

    Hi We are experiencing a posting issue with Bill of exchange Reverse Liability tcode F-20. The documents are posted as below; Check/BOE receipt posted with Posting Date  01.01.2014 Due on          28.02.2014 Issue date     01.01.2014 Used on        0

  • Blurry photos in iPhoto 11 when transferring from iPhoto 6 on different computer

    I recently bought a new MacBook Pro and I'm transferring my files over from my old MacBook.  When I moved my photos from the old one (running iPhoto 6) over to the new one (running iPhoto 11), the images look horrible.  They're all blurry and I can't

  • Invalid web.xml

    I am having a problem within Tomcat (v. 3.3 or 4.0) where the web.xml for the webapps.war file give an error. The error is: At Invalid element 'servlet' in content of 'web-app', expected elements '[error-page, taglib, resource-ref, security-onstraint

  • I accidentally erased my phone app on iphone. How do I recover it?

    i accidentally erased my phone app on iphone. How do I recover it?

  • Factory Calander

    hiiii when i m doing Mb1a Enter good issue then system showing error msg.. "Date 20100302 lies after the end of the valid factory calendar (please correct) " Message no. AG013 somebody suggest me to check ox10 n extend year 2009 up to 2010. i done it