Online response time in discoverer viewer?

Dear sir,
Can anyone help me,
In discoverer view and plus,how to maximise the response time for the viewer,while n-number of visitors requesting the same report.
How to analyse that? what settings are there?
Plz help me , its urgent for my project?
Regards
chandrakumar

In case this helps anyone else - this is the reply I got when I asked SAP this question:
the total user response time is response time + frontend network time.
GUI time is already included in the response time (it is the roll in andwait time from the application server's point of view).
The 'FE (Frontend) net time' is the time consumed in the network for thefirst transfer of data from the frontend to the application server andthe last transfer of data from the application server to the frontend
(during one dialog step).

Similar Messages

  • Discoverer Reporting time in Discoverer view /plus

    Hi I am running reports in the Discoverer View. Some reports are taking more than expected time. I want to know exact time taken by the reports. IS there any tables that storing Execution time for Discoverer reports like start time and end time for the report.
    Thanks.

    Hi,
    You can turn on statistics collection and then look in the EUL5_QPP_STATS table for the start time and duration of all the workbooks run on the system.
    Rod West

  • Discoverer Performance/ Response Time

    Hi everyone,
    I have a few questions regarding the response time for discoverer.
    I have Table A with 120 columns. I need to generate a report based on 12 columns from this table A.
    The questions are whether the factors bellow contribute to the response time of discoverer.
    1. The number of items included in the business area folder (i.e. whether to include 120 cols or just 12 cols)
    2. The actual size of the physical table (120 cols) although I only selected 12 cols. If the actual size of the physical table is only 12 cols, would it improve the performance?
    3. Will more parameters increase the processing time?
    4. Does Joins increase the processing time?
    5. Will using Custom Folder and writing an sql statement to select the 12 columns increase the performance?
    Really appreciate anyone's help on this.
    Cheers,
    Angeline

    Hi,
    NP and Rod, thanks a lot for your replies!
    Actually I was experiencing a different thing that contradicts your replies.
    1. When I reduced the no of items included in my Biz Area from 120 to 12 the response time improve significantly from around 5 minutes to 2-3 minutes.
    2. When I tried to create a dummy table with just 12 cols needed for the report, I could get a very fast response time, i.e. 1 second to generate the report. But of course the dummy table contains much less data (only around 500 K records). Btw, is Discoverer able to handle large database? What is the biggest record size can it handle?
    3. When I add more parameters it seems to add more processing time to the discoverer.
    4. Thanks for the clarification on this one.
    5. And the funny thing is, when I use custom folder to just select the 12 columns, the performance also significantly improves with estimated query time reduced from 2 minutes plus to just 1 mins 30 secs. But still the performance time is inconsistent. Sometimes it only takes around 1 mins 40 secs, but sometimes it can run up to 3 mintues for the same data.
    Now I am creating my report using the custom folder cause it has the best response time so far for me. But based on your replies it's not really encouraged to use the custom folder?
    I need to improve the response time for the Discoverer Viewer as the response time is very slow and users don't really like it.
    Would appreciate anyone's help in solving this issue :) Thanks..
    Cheers,
    Angeline

  • SBWP  transaction - viewing folders/sending documents Long response times

    Hi all,
    I have some complains from some users (about 3-4 out of ~3000 user in my ECC 6.0 system) within my company about their response times on the Business Workplace. In particulat they started complaining about the response time of calling the TCOD SBWP . For 1-2 of them up to 4-5 minutes when myself as well as my other 2 colleagues were getting response of 500ms.
    Then they wanted also to view some folders on the Workplace and they had also response times of minutes instead of seconds.
    I checked that some of their shared folders as well as the Trash Bin had thousands of PDFs. I told to delete them and they deleted most of them. Stil when they want to open a folder it takes >2 minutes while for me to open the same shared folder takes 1-2 seconds.
    I checked in ST03N (user profiles, single records) and 1 of them had long database calls and request time in the Analysis of ABAP/4 database requests (Single Statistical Records).
    I am running out of ideas as I cannot explain why only for those 2-3 users happens to have long response times.
    Is it related to their folders in the Workplace? Where should I focus my investigation for the SBWP like transactions? Is it the case that some Oracle parameters might need to be checked?
    I run the automatic Oracle parameters check (O/S AIX 5.3 , Oracle 10.2 , ECC 6.0) and here are some recommandations:
    fixcontrol (5705630)                   add with value "5705630:ON"                                                                                use optimal OR concatenation; note 176754                    NO                                                          5705630:ON                                         B          1             
    fixcontrol (5765456)                   add with value "5765456:3"                                                                                no further information available                             NO                                                          5765456:3                                          B          1             
    optimpeek_user_binds                   add with value "FALSE"                                                                                avoid bind value peaking                                     NO                                                          FALSE                                              B          1             
    optimizerbetter_inlist_costing         add with value "OFF"                                                                                avoid preference of index supporting inlist                  NO                                                          OFF                                                B          1             
    optimizermjc_enabled                   add with value "FALSE"                                                                                avoid cartesean merge joins in general                       NO                                                          FALSE                                              B          1             
    sortelimination_cost_ratio             add with value "10"                                                                                use non-order-by-sorted indexes (first_rows)                 NO                                                          10                                                 B          1             
    event (10027)                            add with value "10027 trace name context forever, level 1"                                                               avoid process state dump at deadlock                         NO                                                          10027 trace name context forever, level 1          B          1             
    event (10028)                            add with value "10028 trace name context forever, level 1"                                                               do not wait while writing deadlock trace                     NO                                                          10028 trace name context forever, level 1          B          1             
    event (10091)                            add with value "10091 trace name context forever, level 1"                                                               avoid CU Enqueue during parsing                              NO                                                          10091 trace name context forever, level 1          B          1             
    event (10142)                            add with value "10142 trace name context forever, level 1"                                                               avoid Btree Bitmap Conversion plans                          NO                                                          10142 trace name context forever, level 1          B          1             
    event (10183)                            add with value "10183 trace name context forever, level 1"                                                               avoid rounding during cost calculation                       NO                                                          10183 trace name context forever, level 1          B          1             
    event (10191)                            add with value "10191 trace name context forever, level 1"                                                               avoid high CBO memory consumption                            NO                                                          10191 trace name context forever, level 1          B          1             
    event (10411)                            add with value "10411 trace name context forever, level 1"                                                               fixes int-does-not-correspond-to-number bug                  NO                                                          10411 trace name context forever, level 1          B          1             
    event (10629)                            add with value "10629 trace name context forever, level 32"                                                              influence rebuild online error handling                      NO                                                          10629 trace name context forever, level 32         B          1             
    event (10753)                            add with value "10753 trace name context forever, level 2"                                                               avoid wrong values caused by prefetch; note 1351737          NO                                                          10753 trace name context forever, level 2          B          1             
    event (10891)                            add with value "10891 trace name context forever, level 1"                                                               avoid high parsing times joining many tables                 NO                                                          10891 trace name context forever, level 1          B          1             
    event (14532)                            add with value "14532 trace name context forever, level 1"                                                               avoid massive shared pool consumption                        NO                                                          14532 trace name context forever, level 1          B          1             
    event (38068)                            add with value "38068 trace name context forever, level 100"                                                             long raw statistic; implement note 948197                    NO                                                          38068 trace name context forever, level 100        B          1             
    event (38085)                            add with value "38085 trace name context forever, level 1"                                                               consider cost adjust for index fast full scan                NO                                                          38085 trace name context forever, level 1          B          1             
    event (38087)                            add with value "38087 trace name context forever, level 1"                                                               avoid ora-600 at star transformation                         NO                                                          38087 trace name context forever, level 1          B          1             
    event (44951)                            add with value "44951 trace name context forever, level 1024"                                                            avoid HW enqueues during LOB inserts                         NO

    Hi Loukas,
    Your message is not well formatted so you are making it harder for people to read. However your problem is that you have 3-4 users of SBWP with slow runtimes when accessing folders. Correct ?
    You mentioned that there is a large number of documents in the users folders so usually these type of problems are caused by a large number of table joins on the SAPoffice tables specific to your users.
    Firstly please refer to SAP Note 988057 Reorganization - Information.
    To help with this issue you can use report RSBCS_REORG in SE38 to remove any deleted documents from the SAPoffice folders. This should speed up the access to your users documents in folders as it removes unnecessary documents from the SAPoffice tables.
    If your users do not show a significant speed up of access to their SAPoffice folders please refer to SAP Note 904711 - SAPoffice: Where are documents physically stored and verify that your statistics and indexes mentioned in this note are up to date.
    If neither of these help with the issue you can trace these users in ST12 and find out which table is cauing the longest runtime and see if there is a solution to either reduce this table or improve the access method on the DB level.
    Hope this helps
    Michael

  • SAP GoLive : File System Response Times and Online Redologs design

    Hello,
    A SAP Going Live Verification session has just been performed on our SAP Production environnement.
    SAP ECC6
    Oracle 10.2.0.2
    Solaris 10
    As usual, we received database configuration instructions, but I'm a little bit skeptical about two of them :
    1/
    We have been told that our file system read response times "do not meet the standard requirements"
    The following datafile has ben considered having a too high average read time per block.
    File name -Blocks read  -  Avg. read time (ms)  -Total read time per datafile (ms)
    /oracle/PMA/sapdata5/sr3700_10/sr3700.data10          67534                         23                               1553282
    I'm surprised that an average read time of 23ms is considered a high value. What are exactly those "standard requirements" ?
    2/
    We have been asked  to increase the size of the online redo logs which are already quite large (54Mb).
    Actually we have BW loading that generates "Chekpoint not comlete" message every night.
    I've read in sap note 79341 that :
    "The disadvantage of big redo log files is the lower checkpoint frequency and the longer time Oracle needs for an instance recovery."
    Frankly, I have problems undertanding this sentence.
    Frequent checkpoints means more redo log file switches, means more archive redo log files generated. right ?
    But how is it that frequent chekpoints should decrease the time necessary for recovery ?
    Thank you.
    Any useful help would be appreciated.

    Hello
    >> I'm surprised that an average read time of 23ms is considered a high value. What are exactly those "standard requirements" ?
    The recommended ("standard") values are published at the end of sapnote #322896.
    23 ms seems really a little bit high to me - for example we have round about 4 to 6 ms on our productive system (with SAN storage).
    >> Frequent checkpoints means more redo log file switches, means more archive redo log files generated. right?
    Correct.
    >> But how is it that frequent chekpoints should decrease the time necessary for recovery ?
    A checkpoint is occured on every logswitch (of the online redologfiles). On a checkpoint event the following 3 things are happening in an oracle database:
    Every dirty block in the buffer cache is written down to the datafiles
    The latest SCN is written (updated) into the datafile header
    The latest SCN is also written to the controlfiles
    If your redologfiles are larger ... checkpoints are not happening so often and in this case the dirty buffers are not written down to the datafiles (in the case of no free space in the buffer cache is needed). So if your instance crashes you need to apply more redologs to the datafiles to be in a consistent state (roll forward). If you have smaller redologfiles more log switches are occured and so the SCNs in the data file headers (and the corresponding data) are closer to the newest SCN -> ergo the recovery is faster.
    But this concept does not really fit the reality because of oracle implements some algorithm to reduce the workload for the DBWR in the case of a checkpoint.
    There are also several parameters (depends on the oracle version) which control that a required recovery time is kept. (for example FAST_START_MTTR_TARGET)
    Regards
    Stefan

  • How to view response time for each execution in OEM 12c Cloud Control

    Hi All,
    I want to view response time of each execution for my SOA composite in OEM 12c Cloud Control. Can anyone tell me how can I achieve that?
    Thanks in Advance!!

    You should open an SR and support can assist you with it. The only current case where that feature is known to not work is in the case where the Dehydration Store is running on RAC. Otherwise, it should be possible to get it to work. The main thing is that you need to have DB configuration properties (host, port, sid) that have a 3-way string match between the JDBC data source configuration in the WLS server, the JDBC properties you add to the Monitoring Properties of the SOA Infrastructure target, and the connect settings of the Database target in EM that matches to the dehydration DB.
    If there isn't a match in all 3, the Dehydration Diagnostics page will not work. But again, to look into your specific setup, you should file an SR and work with support.

  • Strange response time for an RFC call viewed from STAD on R/3 4.7

    Hello,
    On our R/3 4.7 production system, we have a lot of external RFC calls to execute an  abap module function. There are 70 000 of these calls per day.
    The mean response time for this RFC call is 35 ms.
    Some times a few of them (maybe 10 to 20 per day) are way much longer.
    I am currently analysing with STAD one of these long calls  which lasted 10 seconds !
    Here is the info from STAD
    Response time :                10 683 ms
    Total time in workprocess : 10 683 ms
    CPU time  :                               0 ms
    RFC+CPIC time :                       0 ms
    Wait for work process           0 ms
    Processing time           10.679 ms
    Load time                            1 ms
    Generating time                   0 ms
    Roll (in) time                        0 ms
    Database request time         3 ms
    Enqueue time                      0 ms
    Number      Roll ins            0
                     Roll outs           0
                     Enqueues         0
    Load time   Program             1  ms
                      Screen              0  ms
                     CUA interf.          0  ms
    Roll time    Out                 0  ms
                      In                  0  ms
                    Wait                0  ms
    Frontend    No.roundtrips       0   
                      GUI time            0  ms
                      Net time            0  ms
    There is nearly no abap processing in the function module.
    I really don't uderstand what is this  10 679 ms processing time especially with 0 ms cpu time and 0 ms wait time.
    A usual fast RFC call gives this data
    23 ms response time
    16 ms cpu time
    14 ms processing time
    1 ms load time
    8 ms Database request time
    Does anybody have an idea of what is the system doing during the 10 seconds  processing time ?
    Regards,
    Olivier

    Hi Graham,
    Thank you for your input and thoughts.
    I will have to investigate on RZ23N and RZ21 because I'm not used to use them.
    I'm used to investigate performance problems with ST03 and STAD.
    My system is R/3 4.7 WAS 6.20. ABAP and BASIS 43
    Kernel 6.40 patch level 109
    We know these are old patch levels but we are not allowed to stop this system for upgrade "if it's not broken" as it is used 7/7 24/24.
    I'm nearlly sure that the problem is not an RFC issue because I've found other slow dialog steps for web service calls and even for a SAPSYS technical dialog step  of type <no buffer>. (what is this ?)
    This SAPSYS dialog step has the following data :
    User : SAPSYS
    Task type : B
    Program : <no buffer>
    CPU time                       0 ms
    RFC+CPIC time              0 ms
    Total time in workprocs    5.490 ms
    Response time               5.490 ms
    Wait for work process           0 ms
    Processing time             5.489 ms
    Load time                            0 ms
    Generating time                   0 ms
    Roll (in+wait) time                0 ms
    Database request time          1 ms ( 3 Database requests)
    Enqueue time                       0 ms
    All hundreds of other SAPSYS <no buffer> steps have a less than 5 ms response time.
    It looks like the system was frozen during 5 seconds...
    Here are some extracts from STAD of another case from last saturday.
    11:00:03     bt1fsaplpr02_PLG RFC     R     3     USER_LECKIT     13     13     0     0
    11:00:03     bt1sqkvf_PLG_18      RFC     R     4     USER_LECDIS     13     13     0     0
    11:00:04     bt1sqkvh_PLG_18      RFC     R     0     USER_LECKIT     19     19     0     16
    11:00:04     bt1sqkvf_PLG_18      RFC     R     4     USER_LECKIT     77     77     0     16
    11:00:04     bt1sqkve_PLG_18      RFC     R     4     USER_LECDIS     13     13     0     0
    11:00:04     bt1sqkvf_PLG_18      RFC     R     4     USER_LECDIS     14     14     0     16
    11:00:05     bt1sqkvg_PLG_18      RFC     R     0     USER_LECKIT     12     12     0     16
    11:00:05     bt1sqkve_PLG_18      RFC     R     4     USER_LECKIT     53     53     0     0
    11:00:06     bt1sqkvh_PLG_18      RFC     R     0     USER_LECKIT     76     76     0     0
    11:00:06     bt1sqk2t_PLG_18      RFC     R     0     USER_LECDIS     20     20     0     31
    11:00:06     bt1sqk2t_PLG_18      RFC     R     0     USER_LECKIT     12     12     0     0
    11:00:06     bt1sqkve_PLG_18      RFC     R     4     USER_LECKIT     13     13     0     0
    11:00:06     bt1sqkvf_PLG_18      RFC     R     4     USER_LECKIT     34     34     0     16
    11:00:07     bt1sqkvh_PLG_18      RFC     R     0     USER_LECDIS     15     15     0     0
    11:00:07     bt1sqkvg_PLG_18      RFC     R     0     USER_LECKIT     13     13     0     16
    11:00:07     bt1sqk2t_PLG_18      RFC     R     0     USER_LECKIT     19     19     0     0
    11:00:07     bt1fsaplpr02_PLG RFC     R     3     USER_LECKIT     23     13     10     0
    11:00:07     bt1sqkve_PLG_18      RFC     R     4     USER_LECDIS     38     38     0     0
    11:00:08     bt1sqkvf_PLG_18      RFC     R     4     USER_LECKIT     20     20     0     16
    11:00:09     bt1sqkvg_PLG_18      RFC     R     0     USER_LECDIS     9 495     9 495     0     16
    11:00:09     bt1sqk2t_PLG_18      RFC     R     0     USER_LECDIS     9 404     9 404     0     0
    11:00:09     bt1sqkvh_PLG_18      RFC     R     1     USER_LECKIT     9 181     9 181     0     0
    11:00:10     bt1fsaplpr02_PLG RFC     R     3     USER_LECDIS     23     23     0     0
    11:00:10     bt1sqkve_PLG_18      RFC     R     4     USER_LECKIT     8 465     8 465     0     16
    11:00:18     bt1sqkvh_PLG_18      RFC     R     0     USER_LECKIT     18     18     0     16
    11:00:18     bt1sqkvg_PLG_18      RFC     R     0     USER_LECKIT     89     89     0     0
    11:00:18     bt1sqk2t_PLG_18      RFC     R     0     USER_LECKIT     75     75     0     0
    11:00:18     bt1sqkvh_PLG_18      RFC     R     1     USER_LECDIS     43     43     0     0
    11:00:18     bt1sqk2t_PLG_18      RFC     R     1     USER_LECDIS     32     32     0     16
    11:00:18     bt1sqkvg_PLG_18      RFC     R     1     USER_LECDIS     15     15     0     16
    11:00:18     bt1sqkve_PLG_18      RFC     R     4     USER_LECDIS     13     13     0     0
    11:00:18     bt1sqkve_PLG_18      RFC     R     4     USER_LECDIS     14     14     0     0
    11:00:18     bt1sqkvf_PLG_18      RFC     R     4     USER_LECKIT     69     69     0     16
    11:00:18     bt1sqkvf_PLG_18     RFC     R     5     USER_LECDIS     49     49     0     16
    11:00:18     bt1sqkve_PLG_18     RFC     R     5     USER_LECKIT     19     19     0     16
    11:00:18     bt1sqkvf_PLG_18     RFC     R     5     USER_LECDIS     15     15     0     16
    The load at that time was very light with only a few jobs starting :
    11:00:08 bt1fsaplpr02_PLG RSCONN01                     B 31 USER_BATCH     39
    11:00:08 bt1fsaplpr02_PLG RSBTCRTE                     B 31 USER_BATCH     34
    11:00:08 bt1fsaplpr02_PLG /SDF/RSORAVSH                B 33 USER_BATCH     64
    11:00:08 bt1fsaplpr02_PLG RSBTCRTE                     B 33 USER_BATCH     43
    11:00:08 bt1fsaplpr02_PLG RSBTCRTE                     B 34 USER_BATCH     34
    11:00:08 bt1fsaplpr02_PLG RSBTCRTE                     B 35 USER_BATCH     37
    11:00:09 bt1fsaplpr02_PLG RVV50R10C                    B 34 USER_BATCH     60
    11:00:09 bt1fsaplpr02_PLG ZLM_HDS_IS_PURGE_RESERVATION B 35 USER_BATCH    206
    I'm thinking also now about the message server as there is load balancing for each RFC call ?
    Regards,
    Olivier

  • Discoverer viewer in portal or Drill to detail in Disco report in Portal

    Hi
    Is it possible to embed discoverer viewer into Portal page. I know one way, I can achieve this in older versions by using <IFRAME src="worksheet url"></iframe> as HTML portlet. But, the problem here is that, our Discoverer viewer is SSO protected. So,I don't want user to single sign on again, as they have already have done that when they got into portal.
    My second question is , I have developed Page-Detail report in Disco plus, also I have Drill to Detail in the same report. Drill to Detail is created using "Manage Link" option. When I deploy this report into Portal using Disco Portlet provider, firstly in portal it is not showing the Page Item, also the Drill through link also is not showing.
    Any help on the above questions would be greatly helpful.
    Thanks
    Balaji

    Hi
    Firstly thank you very much for your response.
    The Disco report I have is a Discoverer Plus report with Drill to Detail to another workbook(developed through manage link). But this Drillable link is not showing in the published Disco portlet. Are you saying that it should in the published disco portlet? .. As posted in previous response , I know this would work from Analyze link(which basically lauches viewer)..
    But users would like to see this drillable link in the report in portal page itself(which means on the published portlet itself). To achieve this only, I used Iframes in Html portlet and showing the Discoverer viewer url. It works fine, but only thing is each time user naviges to portal page, it executes the report which takes some time. I am trying to avoid the execution process. I think one solution is I can just schdule these Disco reports and that way it will not query the database each time. I was wondering if there is any other solution maybe I am missing
    Thanks
    Balaji

  • I cant see the graph in discoverer viewer whith was created in the desktop

    Hi,
         I need your help!! I have a problem to see a graph in Discoverer viewer. I will detail you the steps that I follow;
    It was a difficult calculate report to results by Discoverer so, I create a view in the data base which show 3 field, Periodo(number) which is a union whit other field from other folder, State(varchar), Requirement(number).
    From the discoverer desktop y have a crosstab worksheets and a graph, there I can see it without problem, but in the Discoverer viewer I can see the graph.
    I tried to set the worksheets options in “show options” to show me the graphs too, but this options didn’t appeared there. Somebody know what I should have to do to see it?
    On the other hand I changed the “show null values show” option since null to empty, and when I wanted to save the change this error appeared;
    “Oracle BI Discoverer don’t had to found the necessary values to show the events…”
    Thanks for your help!
    Regards
    Dora

    Hi Dora,
    The graphs that you create using Discoverer Plus will show nicely in Viewer. Everything you develop in Pluys works in Viewer. They are companion products.
    You must realize that Discoverer Plus/Viewer and Desktop are two seperate product sets and they use different techniques for graph creation. There are other differencies between the two that suggests that you need to develop all reports using Discoverer Plus in stead of Discoverer Desktop.
    I have experienced workbook created in Plus being destroyed when opened and saved using Desktop. So for a long time I have simply forgotten all about Desktop. If your company is relying on both Viewer and Dersktop I suggest that the time has come to change strategy and technique.
    If you need sparring please mail to [email protected]
    Good Luck!
    Cheers
    Niels Jessen

  • Discoverer Viewer/Plus does not work with Custom EUL owned by APPS user

    I am facing a unique issue with one of my customers. They have a custom EUL named XX_EUL_US and this does not allow connecting through discoverer viewer or discoverer plus.
    I have worked with custom EULS before at other customers and this was never the case. I am being told that this issue is happening as the EUL owner is APPS instead of some other user. Is this the case? Can discoverer viewer not be used with EULs owned by apps user?
    Please not that the Default EUL in ICX profiles is set to this EUL and no other EUL profiles are set.
    Appreciate any thoughts on this.

    Hi,
    I am not sure what you mean when you say you have a "custom EUL named XX_EUL_US". The EUL table should be created in a schema named XX_EUL_US. This will then be the EUL owner. You will then need to connect to Discoverer Administrator as the EUL owner and grant privileges to Apps users and/or responsibilities.
    The user should then be able to connect to Discoverer Plus/View using their apps username, responsibility and password.
    To be able to connect directly from an apps menu you need to have the ICX parameters set correctly. The ICX: Discoverer Default End User Layer Schema Prefix should be set to XX_EUL (without the _US). The ICX: Discoverer End User Layer Language Override should be set to US or not set at all.
    Rod West

  • Discoverer viewer displaying all the reports instead of one report

    Hi Team,
    I have created one menu and attached 4 reports to that.after that that menu is attached to the main menu.This menu is attached to the one xyz responsibility.
    Now i am facing the problem is : After logging to the responsibility,i clicked on individual report ,but it is not displaying only one report.It is showing all the reports in the discoverer viewer.
    Can you please help me in this issue..

    Please post the details of the application release, database version and OS.
    I have created one menu and attached 4 reports to that.after that that menu is attached to the main menu.This menu is attached to the one xyz responsibility.
    Now i am facing the problem is : After logging to the responsibility,i clicked on individual report ,but it is not displaying only one report.It is showing all the reports in the discoverer viewer.Please review these docs and verify that you have completed all the steps.
    How to Create a Link to a Discoverer Workbook in Apps R12 [ID 471303.1]
    How to Create a Link to a Discoverer Workbook in Apps11i [ID 278095.1]
    Thanks,
    Hussein

  • Discoverer Viewer Report with Page Items Slow Performance

    I created a report in Discoverer User that has page items and parameters and runs fine (under 3 minutes) in Discoverer User, but when I try and run it in Discoverer Viewer it takes a long time. It took over 70 minutes before manually terminating it. I made a copy of the report and removed the page items and tried to run it again in Discoverer Viewer and now the report runs in under 3 minutes. Any idea why a report with page items would run in Discoverer user fine, but not discoverer viewer?
    I am using Oracle Business Intelligence Plus 10g (10.1.2.55.26)

    As long as you can access the middle-tier server where Discoverer Viewer is running, you should be able to run and view Discoverer worksheets.
    While it is difficult to even start guessing what the problem might be, if the machine you installed Discoverer on has DHCP addressing, you will run into problems.
    Thanks
    Abhinav
    Oracle Business Intelligence Product Management
    BI on Oracle: http://www.oracle.com/bi/
    BI on OTN : http://www.oracle.com/technology/products/bi/
    Discoverer : http://www.oracle.com/technology/products/discoverer/
    BI Software : http://www.oracle.com/technology/software/products/ias/devuse.html
    Documentation: http://www.oracle.com/technology/documentation/appserver1012.html
    BI Samples : http://www.oracle.com/technology/products/bi/samples/
    Blog : http://oraclebi.blogspot.com/

  • No data found in the discoverer Viewer from Apps responsabiltiy

    Hi
    We have built a new responsibility for the discoverer reports, that is we directly access the discoverer viewer from the separate responsibility,all the report that shared with the users can be viewed from this responsibility,
    but when the user tries to run the report it is saying no data found.but for example if the user is trying to run the hrms report from this discoverer reasonability he get no data found but if we attach the function to the HRMS manager menu and try to access the report from the HRMS manager responsabilty we are able to get the data.
    We did the same setup in dev and test.in dev it is working very fine but in test we are not able to see the records in the report if we try to access.
    Please advice
    thanks
    Regards
    Ramesh Kumar S

    Hi,
    I'm not 100% familiar with HR security, but sounds like profile option setup related to security - perhaps Business Group or Initialization SQL Statement - Oracle / Custom? Check the profile options set for the HRMS Manager responsibility and setup the same for your new responsibility.
    PS. Similar sort of thing exists in General Ledger related to security. See:
    GL security issue in discoverer..
    Regards,
    Gareth

  • Report not working in Discoverer Viewer

    Hi,
    I'm using Discoverer 10.1.2.1 in conjunction with E-business Suite 12.1.3.
    I have a report that runs fine in Desktop or Plus, but when run in viewer (either from the application or from the stand-alone viewer link) it ends with the following error after running for approx 4 minutes:
    The application encountered an invalid state.
    OracleBI Discoverer Viewer was unable to find the necessary data for displaying the results of this event. Please correct any errors and try again.
    I recreated the report from scratch, using the same underlying data, and adding exactly the same items, conditions and sorting. The same responsibilities have been given access, and I verified that the SQL produced in both reports is exactly the same.
    However my new report works! From what I can see the only difference is the name and workbook identifier.
    Does anyone have any idea what could be causing my error? I have searched for the error on this forum and others but haven't found the answer.
    Thanks
    Hazel

    Hi Hazel
    This has come up before and I found this thread that I was involved in several years ago: https://forums.oracle.com/message/3682002
    Here's another one, this time relating to static IP addresses: https://forums.oracle.com/message/1199457
    There's also a document on Oracle Support: The application encountered invalid state.Oracle BI Discoverer Viewer was unable to find the necessary data for displaying the results of this event" (Doc ID 884965.1)
    Also, were you using Portal? I've seen instances where Viewer was unable to run a report that was linked via Portal because the report had changed since it was placed on the Portal.
    Best wishes
    Michael

  • Workbook works fine in discoverer plus but shows error in discoverer viewer

    Hi,
    I have some issues with the parameter of a report.
    one of the parameters is
    mgr_id (eg: 1,2,3) -- from database table.
    manager names are not available in the database but they are known to users like
    1 - david
    2 - alan chris
    3 - peter
    so the requirement is
    (i) the input parameter wud be entered as 1 or 2 or 3 or null ( or in multiple combination or all).
    But the report shud have parameters in the header of the report.
    for eg: if the input parameter is mgr_id = '1' then the report shud have manager name: 'david' in the header.
    if the input is mgr_id = '1',' 2', '3' then on the report header, it shud display as manager name: 'david','alan chris','peter'
    (ii) the default value of mgr_id shud be 'ALL'. if the user enters mgr_id as 'ALL' then it shud take all the values of mgr_id and pull all the values of mgr_id (like 1,2,3)
    and the report header shud display as department name: 'ALL'
    i have developed a custom query where i used,
    select dept_id, decode(mgr.mgr_id,'1','david','2','alan chris','3','peter') as mgr_name, ..... etc.
    in discoverer admin,
    I have created LOVs (item classes) for mgr_name.
    after that going to properties of mgr_name, i have selected mgr_id as indexed item.
    and in the report i have created parameter dept based on mgr_name and selected the option -- allow to select both indexes and values.
    and in the default value i mentioned 'ALL'.
    after that i changed the condition as
    ((mgr_name = :manager name) OR (mgr_name LIKE DECODE(:manager name,'ALL','%')))
    and in the header, i used &department name --- then it displays correctly.
    but the problem is ---
    the report works fine in plus. the problem is only when parameter value is 'ALL'
    when i run the report from discoverer viewer for the very first time (after logging in), the report works fine with correct data.
    but when i run the same report second time with parameter value 'ALL', it gives an error ---
    Invalid value "''" for parameter "department name"
    i dont understand, why it works for first time and doesn't afterwards. If i logout and login again, it works fine in viewer.
    am i doing anything wrong in conditions or declaring the parameters?
    I came to know that, the very first time, when we enter 'ALL' it is taking '%' but when we run for the second time, it is taking NULL. since NULL isn't there in mgr_id or mgr_name, it is throwing that error.
    How do i fix this error?
    Thanks

    Hi Puppethead,
    I have tried with your suggestions.
    the following condition -- did not work
    ( (mgr_name = :manager name)
    OR (mgr_name LIKE DECODE(:manager name, 'ALL', '%', NULL)))
    but for the other condition ---
    ( (mgr_name = :manager name)
    OR (mgr_name LIKE DECODE(nvl(:manager name, 'ALL'), 'ALL', '%', NULL)))
    is wrong because, :manager name cud be null, when it is null, it takes 'ALL' i.e. '%' which shud not be the case.
    if the user enters null, it has to display records with null only but with the above condition it takes '%'
    The main issue which i dont understand is, the report works fine for the first time. for the second time, the report takes null values . why is it taking null values for the second time.
    Thanks

Maybe you are looking for