Slow report performance when filter by 1 date

Post Author: poh_michelle
CA Forum: Formula
Record selection formula : {PartTran.TransDate} in {?FromDate} to {?ToDate}
Report performance was okay when i select #08/10/2007# for {?FromDate} and #10/10/2007# fro ; report displayed within 5 seconds.However, the report slow down wehn i select same date (#10/10/2007#) for both {?FromDate} and {?ToDate}; report displayed after 20 minutes.
What is the reason and what can be done to improve it.
Thanks for any advice.

Post Author: foghat
CA Forum: Formula
Not sure why selecting a single date would slow down your report.  Are you sure something else wasn't/isn't going on? You could give this a try: {PartTran.TransDate} >= {?FromDate}and {PartTran.TransDate} <= {?ToDate}

Similar Messages

  • VERY slow reporting performance

    We've created reports which, when run directly on Oracle through sqlplus (copying the exact SQL from the report) take around 3 minutes, when run from BO (using Web Intelligence) take hours.
    The SQL outputs around 300,000 rows, which equates to a few hundred pages in the report.  Is there a reason why the report would take so long (the session to Oracle is open the whole time)?  I can understand it not taking just the time for Oracle to execute the query, as WebI then needs to take the results and produce the report, but I wouldn't expect it to take hours.
    So far on our servers we've concentrated on making Oracle perform faster, and now we've got that working better our attention has been drawn to the performance issues of the reporting, and we don't have any experience with tuning it, so any advice would be very much appreciated.
    We're running BO Enterprise XI R2 SP3 on Linux with Web Intelligence as the only reporting interface.  The database is Oracle RAC (3 node cluster) 10.2.0.4.

    Correction: I incorrectly said jobserver in the earlier post (I read your post wrong).
    You need to start a trace on the CMS (for querytiming), the Webi Report Server and the Connection Server (which accesses the db)
    For the CMS trace, stop the CMS, in ccm.config, add a '-trace -querytiming' to the end of the cmsLaunch line. Similarly, add a -trace to the command line for the Webi Report Server (I don't remember the exact line; I don't have a XI R2 deployment on Linux available right now)
    To enable tracing on the connection server, do the following:
    (SAP 1197745 - How to enable connection server tracing on a UNIX platform)
    Open cs.cfg in a text editor. These are located by default in
    Linux: bobje/enterprise115/linux_x86/dataAccess/RDBMS/connectionServer/cs.cfg
    Locate the following entry:
    Traces ClassID="csTRACELOG" Active="No"
    Change the "No" to "Yes".
    Traces ClassID="csTRACELOG" Active="Yes"
    Save the file.
    Restart the Web Intelligence Report Server.
    ====================
    NOTE:
    Be aware that connection server traces slow down the server performance. Detailed traces get generated when this level of tracing is enabled. Disable tracing when not needed.
    ====================
    PS: The Webi Report Server contains the Connection Server libraries. The Connection Server runs in-proc within the Webi Report Server. Hence the need to restart the Webi Report Server.
    Also, I encourage you to open a ticket with BO on this. Performance issues are never easy to troubleshoot, even less so on a forum such as this.
    Finally, I don't believe there is any document that lists all command line options. Or rather, if such a document exists, I don't believe it's available for public consumption.

  • Missing report headers when exporting to XLS Data Only

    Hello all,
    I searched for this topic and was unable to find a similar one!
    I have a report that's very long.. (Page Setup:User Defined)
    When I export to Excel Data Only, with option of: Custom: Data is exported according to selected options, the output shows some headers only and skips many.
    I tried going into Design view and putting grids around each feild so that they don't overlap.
    How can I fix this issue?
    Thanks,

    Go to FILE->Export->Report Export Options then select Excel Data Only option from the Format drop down window then press ok.
    Check page headers and footer check box and press ok.
    This should set the options to export report header and footers.
    Hope this helps!
    Achett13

  • SLOW report performance with bind variable

    Environment: 11.1.0.7.2, Apex 4.01.
    I've got a simplified report page where the report runs slowly compared to running the same query in sqldeveloper. The report region is based on a pl/sql function returning a query. If I use a bind variable in the query inside apex it takes 13 seconds to run, and if I hard code a string it takes only a few hundredths of a second. The query returns one row from a table which has 1.6 million rows. Statistics are up-to-date and the columns in the joins and where clause are indexed.
    I've run traces using p_trace=YES from Apex for both the bind variable and hard coded strings. They are below.
    The sqldeveloper explain plan is identical to the bind variable plan from the trace, yet the query runs in 0.0x seconds in sqldeveloper.
    What is it about bind variable syntax in Apex that is causing the bad execution plan? Apex Bug? 11g bug? Ideas?
    tkprof output from Apex trace with bind variable is below...
    select p.master_id link, p.first_name||' '||p.middle_name||' '||p.last_name||' '||p.suffix personname,
    p.gender||' '||p.date_of_birth g_dob, p.master_id||'*****'||substr(p.ssn,-4) ssn, p.status status
    from persons p
    where
       p.person_id in (select ps.person_id from person_systems ps where ps.source_key  like  LTRIM(RTRIM(:P71_SEARCH_SOURCE1)))
    order by 1
    call     count       cpu    elapsed       disk      query    current        rows
    Parse        1      0.00       0.00          0          0          0           0
    Execute      1      0.00       0.01          0          1         27           0
    Fetch        2     13.15      13.22      67694      72865          0           1
    total        4     13.15      13.23      67694      72866         27           1
    Misses in library cache during parse: 0
    Optimizer mode: ALL_ROWS
    Parsing user id: 62  (ODPS_PRIVACYVAULT)   (recursive depth: 1)
    Rows     Row Source Operation
          1  SORT ORDER BY (cr=72869 pr=67694 pw=0 time=0 us cost=29615 size=14255040 card=178188)
          1   FILTER  (cr=72869 pr=67694 pw=0 time=0 us)
          1    HASH JOIN RIGHT SEMI (cr=72865 pr=67694 pw=0 time=0 us cost=26308 size=14255040 card=178188)
          1     INDEX FAST FULL SCAN IDX$$_0A300001 (cr=18545 pr=13379 pw=0 time=0 us cost=4993 size=2937776 card=183611)(object id 68485)
    1696485     TABLE ACCESS FULL PERSONS (cr=54320 pr=54315 pw=0 time=21965 us cost=14958 size=108575040 card=1696485)
    Rows     Execution Plan
          0  SELECT STATEMENT   MODE: ALL_ROWS
          1   SORT (ORDER BY)
          1    FILTER
          1     HASH JOIN (RIGHT SEMI)
          1      INDEX   MODE: ANALYZED (FAST FULL SCAN) OF
                     'IDX$$_0A300001' (INDEX)
    1696485      TABLE ACCESS   MODE: ANALYZED (FULL) OF 'PERSONS' (TABLE)
    Elapsed times include waiting on following events:
      Event waited on                             Times   Max. Wait  Total Waited
      ----------------------------------------   Waited  ----------  ------------
      db file scattered read                       1276        0.00          0.16
      db file sequential read                       812        0.00          0.02
      direct path read                             1552        0.00          0.61
    ********************************************************************************Here's the tkprof output with a hard coded string:
    select p.master_id link, p.first_name||' '||p.middle_name||' '||p.last_name||' '||p.suffix personname,
    p.gender||' '||p.date_of_birth g_dob, p.master_id||'*****'||substr(p.ssn,-4) ssn, p.status status
    from persons p
    where
       p.person_id in (select ps.person_id from person_systems ps where ps.source_key  like  LTRIM(RTRIM('0b')))
    order by 1
    call     count       cpu    elapsed       disk      query    current        rows
    Parse        1      0.02       0.04          0          0          0           0
    Execute      1      0.00       0.00          0          0         13           0
    Fetch        2      0.00       0.00          0          8          0           1
    total        4      0.02       0.04          0          8         13           1
    Misses in library cache during parse: 1
    Optimizer mode: ALL_ROWS
    Parsing user id: 62  (ODPS_PRIVACYVAULT)   (recursive depth: 1)
    Rows     Row Source Operation
          1  SORT ORDER BY (cr=10 pr=0 pw=0 time=0 us cost=9 size=80 card=1)
          1   FILTER  (cr=10 pr=0 pw=0 time=0 us)
          1    NESTED LOOPS  (cr=8 pr=0 pw=0 time=0 us)
          1     NESTED LOOPS  (cr=7 pr=0 pw=0 time=0 us cost=8 size=80 card=1)
          1      SORT UNIQUE (cr=4 pr=0 pw=0 time=0 us cost=5 size=16 card=1)
          1       TABLE ACCESS BY INDEX ROWID PERSON_SYSTEMS (cr=4 pr=0 pw=0 time=0 us cost=5 size=16 card=1)
          1        INDEX RANGE SCAN IDX_PERSON_SYSTEMS_SOURCE_KEY (cr=3 pr=0 pw=0 time=0 us cost=3 size=0 card=1)(object id 68561)
          1      INDEX UNIQUE SCAN PK_PERSONS (cr=3 pr=0 pw=0 time=0 us cost=1 size=0 card=1)(object id 68506)
          1     TABLE ACCESS BY INDEX ROWID PERSONS (cr=1 pr=0 pw=0 time=0 us cost=2 size=64 card=1)
    Rows     Execution Plan
          0  SELECT STATEMENT   MODE: ALL_ROWS
          1   SORT (ORDER BY)
          1    FILTER
          1     NESTED LOOPS
          1      NESTED LOOPS
          1       SORT (UNIQUE)
          1        TABLE ACCESS   MODE: ANALYZED (BY INDEX ROWID) OF
                       'PERSON_SYSTEMS' (TABLE)
          1         INDEX   MODE: ANALYZED (RANGE SCAN) OF
                        'IDX_PERSON_SYSTEMS_SOURCE_KEY' (INDEX)
          1       INDEX   MODE: ANALYZED (UNIQUE SCAN) OF 'PK_PERSONS'
                      (INDEX (UNIQUE))
          1      TABLE ACCESS   MODE: ANALYZED (BY INDEX ROWID) OF
                     'PERSONS' (TABLE)

    Patrick, interesting insight. Thank you.
    The optimizer must be peeking at my bind variables with it's eyes closed. I'm the only one testing and I've never passed %anything as a bind value. :)
    Here's what I've learned since my last post:
    I don't think that sqldeveloper is actually using the explain plan it says it is. When I run explain plan in sqldeveloper (with a bind variable) it shows me the exact same plan as Apex with a bind variable. However, when I run autotrace in sqldeveloper, it takes a path that matches the hard coded values, and returns results in half a second. That autotrace run is consistent with actually running the query outside of autotrace. So, I think either sqldeveloper isn't really using bind variables, OR it is using them in some other way that Apex does not, or maybe optimizer peeking works in sqldeveloper?
    Using optimizer hints to tweak the plan helps. I've tried both /*+ FIRST_ROWS */ and /*+ index(ps pk_persons) */ and both drop the query to about a second. However, I'm loath to use hints because of the very dynamic nature of the query (and Tom Kyte doesn't like them either). The hints may end up hurting other variations on the query.
    I also tested the query by wrapping it in a select count(1) from ([long query]) and testing the performance in sqldeveloper and in Apex. The performance in that case is identical with both bind variables and hard coded variables for both Apex and SqlDeveloper. That to me was very interesting and I went so far as to set up two bind variable report regions on the same page. One region wrapped the long query with select count(1) from (...) and the other didn't. The wrapped query ran in 0.01 seconds, the unwrapped took 15ish seconds with no other optimizations. Very strange.
    To get performance up to acceptable levels I have changed my function returning query to:
    1) Set the equality operator to "=" for values without wildcards and "like" for user input with wildcards. This makes a HUGE difference IF no wildcard is used.
    2) Insert a /*+ FIRST_ROWS */ hint when users chose the column that requires the sub-query. This obviously changes the optimizer's plan and improves query speed from 15 seconds to 1.5 seconds even with wildcards.
    I will NOT be hard coding any user supplied values in the query string. As you can probably tell by the query, this is an application where sql injection would be very bad.
    Jeff, regarding your question about "like '%' || :P71_SEARCH_SOURCE1 || '%'". I've found that putting wildcards around values, particularly at the beginning will negate any indexing on the column in question and slows performance even more.
    I'm still left wondering if there isn't something in Apex that is breaking the optimizer "peeking" that Patrick describes. Perhaps something in the way it switches contexts from apex_public_user to the workspace schema?

  • How to improve slow PowerPivot performance when adding/modifying measures, calculated columns or Relationships?

    I have been using PowerPivot for a couple of months now and whilst it is extremely quick when pulling in data to populate Pivot Tables, it is extremely slow to make the following kind of changes to the Data Model:
    - Add a Measure / Calculated Field
    - Add a Calculated Column
    - Rename a Calculated Field
    - Re-name a Calculated Column
    - Modify a relationship
    - Change a tables properties
    - Update a table
    In the status bar of excel I get a very quick 'calculating', then it spends a lot of time 'reading data',
    then it 'finalises' after which nothing is in the status bar but it still takes approx. 45 seconds before the program becomes responsive again. This waiting time does not change depending on the action, it is the same if I rename a
    column as it is if I add a new measure.
    My question is what affects performance of these actions and how do I improve it?
    To give you an idea of where my data comes from, I have:
    - 7 tables that feed into the Data Model directly from within the workbook which contains the data model itself. These are a combination of static tables and tables that connect to a MySQL database.
    - 6 separate workbooks which contain static data that is updated manually periodically (copied and pasted from another source)
    - 5 separate workbooks which contain dynamic tables that are linked to our MySQL database and update when opened.
    Now I realise that this is probably where my issue is, however I have no idea how to fix it. You do not seem to be able to connect to a MySQL database directly within the PowerPivot window itself so there is no way to generate and update tables without
    first creating them either in a worksheet or separate workbook (as far as I know).  If I try to create all of the tables directly within the single workbook containing the Data Model I get performance and crashing issues hence why I separate tables into
    individual workbooks.
    Any advice on how to improve performance would be tremendously appreciated. I'm new and keen to learn, I'm aware this set-up is far from best practice.
    Hardware wise I am using:
    - Windows 8 64-bit
    - Excel 2013 64-bit
    - Intel Core i7 processor
    - 6 GB Ram
    Thanks,
    James

    Darren,
    I think the point I was making is its in memory, geez... BTW what do all applications do when they run out of paged memory,  if PowerPivot is using all available memory then wouldn't this force the other applications to use Virtual or essentially write
    back and forth to the disks? I think Virtual memory white to disk ??, lol Also, there are parts if the architecture of Excel 2013 that when importing data into PowerPivot require memory and when working in SharePoint the PowerPivot data is cached to disk
    unless recently refreshed... But this conversation isn't help the James who asked the question and as much as I would love to continue its become a little boring..
    Hi James,
    If you download one the ODBC MySQL Connectors
    http://dev.mysql.com/downloads/connector/odbc/ and I believe yours is the first one for x64 systems and connect directly to the data you should be able to reduce the number of workbooks your opening and if you notice in the following graphic these
    connection are automatically refreshed by default, the parts in red are the differences between PowerPivot 2010 and 2013
    You should notice a lot of improvement especially when refreshing data please let us know how it goes...
    After registering the ODBC Driver
    Click Add. on the User-DSN sheet, choose the “MySQL ODBC 5.x driver”, fill in the credentials, choose a database (from the select menu) and a data source name and you’re done.
    Back in Excel you go on the PowerPivot section of the ribbon and open the PowerPivot window  (the green icon on the left side). In the ‘Home’ section of that window you will see a small gray cylindrical symbol (the international
    symbol for “database”) which will suggest to you different data sources to choose from. Take the one where it says “ODBC”.
    In the next dialog you click on create, choose the adapter, and then Ok. Back in the assistant you can check the connection and proceed.
    Now you have the choice between importing the data from tables using the import assistant or Query depends on your skillset..
    Cheers,
    Ivan
    Ivan Sanders <a href="http://www.linkedin.com/in/iasanders">My LinkedIn </a> , <a href="http://msmvps.com/blogs/ivansanders">My Blog</a>, <a href="http://twitter.com/iasanders"> @iasanders</a>,
    <a href="http://shop.oreilly.com/product/0790145372703.do">BI in SP2013</a>, <a href="http://sharepointdemobuilds.codeplex.com">SP2013 Content Packs</a>.

  • Interactive report error when filter

    Hi,
    I'm new to Apex and get the following error when I create an interactive report:
    "Invalid set of rows requested, the source data of the report has been modified.
    reset pagination"
    This error occurs when I scroll through the pages of the report and get to say the third page of data. I then apply a filter to restrict the data returned to less than a page worth of data, click "Go" and get the error.
    I've searched the forum and understand that this is occurring as Apex is still trying to display 3 pages of data when there is only one page worth of data and I know that I somehow need to reset the pagination, but I cannot find a way to do this. I have tried creating a process to reset the pagination, but this does not seem to fire when the "Go" button is clicked on the interactive report search bar.
    I have tried to reproduce this in my online Apex workspace, but it works fine there. I think this is probably down to using version 4.0 online, but my company is still on version 3.1.0.00.32.
    Is this a bug in the version of Apex that I'm using or is there a way that I can reset the pagination on an interactive report (when the "go" button is clicked)?
    thanks
    Adrian

    The best thing is to run the page with debug and see if the reset pagination is firing or not. Its been a long time since I used 3.1.x but I think reset pagination 3.1 works.
    Maybe something to do with the condition on the reset pagination that prevents it from running? Make reset pagination 'unconditional' and see the result.
    Regards,

  • BPC 7.5NW Report Issues When Requesting Year To Date Data

    Dynamic reports built with 2 row expansions (Cost Ctr, Account) up to 3 columns (not expansions, set to different categories with excel formulas for variance calculations...Act vs. Bud for example) result in error messages (error retrieving data and another related error messge) when requesting YTD data vs. Periodic data.  The report runs fast and properly with Periodic measure.  When running with the YTD measure, attempts to reduce the report data volume by removing comparative categories and removing formatting (after ranges) still results in YTD data retrieve errors.
    Has anyone else had similar issues with BPC 7.5NW, SP9, and YTD data not pulling into reports?  What are the common causes for issues with YTD data retrieval?  Given the design by SAP of the BPC NW product and promotion of using BPC for both Periodic and YTD data, should I expect this report to work properly, meaning that I can get output with YTD data without long runtimes or errors that result in terminating the report?  Thank you.

    Hi,
    This would do it :SQL> !cat q.sql
    with t (dt, val) as (
         select to_date('2011/01/04','yyyy/mm/dd'), 123 from dual union all
         select to_date('2011/04/09','yyyy/mm/dd'), 234 from dual union all
         select to_date('2011/06/28','yyyy/mm/dd'), 345 from dual union all
         select to_date('2011/10/18','yyyy/mm/dd'), 456 from dual union all
         select to_date('2011/12/23','yyyy/mm/dd'), 567 from dual union all
         select to_date('2012/01/07','yyyy/mm/dd'), 678 from dual union all
         select to_date('2012/05/13','yyyy/mm/dd'), 789 from dual union all
         select to_date('2012/07/19','yyyy/mm/dd'), 890 from dual union all
         select to_date('2012/08/30','yyyy/mm/dd'), 901 from dual union all
         select to_date('2012/09/22','yyyy/mm/dd'), 012 from dual
    ------ end of sample data ------
    select *
    from t
    where dt >= trunc( add_months(trunc( &&dateexec. ,'month'),-1) ,'year')
    and dt < trunc( &&dateexec. ,'month')
    undefine dateexecThe dateexec parameter is just here to "simulate" sysdate of different points in year.
    <i>(your actual query would have sysdate where the dateexec parameter appears)</i>
    SQL> @q
    Enter value for dateexec: sysdate
    DT                         VAL
    07/01/2012 00:00:00        678
    13/05/2012 00:00:00        789
    19/07/2012 00:00:00        890 sysdate does the current year up to previous month.
    SQL> @q
    Enter value for dateexec: to_date('20120122','yyyymmdd')
    DT                         VAL
    04/01/2011 00:00:00        123
    09/04/2011 00:00:00        234
    28/06/2011 00:00:00        345
    18/10/2011 00:00:00        456
    23/12/2011 00:00:00        567when ran in january, it retrieves all the previous year.

  • Mac mini VPN headless, slow graphics performance when no monitor plugged in

    I am using a Mac mini, headless, with no monitor, keyboard or mouse, via VPN from a Windows machine over a LAN. Everything works fine and setting everything up was a breeze. However, something very odd I noticed was that if my Mac mini does not have a monitor plugged in to its video output, the rendering of the graphics over the VPN is very slow. It is most noticable when hovering over the dock and activating the animations etc, though it is noticable all the time. It's so bad that it would be IMPOSSIBLE to watch a video clip or simple animation. It's as though frames are being dropped. Now, this can be immediately fixed if a monitor is plugged into the Mac mini and turned on, and the VPN to the Mac mini from my Windows machine works flawlessly and the graphics are rendered perfectly, with little to no delay or frame skipping.
    I have a hunch that the Mac could be disabling some kind of graphics acceleration when the monitor is not plugged in, but I am not a Mac expert by far, this is my first Mac experience. It was purchased solely for the VPN and headless use, so if I can't get it working properly, it's a waste.
    Note, I have tried several VPN client/viewer combinations and they do not change the situation.
    Thanks in advance.
    Message was edited by: HopkinFrog
    Also, when I try to use the remote with the Mac mini (pressing the menu button), over VPN with no monitor attached to the machine, I simply get a white blank screen until I press the menu button again.

    I must have been HALF ASLEEP when I wrote this.
    I do not mean VPN, I mean VNC!!. Sorry for the confusion.

  • How to Suprass a report layout when there is no data

    Hi Community,
    I have a Sql Query which Gives me the list of pre-treatments which doesn't have the corresponding post treatment
    This query sometimes doesn't not return any rows
    At that time is there a way that i can suprass the the whole layout from printing on the report
    i am using oracle 6i Reports
    Please help me out!!!!

    Create a summary column (lets say "CS_TotalRecords") at report level, and set the following properties:
    Function: Count
    Source: column which can not be null (if report displays some data)
    Reset At: Report
    Create a format trigger at most outer frame (the frame you want to hide and all object within that), and write following code:
    BEGIN
    IF NVL(:CS_TotalRecords,0) = 0 THEN
    RETURN (FALSE);
    END IF;
    RETURN (TRUE);
    END;
    Hope this helps.

  • Slower/worse performance when publishing for deployment vs testing

    When I package my app for deployment (ad hoc right now) the performace of the game drops dramatically.  It's almost unplayable.  In debugging it works much better.  What's different about this build process?  Why does it take longer to publish and give worse results?
    I've tried with iphone packager and with ADT 2.6.

    What about the disk layout? Do you have the same or fewer spindals? Are the disk units rating (average IO, sustained IO) the same or better? What about the total load on the machine? The network card?
    There are many factors that have to be considered. Getting back to just Oracle itself:
    Were any database parameters changed?
    Were the statistics recalculated after the move? In fact, how was the migration accomplished: copied files vs exp/imp?
    Were system statistics in use and if so were they regathered on the new machine?
    Looking at each of the above may lead you to the problem.
    HTH -- Mark D Powell --

  • Extremely Slow Report Performance

    I have a report that I have been using for several years with no problems until recently.  I made a change to the stored procedure that the report uses, and now the report will take several hours to run; whereas previously it took less than a minute.  The change to the proc was very minor and involved only changing some verbage on a static text field that the procedure returned.  The output of the procedure was not changed, and I can run the procedure using SQL management studio in less than 2 seconds.  However, as I said Crystal Report takes several hours to run this same stored procedure.  I even tried creating a brand new report and using this procedure as the datasource, and it did the same thing.  This makes absolutely no sense to me.  Can anyone tell me why this is happening?  I'm using Crystal Reports 2008 with sp2.

    The verify database takes just as long as running the report itself, so it doesn't help.  The portion of the proc that was changed was as follows:
    This:
    set @bodyBlock2 =
    'NOTA:  En caso de no tener la(s) devolucion(es) solicitada(s), favor de ' +
    'anotar en la presente el motivo por el cual no se est entregando la ' +
    'devolucion.
    Para cualquier duda y/o aclaracion al respecto, quedo de usted en la linea ' +
    'de Help Desk
    58 64 17 07 or 01 800 509 17 84, lada sin costo.'
    Was changed to this:
    set @bodyBlock2 =
    'NOTA:  En caso de no tener la(s) devolucion(es) solicitada(s), favor de ' +
    'anotar en la presente el motivo por el cual no se est entregando la ' +
    'devolucion.
    Para cualquier duda, aclaración o reporte ponemos a su disposición los siguientes
    teléfonos 01800 509 1784, 01800 581 2490, lada sin costo o al (55) 5864 1700
    extensiones 1757, 1802, 1811, 1791, 1707, 1726, 1762, 1767, 1781, 1792, 1820 o 1838'

  • Header Fields are not displayed when there is no data in the report

    Hi,
    I am having this strange situation in my webi report.
    When I refresh my report and when there is no data, column names are not being displayed in my report. All it is showing is two empty rows. When I have data, then report is displaying data along with column headings.
    Please suggest me where I am going wrong.
    Thanks,
    VenMen

    Thanks for the update.
    Can you try this.
    Run the report first time for some data.
    When the report is rendered select the table that is displayed.
    On the left pane side properties tab.
    Expand display and check all the options
    Show rows with empty measures
    show rows with empty dimensions
    Show when empty
    Hope this will help
    Regards
    Kultar

  • Report fetched by windows OS report "Performance by System"

    When we fetch the report from the reporting pane  in "Windows server operating server reporting" (Performance by System), sometimes the data shown in the column form comes in the yellow and red color (which shows warning and critical respectively).
    My question is the threshold for these performance rules are mentioned in those perfmon collection rule or the rules from which the alert gets generated.
    Appreciate your help.

    The threshold of these performance rules, click on the alert and display details then show threshold of this alerts.
    For more details, you can refer below link
    http://technet.microsoft.com/en-us/library/hh457556.aspx
    http://technet.microsoft.com/en-us/library/cc180267.aspx
    http://download.doubletake.com/_download/dt53/docs/RecoverNow/User%27s%20Guide/Content/SCOM.htm
    Please remember, if you see a post that helped you please click "Vote As Helpful" and if it answered your question, please click "Mark As Answer"
    Mai Ali | My blog: Technical | Twitter:
    Mai Ali

  • Interactive report and default-filter

    ahoj!
    i have a question regarding interactive reports and default-filter for a date: is it possible to use the current date for the default-filter? apex need 'dd.MM.yyyy' for the filter... already tried to_date(sysdate, 'dd.MM.yyyy')
    thx in advance,
    christian

    Hi,
    I've just been trying the "in the last" option and had no problems for any number that I entered. Are you just entering 1 into the box? What error do you get?
    I've loaded the page with Debug switched on, and get:
    select
           null as apxws_row_pk,
           "DATE_ID",
           "ATD_DATE",
           "CHECK",
           count(*) over () as apxws_row_cnt
    from (
    select  *  from (
    select
    apex_item.checkbox(1, DATE_ID) "CHECK",
    "DATE_ID",
    "ATD_DATE"
    from "#OWNER#"."ATD_DATES"
    )  r
    where ("ATD_DATE" between systimestamp - (1 * :APXWS_EXPR_1) and systimestamp)
    ) r where rownum <= to_number(:APXWS_MAX_ROW_CNT):APXWS_EXPR_1 would contain the value 1 as that is what I've entered for the filter. My report's sql statement is just the innermost nested select statement, the rest has been added by the IR functionality and the filter.
    Andy

  • Report performance and data quality

    Hi,
    Can someone help give explanations to following questions :
    1.) Does BW Report show how current is my data?
    2.) What are the reason why the performance of my BW Report is slow?
    3.) What are the reason why my BW Report is have missing data?
    4.) Why is my BW Report have incorrect data?
    5.) Why doesnu2019t my BW Report Data match SAP R/3 Data?
    Thanks,
    Milind
    Please do not raise generic questions across multiple forums
    Edited by: Arun Varadarajan on Apr 9, 2010 2:08 AM

    Milind,
    1.) Does BW Report show how current is my data?
    You should be able to see the data currency when you run in the web - which method are you using - BEx or Web...?
    2.) What are the reason why the performance of my BW Report is slow?
    It could be due to anything - please search the forums for the same on how to identify possible performance bottlenecks
    3.) What are the reason why my BW Report is have missing data?
    It depends - Missing data loads etc etc
    4.) Why is my BW Report have incorrect data?
    You should be knowing that...? I can just say that it has incorrect data because ...." The sun rises in the east...".!!! more akin to asking "Why did the chicken cross the road"
    5.) Why doesnu2019t my BW Report Data match SAP R/3 Data?
    You should ask SAP that question...
    Honestly I am not sure what the reason behind such generic questions are.... if you are looking for answers - then you need to be more specific - if these are more like Interview Questions asked to you - I guess you should be able to answer them or ask further questions to clarify the question further....

Maybe you are looking for

  • Single invoice for multiple payment terms

    Hi Friends, I have given 2 different payment terms for 2 line items in single order and the billing is splitted due to this reason. I want to know where the system is checking this control. Because I want to create single Invoive even the payment ter

  • A few FCP to Pr questions

    I'm wondering if there's a tutorial somewhere that highlights the differences between FCP and Pr, and has tips on quickly getting up to speed. Also, I'm perplexed about the red bar in the timeline.  I imported a couple of FCP sequences via XML, and m

  • Installing the VIM Pack from sunfreeware

    Has anybody had issues with this package inside zones? Everything works like a charm, but since this is installed in the /usr/local directory structure no users other than root can access the vim editor. Anybody know a way around this? Edited by: sms

  • Error 1935 when installing VISA 3.2 WITHOUT Framework 2.0 installed

    I want to use a NI PCMCIA-232/2 card. I installed the latest update of NI serial 1.7 for this purpose. (Win XP) This installs also NI VISA 3.2. During the installation of VISA 3.2 I got the folllowing message: Error 1395. An error occured during the

  • Facebook notifications from deleted Facebook account

    Hi all, I am having a problem on my late 2010 MBA running 10.8.2.  The other day in order to do some various posting to facebook i canged one of the accounts in System Pref's from my facebook details to hers (to clarify i removed my account and place