Slow performance of report with Cal year

Experts,
  Hope all is OK . We have  BW appraisal reports, one of which is Appraisal markings by grade report having Calendar year as selection variable.
If I run this report with calendar year as selection parameter , this reports is taking very very long time and is not doing any thing .
If I remove Calendar year in the report, then the report is executing in a reasonable amount of time.
But we need calendar year as a variable in the report. Please help me why with this variable taking very long time ?
I appreciate any advice.
Thanks

Hello Manoj,
Take into consideration creating a aggregate on Calyear.
Your problem is because of  variable usage, your problem is data selection of calyear. If you create and aggregation on this cube by Calyear, selection will be very fast.
Sarhan.

Similar Messages

  • Slow performance refreshing report with multiple subreports in main report

    I am using multiple subreports in my main report that appear to be slowing down the performance of the report considerably. Is there something I can do to increase performance with subreports.

    Hi Brian,
    Subreports are a known performance hog.
    I would suggest going back to the drawing board and rethinking on the layout of the report. The best thing, of course, would be to completely get rid of the subreports by creating a single SQL statement that takes care of all the report's requirements.
    Sometimes it is imperative to have subreports, in which case, you should try to make sure any selection filters from both the Main Report and the Subreport are passed back to the database.
    Also, avoid using Page N of M for better performance - with or without Subreports.
    -Abhilash

  • Performance of reports with subquery

    I am experiencing a performance issue for those reports with subquery. Is there anything I could do to get a better performance in terms of fine tuning my report. such as capturing result (client list) in a filter first instead of using a subquery.
    I'm open to any suggestions.
    Thanks,
    Ted

    Thank you for your suggestion.
    Is there anything I could do on OBIEE side? such as run the subquery first and get it cached before running the main report?
    Subquery would return around 5,000 clients. I use it to subset clients which are not in the subquery.
    Thanks,
    Ted

  • Slow performance displaying mails with attachments

    Hi
    I have the following configuration:
    - Windows XP (SP2)
    - Oracle Connector for Outlook version 10.1.3.0.2
    - Outlook 2003 (SP2)
    The problem I have is it is extremely slow when displaying messages that contain attachments even when I am in offline mode.
    When connected I watch the network connection and there seems to be little activity, but it takes 30-40 seconds to display messages.
    Any ideas what could be wrong with my configuration?
    Derek

    I eventually tracked down the source of the slow performance. Nothing to do with the Oracle Connector for Outlook.
    There was a Norton Anit-Virus Add-In (SAVCORP90) that was causing the issue. When I disabled it the message display performance returned to normal.

  • Slow Performance of JApplet with Firefox Plugin in Linux

    Hi everyone,
    I recently revisited an applet I wrote 8 years ago. It's a simple ticker tape. I need a ticker tape for a larger project I am working on, and I had to convert it from AWT to Swing.
    Anyway, with the new 1.6.0_xx plugin for Firefox 2.0.0.5 under Fedora 7, the performance is slow and the text scrolls by in a choppy manner. However, if I run the mouse over the JApplet area or the Java console itself while it is running, the text scrolls by smoothly. This does not happen with the 1.5.0_xx series plugin, and although the JApplet was compiled with the 1.4.2_15 compiler, I cannot find a plugin from that series that works with Firefox 2.0.0.x to test if the problem is there as well.
    So, can anyone help me out? Is this a known issue with the newest plugin? I don't think there's an error in the code. Both the embedded applet and source code can be found at this URL: http://www.brian-borowski.com/Ticker/
    Thanks in advance for your help.
    -Brian

    I tried to do a simular thing, but I did not notice a performance decrease. However it might be possible that this happens. For example, if you are calling a time critical vi in the second while loop, it will lock all other processes until it is finished. Or maybe you are calling a shared resource (allocate an array, access to a file, using a mutex) that the other loop holds. Until the shared resource is released the first while loop must wait before it can access that resource (memory, file, etc.).
    Without looking at the source code it will be hard to say what is causing it. I recommend to remove part by part subVI's from the second loop, to debug where the problem exists. If you want I can have a look at the code, please reply to this thread.
    ErikvH
    A
    pplications Engineering
    National Instruments
    Attachments:
    Digital_Output.vi ‏56 KB

  • Slow performance of application with 2 while-loops

    I made an application with one while-loop addressing several FP devices which runs as expected in a FP-2015. Loop time is 100 ms.
    If I add another while-loop with a timer, the performance is very slow. How come?

    I tried to do a simular thing, but I did not notice a performance decrease. However it might be possible that this happens. For example, if you are calling a time critical vi in the second while loop, it will lock all other processes until it is finished. Or maybe you are calling a shared resource (allocate an array, access to a file, using a mutex) that the other loop holds. Until the shared resource is released the first while loop must wait before it can access that resource (memory, file, etc.).
    Without looking at the source code it will be hard to say what is causing it. I recommend to remove part by part subVI's from the second loop, to debug where the problem exists. If you want I can have a look at the code, please reply to this thread.
    ErikvH
    A
    pplications Engineering
    National Instruments
    Attachments:
    Digital_Output.vi ‏56 KB

  • Slow performance on iMac with Lion

    The performance of my iMac (24-inch, early 2008, 2.8 GHz Intel Core 2 Duo, 2 GB RAM) has degraded significantly since I upgraded to Lion.  I downloaded Lion from the App store and did a direct install before realizing I could do a "clean install".  I've corrected my disk permissions and have over 180 GB of free memeory on my hard drive.  I've done everything I can think of to help address performance issues except:
    1.  Perform a clean install of Lion
    2.  Upgrade my RAM.
    I've created Lion Bootable DVD in anticipation of needing to do a clean install, but am unsure of what the potential negative impact of this would be (i.e. lost data).  It would seem upgrading my RAM to 4GB would be a good move from what several threads are recommending as 2GB RAM appears to be the bare minimum.
    A question to the group is what are the risks of a clean install and how do I minimize them?  
    Will an upgrade in RAM solve this problem (have others experienced this issue and solved it with this approach)?
    Are there any other things I should consider?
    I love Apple products and really want to like Lion, but so far not impressed as this has been a significantly negative experience.
    Thanks

    Bermudagord wrote:
    I've created Lion Bootable DVD in anticipation of needing to do a clean install, but am unsure of what the potential negative impact of this would be (i.e. lost data).
    A question to the group is what are the risks of a clean install and how do I minimize them?
    A clean install would be adviseable for Lion. Be sure to have a good backup preferably SuperDuper or CarbonCopy Cloner
    When you are in Setup Assistant do your migration. Migrate only apps and data to minimize the chance of any other corruption in your previous system.
    If you want to revert to SL just backup your system in the same manner and use your SL install disk, then migrate what you need.

  • Slow performance of treeTable with 35,000 rows

    I have a page that has a vertical Panel Splitter. Thre first facet of the
    Panel Splitter
    contains a Panel Strech Layout. Its center facet contains a treeTable whose
    value has been
    derived from #{pageFlowScope.mapperUIBean.
    targetTreeModelForCurrentMappingSet}. The targetTreeModelForCurrentMappingSet
    has been populated in the mapperUIBean with over 35,000 rows.
    When I run the page and click on any row (either the first or last row), it
    processes the
    selectionListener method that was declared but it takes about 25-30 seconds
    before it processes it.
    The same thing is true when I drag something into a row.
    I am on jdev 11.1.1.6.1.
    Thanks for any help provided.

    Hi,
    hard to tell. There is nothing else than the description of the behavior, which is not enough for a guess. If the rows are saved in a managed bean in page flow scope, the question is how you get them out into the tree table. If you can intercept this and print time informations, it may give you an idea
    Frank

  • Cal year / Month is not populating correctly in Bex report

    Hi Experts,
    We have a bex query on virtual provider with only one column in it i.e 'cal year / month'. I'm having the problem with this column.
    In the excel output of this report the cal year /month is displaying in this sequence.
    01/0000     02/0000     03/0000     04/0000     05/0000     06/0000     07/0000     08/0000     09/0000     08/2008     09/2008     10/2008     11/2008     12/2008     01/2009     02/2009     03/2009     04/2009     05/2009     03/2010     04/2010     05/2010     06/2010     07/2010     08/2010     09/2010     10/2010     11/2010     12/2010     01/2011     02/2011     03/2011     04/2011     05/2011     06/2011     07/2011     08/2011     09/2011     10/2011     11/2011     12/2011     01/2012     02/2012     03/2012     04/2012     05/2012     06/2012     07/2012     08/2012
    I'm not getting why the first 9 columns its displaying as 01/0000     02/0000     03/0000     04/0000     05/0000     06/0000     07/0000     08/0000     09/0000     .
    It should display as
    08/2008     09/2008     10/2008     11/2008     12/2008     01/2009     02/2009     03/2009     04/2009     05/2009
    06/2009        07/2009        08/2009         09/2009         10/2009        11/2009        12/2009        01/2010         02/2010         03/2010     04/2010     05/2010     06/2010     07/2010     08/2010     09/2010     10/2010     11/2010     12/2010     01/2011     02/2011     03/2011     04/2011     05/2011     06/2011     07/2011     08/2011     09/2011     10/2011     11/2011     12/2011     01/2012     02/2012     03/2012     04/2012     05/2012     06/2012     07/2012     08/2012
    My data source is fetching the correct data including this cal year/ month.
    Thanks in advance for your help.
    Regards,
    Brahma Reddy

    It seems there are some records of "cal year/month" with value 01/0000 02/0000 03/0000 04/0000 05/0000 06/0000 07/0000 08/0000 09/0000
    2 possible solution:
    1) Check the data again, you may simply filter them by restrict the cal year/month to the specific value (01/0000)
    2) restrict the cal year/month, only allow 2008/8 + value in the result.

  • Cal.Year/ Month - Applet - MM.YYYY and  HTML- MMM.YYYY

    Hi All.
    I have created Webi Report on top of Bex Query. 
    While running the webi report Cal.Year/ Month Prompts list of vales showing format as MM.YYYY (01.2014) in Applet version.
    and running the same report with HTML version, Cal.Year/ Month Prompts list of vales showing different format as MMM.YYYY (Jan.2014).
    I have checked at bex Q level Cal.Year/ Month variable have List of values MM.YYYY (01.2014) format.
    Cal.Year/ Month - @
    Bex Q : MM YYYY
    Applet - MM.YYYY  
    HTML- MMM.YYYY
    Need to understand same report Cal.Year/ Month Prompts list of vales showing  different format (Applet - MM.YYYY. HTML- MMM.YYYY).
    is there any way to fix the issue display same format in the both version.

    Hi Sateesh,
    Thanks for reply.
    Are you using Unv /BICS ?  BICS
    Did you checked both applet and HTML version with the same user ? Yes, Same user
    HTML and Applet version report pointed to same BO and BW system ? Yes, Both BO and BW systems are same. (development systems)
    In BEx , variable values shows both text and key and when value selected it only show key in the entry field box .- Yes, bex variable and characteristic defined display as KEY.
    I have created Cal.Year/ Month Prompts at webi level….while refresh the report. Cal.Year/ Month Prompts  list of vales showing  different format (Applet -MM.YYYY (01.2014). HTML- MMM.YYYY (Jan.2014)).

  • Discoverer Viewer Report with Page Items Slow Performance

    I created a report in Discoverer User that has page items and parameters and runs fine (under 3 minutes) in Discoverer User, but when I try and run it in Discoverer Viewer it takes a long time. It took over 70 minutes before manually terminating it. I made a copy of the report and removed the page items and tried to run it again in Discoverer Viewer and now the report runs in under 3 minutes. Any idea why a report with page items would run in Discoverer user fine, but not discoverer viewer?
    I am using Oracle Business Intelligence Plus 10g (10.1.2.55.26)

    As long as you can access the middle-tier server where Discoverer Viewer is running, you should be able to run and view Discoverer worksheets.
    While it is difficult to even start guessing what the problem might be, if the machine you installed Discoverer on has DHCP addressing, you will run into problems.
    Thanks
    Abhinav
    Oracle Business Intelligence Product Management
    BI on Oracle: http://www.oracle.com/bi/
    BI on OTN : http://www.oracle.com/technology/products/bi/
    Discoverer : http://www.oracle.com/technology/products/discoverer/
    BI Software : http://www.oracle.com/technology/software/products/ias/devuse.html
    Documentation: http://www.oracle.com/technology/documentation/appserver1012.html
    BI Samples : http://www.oracle.com/technology/products/bi/samples/
    Blog : http://oraclebi.blogspot.com/

  • Extremely slow performance in comparison to RDC-based report

    Today our application uses the Crystal Reports RDC. In the last
    ten years we have made good experience with Crystal Reports.
    In the next few months we want to migrate our application to
    JAVA.
    Because the reports offer a great reuse-potential, we would like
    to implement them via CR4E.
    But during the first benchmark tests we made a very bad experience.
    Running (Printing) in CR4E the reports needed from 20 up to 480 percent
    more time in comparison to the RDC-based reports.
    When using POJOs need from 25 up to 570 percent more time.
    When printing the reports from within the CR2008 designer there is
    no performance difference.
    The reports have from 4 to 82 pages.
    We didn´t expect such a huge performance difference.
    Is CR4E in general slower than the RDC-based Crystal Reports?
    Or are there any parameters in CR4E in addition to the RDC-based
    solution for tuning the reports?
    Edited by: Rainer Graedtke on Nov 9, 2009 2:44 PM

    Hi Ted,
    thanks for the fast response.
    Is there another "datasource option" with better performance than POJO?
    In the upcoming JAVA-version of our application we would like to "feed"
    our reports from memory, not from the underlying database.
    One more statement concerning our performance problem:
    In our reports we make extensive use of subreports.
    Each report has the minimum of two subreports, one for the pageheader,
    one for the pagefooter.
    This is necessary because we have changing headers and footers
    depending on the contents of the details.
    The report with the largest performance loss contains 29 subreports!
    RDC-based it prints in 10 seconds, Java-based we need over 50 seconds.
    Is there any suggestion how to improve the performance of our reports in
    JAVA/CR4E other than reducing the number of subreports?
    Another question concerning the keycode-problem:
    We have designed the reports in the CR2008 Designer and we want to run
    them under CR4E. Is this a valid approach?
    Sincerely Rainer
    Edited by: Rainer Graedtke on Nov 10, 2009 10:51 AM

  • Slow Performance with large library (PC)

    I've been reading many posts about slow performance but didn't see anything addressing this issue:
    I have some 40000 photos in my catalog and despite generating previews for a group of directories, LR still is very slow in scrolling through the pics in these directories.
    When I take 2000 of these pics and import them into a new catalog - again generating previews, the scroll through the pics happens much much faster.
    So is there some upper limit of recommended catalog size for acceptable performance?
    Do I need to split my pics up by year? Seems counter productive, but the only way to see the pics at an acceptable speed.

    I also have serious performance issues, and i don´t even have a large database catalog, only around 2.000 pictures, the db file itself is only 75 mb big. Done optimization - didn´t help. What i encountered is that the cpu usage of LR 1.1 goes up and STAYS up around 85% for 4-5 minutes after programm start - during that time, zooming in to an image can take 2-3 minutes! After 4-5 minutes, cpu usage drops to 0%, the background task (whatever LR does during that time!) has finished and i can work very smoothly. preview generation cannot be the problem, since it also happens when i´m working in a folder that already has all previews build, close LR, and start again instantly. LR loads and AGAIN i´ll have to wait 4-5 minutes untill cpu ussage has dropped so i can continue working with my images smoothly.
    This is very annoying! I will stop using LR and go back to bridge/acr/ps, this is MUCH much faster. BUMMER!

  • Bug report: Energy Manager causes slow performance...

    Very slow Windows Startup..
    Energy Manager's plans and the problem occured using any plan as I remember, and it wasn't an issue with the maximum processor state. I had heavy disk usage, especially access to the Page File, which is what the person in the second link I posted also described. CPU usage was low. The disk usage was seemingly not related to Energy Manager (Chrome was usually the worst), but removing Energy Manager completely resolved the problem. If you would find it helpful. I'm quite confident that even with Maximum Processor state at %100, CPU usage was low when it was slow, and that disk access was the culprit.
    previously have use 1.0.0.31 or 32 not sure.
    Windows 8.1
    Im using Lenovo Energy Management Driver
    1.0.0.28
    4.6.2014

    I actually have the same problem.  Sorry to work on a old thread, but...
    I've got my Lenovo B50-70 less than 1 month ago, and had switched it to SSD and had set everything up and is working perfectly.  However there are a bunch of lenovo programs that came pre-installed.  This computer is used mainly for work and I really don't like programs that suddenly pop-up and ask you to do stuff, as it is quite distracting, and with Lenovo Energy Manager being one of them so I removed it.  (I did do some research before I remove...)
    Everything was fine, until when I try to switch on my wifi (I usually have it off, as I use LAN most of the time)...  For the B50-70 there is a airplane button on F7, which allow the user to switch wifi and bluetooth on/off.  That button was unresponsive...  After some digging it is actually related to the power energy manager being removed.  So I've downloaded the latest version from Lenovo and it was working again...
    However though, whenever I start my computer it is taking so long to start...  And I tried uninstalling the power energy manager and it solved the problem...  And I am stuck with either having no wifi on/off ability or a really slow boot, which loses the whole purpose of the SSD.  After you enter your username and password it originally takes around 5~10 seconds to login, now it takes over 1 minute.
    I have found this other link, which says Lenovo Power Manager having the same problem.
    https://forums.lenovo.com/t5/ThinkVantage-Technologies/Bug-report-Power-Manager-causes-slow-performa...
    There is a solution, which you can change the registry to switch a function off.  I've tried it and it didn't work, which is not surprising as it's a different program..  But I am hoping that Lenovo can provide a similar solution for Lenovo Energy Manager...

  • Extremely slow performance with ojdbc6.jar on IBM JVMs (Java 6)

    We are consistently seeing slow performance (easily demonstrable by the simplest of test cases) while using ojdbc6.jar on IBM JDKs. Pefrormance is normal when we simply opt for ojdbc14.jar under the same JDK, same java program.
    Works well
    =======
    JRockit, ojdbc6.jar
    JRockit, ojdbc14.jar
    IBM JDK, ojdbc14.jar
    A lot slower
    =======
    IBM JDK, ojdbc6.jar
    All we had to do was to write a simple JDBC access program and measure time taken around
    Connection conn = DriverManager.getConnection(jdbcURL, dbProps);
    The first 3 combinations mentioned above return comparable numbers. The 4th combination takes 3-4 times as much.
    We initially saw this during performance test of our software on WebSphere (also supported on Weblogic). Later was able to eliminate WebSphere and Weblogic as the problem areas and narrow it down to the JVM. Reproduce-able on development machines.
    We have been using Oracle/WLS since BEA used to market 'Tengah' servers. This is the worst we have seen as far as the qaulity of support staff goes. We first level support seems clueless. We filed this case and the support comes back tells us that IBM JDKs are not supported for WLS! Oh, we have been running WLS/AIX/IBM JVM combination only for 5+ years.

    A quick solution to test this would be to do the following:
    mv /dev/random /dev/random.bk
    ln /dev/urandom /dev/random
    Note: The quick solution will disappear after a reboot. You could add it to rc.local to automatically make this change, but I wouldn't recommend it.
    The following is from [http://www.usn-it.de/index.php/2009/02/20/oracle-11g-jdbc-driver-hangs-blocked-by-devrandom-entropy-pool-empty/]:
    "Oracle 11g JDBC driver hangs blocked by /dev/random – entropy pool empty
    On a headless (=without console) network server, the 11g JDBC driver used for (java) application connect may cause trouble. In my case, it refused to connect to the DB without any error, trace or log entry. It simply hung. After several hours, it connected one time, and freezed again. Remote debugging done by the development clarified that it locks after calling SeedGenerator() and SecureRandom().
    Reason: The JDBC 11g needs about 40 bytes of secure random numbers, gathered from /dev/random, to encrypt its connect string.
    But public-available “man 4 random” says:
    When read, the /dev/random device will only return random bytes within the estimated number of bits of noise in the entropy pool. /dev/random should be suitable for uses that need very high quality randomness such as one-time pad or key generation. When the entropy pool is empty, reads from /dev/random will block until additional environmental noise is gathered.
    So far so good, now the question arises: Why does this mystic “entropy pool” runs out of gas?
    The answer is as simple as unsatisfying: because too less entropy “noise” was generated by the system. You can check the “filling level” (maybe zero?) of your pool and the overall size of the pool (usually 4096) by issuing
    cat /proc/sys/kernel/random/entropy_avail
    cat /proc/sys/kernel/random/poolsize
    Hint: /dev/random will deliver one new random number as soon as the pool has reached more than 64 entropy units.
    So why does my box not generate more entropy noise?
    Because only few drivers will fill the entropy pool, first of all keyboard and mouse. Sounds very useful on a server in a datacenter, isn’t it? Some block device and network drivers seem to do so as well, and I have read from guys on the net changing their network card and driver to enjoy this “feature”! But let’s stop ranting, /dev/random is simply made for high security randomness, and if it can’t make sure that randomness is as good as possible in this deterministic world, it stops. Intelligent people have created /dev/urandom for that, like “man 4 random” clearly states:
    A read from the /dev/urandom device will not block waiting for more entropy. As a result, if there is not sufficient entropy in the entropy pool, the returned values are theoretically vulnerable to a cryptographic attack on the algorithms used by the driver. Knowledge of how to do this is not available in the current non-classified literature, but it is theoretically possible that such an attack may exist. If this is a concern in your application, use /dev/random instead.
    Now let’s get back on our JDBC problem. Oracle JDBC 11g seems to use /dev/random by default, which causes usually no trouble on clients running with console access by a user, because his/her unpredictable :) actions will keep the entropy pool well-fed. But to make it usable on a headless server with a latently empty entropy pool, you should do several things, in descending security order (without warranty):
    1. Involve an audio entroy daemon like AED to gather noise from your datacenter with an open microphone, maybe combine it with a webcam noise collector like VED. Other sources are talking about “Cryptographic Randomness from Air Turbulence in Disk devices“. :)
    2. Use the Entropy Gathering Daemon to collect weaker entropy from randomness of userspace programs.
    3. Talk your JDBC into using /dev/urandom instead:
    -Djava.security.egd=file:///dev/urandom"

Maybe you are looking for