Benchmark response times using BIA.

Hi all
I ran some queries using BIA and logged their processing times. Now, I want to compare these processing times with "benchmark response times using BIA" just to make sure that these queries are performing optimally. Could you pls provide pointers/links to the same.
Regards
Niranjana

Following are the steps we do in order to get the benchmark data.
Assume you have Indexed a cube in BIA.
Without BIA:
1. Goto transaction RSRT2-->Select the query and click "Execute + Debug"
2. Under section "BIA Server"-->Check "Donot user BIA Index"
3. Under section "Others"--> Check "Display statistics data" and "Do not user cache"
4. Execute....After query complete click on Back button to see the statistics
5. Sum up the "Duration column" and Subtotal by column "Event Text"
6. We are interested on the item "datamanager" - note down the value
With BIA:
7. Goto transaction RSRT2-->Select the same query and click "Execute + Debug"
8. Under section "BIA Server"-->DONOT Check "Donot user BIA Index"
9. Under section "Others"--> Check "Display statistics data" and "Do not use cache"
10. Execute....After query complete click on Back button to see the statistics
11. Sum up the "Duration column" and Subtotal by column "Event Text"
12. We are interested on the item "datamanager" - note down the value
13. Click on "Aggregation layer"-->You should note the on the "aggregation" column the cube name should be suffixed with "$X" ( which indicates it uses BIA)
14. Here "viewed at" is sum of other two front columns...
In the above way you can benchmark between non-BIA vs BIA query performance. if you see even with BIA the performance is slow...then it could be bcoz of OLAP time...lot of calculation in the frontend etc...
In general we see the performance of BIA is fantastic...except sometimes the numbers are not matching between BIA and non-BIA queries...where we need to rebuild the indexes.
Hope it helps
Gopi

Similar Messages

  • Unable to capture the Citrix network response time using OATS Load testing.

    Unable to capture the Citrix network response time using OATS Load testing. Here is the scenario " in our project users logs into Citrix network and select the Hyperion application and does the Transaction and the Clients wants us to simulate the same scenario for load testing. We have scripted starting from Citrix Login and then launching Hyperion application. But the time taken to launch the Hyperion Application from Citrix network has not been captured whereas Hyperion Transaction time have been recorded. Can any help to resolve this issue ASAP?

    Hi keerthi,
    1. I have pasted the code for the first issue
    web
                             .button(
                                       122,
                                       "/web:window[@index='0' or @title='Manage Network Targets - Oracle Communications Order and Service Management - Order and Service Management']/web:document[@index='0' or @name='1824fhkchs_6']/web:form[@id='pt1:_UISform1' or @name='pt1:_UISform1' or @index='0']/web:button[@id='pt1:MA:0:n1:1:pt1:qryId1::search' or @value='Search' or @index='3']")
                             .click();
                        adf
                        .table(
                                  "/web:window[@index='0' or @title='Manage Network Targets - Oracle Communications Order and Service Management - Order and Service Management']/web:document[@index='0' or @name='1c9nk1ryzv_6']/web:ADFTable[@absoluteLocator='pt1:MA:n1:pt1:pnlcltn:resId1']")
                        .columnSort("Ascending", "Name" );
         }

  • Response times using BEX/Web vs. BEX /Excel user interface

    Our customer in Sweden is currently rolling out SAP BW to local sales offices around the globe and they are having huge response time problems with the BEX/Excel interface running from Australia towards the BW servers in Sweden. If running the Web interface the response time is acceptable.
    Question: Is this problem caused purely by bandwidth limitations or is it the distance itself that is the problem. If they double their bandwidth capacity, will the response times using the BEx web interface be reduced by 50% or will it remain unchanged?
    /Fredrik Osterstrom, Capgemini,

    Hi,
    It is generally much slower to display queries using Excel than the HTML version because of the way they display data and the way they are designed/engineered.
    But generally of course, it should also be on an acceptable basis.
    If you still  could not figure out the problem, I advice you  to display the query using HTML version via Browser and then download to Excel aftewards. It is normally fast. This is just a work around till you figure out the probable cause. There are many factors that contribute to this.
    Try isolation technique.
    1.) Is the query slow on displaying using Excel on all desktop computers or only selected computers?
    2.) Is the desktop computer fast enough w/ ample memory to handle the processing?
    3.) Are there any programs running/installed on the computer which directly affects Excel processing?
    Hope I have helped, and if I did, please grant points!
    --Jkyle

  • Experiencing very slow response time using AirPort while wired response is fine. Suggestions?

    After a surfing session a few weeks ago the response time for my AirPort internet connection became painfully slow.  When I plug in to the wired connection it works fine.  Resetting the router and modem doesn't seem to fix it.  What might have happened and what can I do to fix it?

    It is very possible that you may have some form of Wi-Fi interference that appears during these hours that is preventing your AirPort Base Station from providing a clean RF signal.
    I suggest you perform a simple site survey, using utilities like iStumbler, Wi-Fi Explorer or AirRadar to determine potential areas of interference, and then, try to either eliminate or significantly reduce them where possible.

  • Slow response time for HTTP requests

    I have an 8310 Pearl. For the past week, I've noticed that the response time using the Internet browser was very slow. One work-around was to cancel the request and then submit it again. The second request returned the page MUCH faster.
    I cleared my history, and this seemed to have helped. Has anyone else noticed this? Thanks

    Welcome to the forums
    Try a Battery Pull:
    1. Remove the battery while phone is turned on
    2. Wait a few seconds before inserting it again
    3. Allow the long boot up process to finish (takes a few minutes usually)
    This clears most software issues that occur on blackberries from time to time. You could also use Quickpull, a free app that simulates a battery pull without physically removing the battery: http://software.crackberry.com/product.asp?id=27660
    A battery pull reclaims memory lost during application usage and clears minor software errors that creep in over time. The blackberry runs an OS just like your PC runs windows and the battery pull is the equivalent of a reboot/restart on your computer.
    Kijana
    Please remember to:
    1. Mark Accept as Solution on the appropriate post once your issue has been resolved
    2. Give Kudos to helpful posts (click the star next to the post)
    Thanks

  • Portal Page Response time

    Hi All,
    Enter the portal page url in IE,
    I observed portal page response time using the HTTPWatch, it was taking 6 to 7 mintes to get response .
    What are the things to do improve the response time.
    Please suggest me possible solutions.
    Thnaks,
    Venkata Sarvabatla

    Too big a topic :). Involves download a profiler (commercial ones like JProbe or OptimizeIt or YourKit or ClearApp or free ones available on a google search) which can profile CPU execution and installing and configuring your JVM to send data to it. Then running the problematic part of your application, the profiler will tell you how long each method call took, how many calls were made, the call graphs etc. Once you know which part of your application is taking time , you would then inspect the code and see what the problem is. Im assuming you have already done your rule of thumb tuning (on VM, OS, DB etc).
    Google will give you plenty of examples, the tool you choose will also be important , whether you already have a commercial profiler available to your project , or whether youll need to configure some open source one.
    This is my writeup for TPTP (free open source)
    http://theworkaholic.blogspot.com/2009/10/profiling-bea-weblogic-portal-apps.html
    If you havent profiled/tuned before , i suggest you find someone on your project who has and watch them while they do it. Setting up profilers on app servers and writing tests is somewhat painful the first time around and you can learn much more from someone else.

  • Ranking response times for Fire Department

    Hello,
    I'm in the middle of powerpivot pro's online course, but I'm looking for a little feedback to see if I'm on the right track with a project. I work for a fire department and we collect the response times of our units. As you can guess, we have goals we measure
    against and I'm hoping powerpivot will be a good choice to work with. Simply put, for each incident (fire, medical) I create a response time using a calculated column in powerpivot window subtracting the arrival time from the dispatch time (format HH:MM:SS).
    I am populating the workbook with a database connection to our SQL database. I group the incidents using the unique incident number in a pivot table, so it list every unit that was on the run. My idea would be to create a measure listing each unit's response
    time, looking for the minimum time and compare the minimum against our goal. The result returns either true or false. 
    I"m working on the solution, but I'm just wondering if my idea sound plausible using powerpivot?
    I'm not looking for anyone to do my work, but looking for any suggestions to push me in the right direction.
    Thank you for looking , Brent

    Hello Paul, 
         The calendar table was something I just put in because of a lesson I had with Rob's course, so I was in the processing of adding it to the project as a means of reinforcing the lesson. I will take another look at our SQL database, but
    in the database, the incident and apparatus tables are primarily linked by the incident key. I will double check to see if the apparatuses in the table have another relationship with a look up table and I can change the data model. Initially, I was looking
    for the simplest model to start with and then build from there. 
          As far as the error, the min function is working correctly to find the minimum time of the incident, but I think my confusion comes with the grouping of the rows. As you can probably tell, there is one incident key for each fire call
    and many apparatus are assigned to the call with the incident key as the primary relationship. If I add all the incidents to the pivot table, it correctly identifies each incident as being met or not met. But, if I remove the incident keys from the pivot and
    just keep the months, the pivot table correctly filters down to the month and find the minimum of the entire month and I loose the details of each incident's value of met or not met. Thank to Rob, I understand what is happening, but not how to fix it (I'm
    finishing up the calendar table now and the X factors are next, which will probably help me with this issue).
    The reporting for now, is pretty simplistic. It is being able to identify each incident, if it met the goal and then able to report out of the incidents how many we made and how many we missed. If we had 1,000 calls and we made 500 of them, we would be at
    50%. The goal is a simple true or false based on the minimum time of the first arriving unit ( I used 200 seconds for this project). 
    I do have Excel 2013, so I can open the file. I created the original at home with 2010. 
    I hope that make sense.  Thank you for the help. I have the day off and I'm working through Rob's video course most of the day today.  Thanks, Brent

  • Report to calculate avg response time for a transaction using ST03.

    Hi Abap Gurus ,
    I want to develop a report which calculates the average response time(ST 03) for a transaction on hourly basis.
    I have read many threads like in which users are posting which tables/FM to use to extract data such as  dialog step and total response time .
    I am sure many would have created a report like this , would appeciate if you can share pseudo code for same.Any help regarding the same is highly appreciated...
    Cheers,
    Karan

    http://jakarta.apache.org/jmeter/

  • Why is the response time when using clickwheel...

    I've bought a new 60gb video and, in comparison to my 20gb 3rd generation, the response time when using the clickwheel (next track/pause/play/bak track) is annoyingly long on start-up. And the display takes seconds to catch up with what's playing sometimes freezing with half of the one 'page' showing and half of another. Should I expect this because of the larger capacity/more complex iPod?
    Also, in an earlier post I noted that when the iPod first fires up, tracks seem to stop and start - just for a second or so, 3 or 4 times within the first 15 seconds. It's not the track on iTunes as it never happens on my 3g.
    Dell   Windows XP  

    Definately sounds like a faulty harddrive.
    You could always try restoring it with the most recent updater, and reloading everything - kind of like a fragmented harddrive, I am noticing better performance since I last wiped it clean and relaoded it (due to another error - actually)....

  • "Windows 8 using 100% of HDD with high average response times and low read/write speed"

    Turns out this is a fairly well known windows 8.1 issue that has been plaguing users since at least 2013 and there is no one simple fix, so it may not be *entirely* hp's fault; but I've had two of these laptops, both the same model, the first one needing to be returned and exchanged for an entirely unrelated issue (hardware failure: ethernet port nonfunctional with lights stuck on). Both are refurbished. Both have been extreemly slow and unresponsive even compared to a lesser Brazos powered laptop I had before, but I've only recently decided to investigate why. 
    So if there is something HP specific going on here, I hope there is one simple fix. My average response time has gone up as far as well over a minute (>60,000ms), so I may be an outlier case here compared to the typical windows 8.1 hardrive responsiveness/bandwidth problem. 
    Edit: there is a case with another HP pavilion laptop (intel powered though, so it may be an intel storage driver issue described in the first link) being much, much worse.
    This question was solved.
    View Solution.

    Guess what just now happened again
     So using DISM did _not_ fix it.

  • Any way to improve response time with iPhoto 8.1.2 using Mountain Lion ?

    Any way to improve response time with iPhoto 8.1.2 using Mountain Lion ?  Can you store photos on a separate hard drive and use a smaller file for openning iphoto?

    How did you move your iPhoto library to the new system?  the recommended way is Connect the two Macs together (network, firewire target mode, etc)  or use an external hard drive formed Mac OS extended (journaled) and drag the iPhoto library intact as a single entity from the old Mac to the pictures folder of the new Mac - launch iPhoto on the new mac and it will open the library and convert it as needed and you will be ready move forward.
    LN

  • Response time https/http using CSM+SSL versus CSM

    Im using C6K and CSM for load balancing and an SSL module for https termination. That means clients can access the serverfarm through https/http and compare response time values.
    What should be the expected delay generated from SSL(https) compared to cleartext(http) ?
    My customers are complaining about to much delay...
    Thanks in advance for your answer.
    Volker

    it's difficult to answer.
    There will be an initial delay required to establish the ssl session, then for each packet there will be additional delay to encrypt/decrypt the packet.
    It also depends on load.
    An CSM can handle lot of traffic but not an SSLM.
    So, if you see something wrong in a sniffer trace like packet drops, retransmits, huge delays per packet [like 300ms], then we need to investigate what is going on.
    Otherwise, this would be normal.
    Regards,
    Gilles.

  • Mac system, When I use shockwave, the response time is very slow. I have updated Shockwave.

    I am running OS X 10.6.8. I am trying to do a jigsaw puzzle, but the cursor movement is slow and erratic. I have updated Adobe Shockwave, but still have the problem. The last time I updated Firefox I had the same problem and went back to the old version. Any help would be greatly appreciated.

    tr3v wrote:
    Thanks for replying.
    1/ No, but I am running Vista, so assume this will be the same? Or should I just look at the tmp and temp environment variables to see what folders are being used?
     In Vista type temp in the Search programs and files box and double click on the temp files icon that appears above. Delete all the files in the folder that you can. It is safe as they are exactly what they are called and that is temporary files. If you haven't done this ever or in quite a while there should be a noticeable increase in the operating system's response time.
    2/ Yes. It completes after a very long time. 
    5/ No - but will try this out too. 
    ****Please click on Accept As Solution if a suggestion solves your problem. It helps others facing the same problem to find a solution easily****
    2015 Microsoft MVP - Windows Experience Consumer

  • Using trace response time much faster than without tracing

    Oracle 10.2.0.3 64bit for windows 2003
    one very simple statement, "delete from taba where id=**"
    when run it without tracing, took about 2 mins
    when adding 10046 trace, the response time is less than 2 seconds
    repeated this dozens times, all the same problem.
    anyone has this problem before?
    appreciate your input.

    Do you have trigger for DELETE statement on this table ?yes, 4 of them behind it
    Do you have some dependent object like materialized view using this table ?no
    Is this a single instance database or RAC database ?single instance
    Do you have concurrent DML running on this table when runnnig this tests ?no
    below if the tkprof
    TKPROF: Release 10.2.0.3.0 - Production on Thu Jan 6 12:31:57 2011
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Trace file: pacs_ora_5800.trc
    Sort options: default
    count    = number of times OCI procedure was executed
    cpu      = cpu time in seconds executing
    elapsed  = elapsed time in seconds executing
    disk     = number of physical reads of buffers from disk
    query    = number of buffers gotten for consistent read
    current  = number of buffers gotten in current mode (usually for update)
    rows     = number of rows processed by the fetch or execute call
    error connecting to database using: rpt3/rpt3
    ORA-12560: TNS:protocol adapter error
    EXPLAIN PLAN option disabled.
    delete TABLEAs
    where
    id=:"SYS_B_0"
    call     count       cpu    elapsed       disk      query    current        rows
    Parse        1      0.00       0.00          0          0          0           0
    Execute      1      0.00       0.01          0          3         13           1
    Fetch        0      0.00       0.00          0          0          0           0
    total        2      0.00       0.01          0          3         13           1
    Misses in library cache during parse: 1
    Misses in library cache during execute: 1
    Optimizer mode: ALL_ROWS
    Parsing user id: 36 
    Rows     Row Source Operation
          1  DELETE  TABLEAS (cr=4 pr=0 pw=0 time=706 us)
          1   INDEX UNIQUE SCAN PK_TABLEA_ID (cr=3 pr=0 pw=0 time=39 us)(object id 23307)
    Elapsed times include waiting on following events:
      Event waited on                             Times   Max. Wait  Total Waited
      ----------------------------------------   Waited  ----------  ------------
      SQL*Net message to client                       1        0.00          0.00
      SQL*Net message from client                     1        3.60          3.60
    INSERT INTO TMP_TABLEAS_TRIGGER
    VALUES
    (:B1 )
    call     count       cpu    elapsed       disk      query    current        rows
    Parse        1      0.00       0.00          0          0          0           0
    Execute      1      0.00       0.00          0          1          7           1
    Fetch        0      0.00       0.00          0          0          0           0
    total        2      0.00       0.00          0          1          7           1
    Misses in library cache during parse: 0
    Optimizer mode: ALL_ROWS
    Parsing user id: 36     (recursive depth: 1)
    SELECT DISTINCT ID
    FROM
    TMP_TABLEAS_TRIGGER
    call     count       cpu    elapsed       disk      query    current        rows
    Parse        1      0.00       0.00          0          0          0           0
    Execute      1      0.00       0.00          0          0          0           0
    Fetch        2      0.00       0.00          0          3          0           1
    total        4      0.00       0.00          0          3          0           1
    Misses in library cache during parse: 0
    Optimizer mode: ALL_ROWS
    Parsing user id: 36     (recursive depth: 1)
    Rows     Row Source Operation
          1  HASH UNIQUE (cr=3 pr=0 pw=0 time=158 us)
          1   TABLE ACCESS FULL TMP_TABLEAS_TRIGGER (cr=3 pr=0 pw=0 time=37 us)
    UPDATE TABLEAS SET SUPERCEDES = 0
    WHERE
    ID IN (SELECT ID FROM TABLEAAA WHERE AAID IN(SELECT AAID FROM
      TABLEAAA WHERE ID=:B1 ) ) AND SUPERCEDES=:B1
    call     count       cpu    elapsed       disk      query    current        rows
    Parse        1      0.00       0.00          0          0          0           0
    Execute      1      0.12       0.11          0      50411          2           1
    Fetch        0      0.00       0.00          0          0          0           0
    total        2      0.12       0.11          0      50411          2           1
    Misses in library cache during parse: 0
    Optimizer mode: ALL_ROWS
    Parsing user id: 36     (recursive depth: 1)
    Rows     Row Source Operation
          0  UPDATE  TABLEAS (cr=50411 pr=0 pw=0 time=117980 us)
          1   NESTED LOOPS  (cr=50411 pr=0 pw=0 time=117906 us)
      16731    VIEW  VW_NSO_1 (cr=217 pr=0 pw=0 time=17208 us)
      16731     SORT UNIQUE (cr=217 pr=0 pw=0 time=17206 us)
      16731      TABLE ACCESS BY INDEX ROWID TABLEAAA (cr=217 pr=0 pw=0 time=33524 us)
      16733       NESTED LOOPS  (cr=71 pr=0 pw=0 time=58 us)
          1        INDEX RANGE SCAN PK_TABLEA_AA (cr=3 pr=0 pw=0 time=30 us)(object id 23313)
      16731        INDEX RANGE SCAN IDX_TABLEAAA_AA_ID (cr=68 pr=0 pw=0 time=44 us)(object id 23316)
          1    TABLE ACCESS BY INDEX ROWID TABLEAS (cr=50194 pr=0 pw=0 time=90408 us)
      16730     INDEX UNIQUE SCAN PK_TABLEA_ID (cr=33464 pr=0 pw=0 time=45248 us)(object id 23307)
    UPDATE TABLEAS SET SUPERCEDEDBY = 0
    WHERE
    ID IN (SELECT ID FROM TABLEAAA WHERE AAID IN(SELECT AAID FROM
      TABLEAAA WHERE ID=:B1 ) ) AND SUPERCEDEDBY=:B1
    call     count       cpu    elapsed       disk      query    current        rows
    Parse        1      0.00       0.00          0          0          0           0
    Execute      1      0.10       0.11          0      50411          2           1
    Fetch        0      0.00       0.00          0          0          0           0
    total        2      0.10       0.11          0      50411          2           1
    Misses in library cache during parse: 0
    Optimizer mode: ALL_ROWS
    Parsing user id: 36     (recursive depth: 1)
    Rows     Row Source Operation
          0  UPDATE  TABLEAS (cr=50411 pr=0 pw=0 time=117770 us)
          1   NESTED LOOPS  (cr=50411 pr=0 pw=0 time=117660 us)
      16731    VIEW  VW_NSO_1 (cr=217 pr=0 pw=0 time=17251 us)
      16731     SORT UNIQUE (cr=217 pr=0 pw=0 time=17250 us)
      16731      TABLE ACCESS BY INDEX ROWID TABLEAAA (cr=217 pr=0 pw=0 time=33528 us)
      16733       NESTED LOOPS  (cr=71 pr=0 pw=0 time=63 us)
          1        INDEX RANGE SCAN PK_TABLEA_AA (cr=3 pr=0 pw=0 time=38 us)(object id 23313)
      16731        INDEX RANGE SCAN IDX_TABLEAAA_AA_ID (cr=68 pr=0 pw=0 time=53 us)(object id 23316)
          1    TABLE ACCESS BY INDEX ROWID TABLEAS (cr=50194 pr=0 pw=0 time=90462 us)
      16730     INDEX UNIQUE SCAN PK_TABLEA_ID (cr=33464 pr=0 pw=0 time=45175 us)(object id 23307)
    DELETE FROM TABLEAAA
    WHERE
    ID = :B1
    call     count       cpu    elapsed       disk      query    current        rows
    Parse        1      0.00       0.00          0          0          0           0
    Execute      1      0.00       0.00          0          3         10           1
    Fetch        0      0.00       0.00          0          0          0           0
    total        2      0.00       0.00          0          3         10           1
    Misses in library cache during parse: 0
    Optimizer mode: ALL_ROWS
    Parsing user id: 36     (recursive depth: 1)
    Rows     Row Source Operation
          0  DELETE  TABLEAAA (cr=4 pr=0 pw=0 time=626 us)
          1   INDEX RANGE SCAN PK_TABLEA_AA (cr=3 pr=0 pw=0 time=39 us)(object id 23313)
    INSERT INTO TMP_TABLEAAA_RECORDS (AAID, STATUS,TABLEAID)
    VALUES
    (:B3 , :B2 ,:B1 )
    call     count       cpu    elapsed       disk      query    current        rows
    Parse        1      0.00       0.00          0          0          0           0
    Execute      1      0.00       0.00          0          1          7           1
    Fetch        0      0.00       0.00          0          0          0           0
    total        2      0.00       0.00          0          1          7           1
    Misses in library cache during parse: 0
    Optimizer mode: ALL_ROWS
    Parsing user id: 36     (recursive depth: 2)
    SELECT DISTINCT AAID, STATUS,TABLEAID
    FROM
    TMP_TABLEAAA_RECORDS
    call     count       cpu    elapsed       disk      query    current        rows
    Parse        1      0.00       0.00          0          0          0           0
    Execute      1      0.00       0.00          0          0          0           0
    Fetch        2      0.00       0.00          0          3          0           1
    total        4      0.00       0.00          0          3          0           1
    Misses in library cache during parse: 0
    Optimizer mode: ALL_ROWS
    Parsing user id: 36     (recursive depth: 2)
    Rows     Row Source Operation
          1  HASH UNIQUE (cr=3 pr=0 pw=0 time=162 us)
          1   TABLE ACCESS FULL TMP_TABLEAAA_RECORDS (cr=3 pr=0 pw=0 time=35 us)
    SELECT COUNT(*)
    FROM
    TABLEAAA WHERE AAID = :B1
    call     count       cpu    elapsed       disk      query    current        rows
    Parse        1      0.00       0.00          0          0          0           0
    Execute      1      0.00       0.00          0          0          0           0
    Fetch        1      0.01       0.00          0         68          0           1
    total        3      0.01       0.00          0         68          0           1
    Misses in library cache during parse: 0
    Optimizer mode: ALL_ROWS
    Parsing user id: 36     (recursive depth: 2)
    Rows     Row Source Operation
          1  SORT AGGREGATE (cr=68 pr=0 pw=0 time=2936 us)
      16730   INDEX RANGE SCAN IDX_TABLEAAA_AA_ID (cr=68 pr=0 pw=0 time=8 us)(object id 23316)
    UPDATE TABLEAS SET SUPERCEDES = 0
    WHERE
    ID IN (SELECT ID FROM TABLEAAA WHERE AAID IN(SELECT AAID FROM
      TABLEAAA WHERE AAID=:B2 ) ) AND SUPERCEDES=:B1
    call     count       cpu    elapsed       disk      query    current        rows
    Parse        1      0.00       0.00          0          0          0           0
    Execute      1      0.29       0.29          0     121996          0           0
    Fetch        0      0.00       0.00          0          0          0           0
    total        2      0.29       0.29          0     121996          0           0
    Misses in library cache during parse: 0
    Optimizer mode: ALL_ROWS
    Parsing user id: 36     (recursive depth: 2)
    Rows     Row Source Operation
          0  UPDATE  TABLEAS (cr=121996 pr=0 pw=0 time=295237 us)
          0   HASH JOIN SEMI (cr=121996 pr=0 pw=0 time=295234 us)
          0    TABLE ACCESS FULL TABLEAS (cr=121996 pr=0 pw=0 time=295131 us)
          0    VIEW  VW_NSO_1 (cr=0 pr=0 pw=0 time=0 us)
          0     MERGE JOIN CARTESIAN (cr=0 pr=0 pw=0 time=0 us)
          0      TABLE ACCESS FULL TABLEAAA (cr=0 pr=0 pw=0 time=0 us)
          0      BUFFER SORT (cr=0 pr=0 pw=0 time=0 us)
          0       INDEX RANGE SCAN IDX_TABLEAAA_AA_ID (cr=0 pr=0 pw=0 time=0 us)(object id 23316)
    UPDATE TABLEAS SET SUPERCEDES = 0
    WHERE
    ID=:B1
    call     count       cpu    elapsed       disk      query    current        rows
    Parse        1      0.00       0.00          0          0          0           0
    Execute      1      0.00       0.00          0          3          0           0
    Fetch        0      0.00       0.00          0          0          0           0
    total        2      0.00       0.00          0          3          0           0
    Misses in library cache during parse: 0
    Optimizer mode: ALL_ROWS
    Parsing user id: 36     (recursive depth: 2)
    Rows     Row Source Operation
          0  UPDATE  TABLEAS (cr=3 pr=0 pw=0 time=33 us)
          0   INDEX UNIQUE SCAN PK_TABLEA_ID (cr=3 pr=0 pw=0 time=30 us)(object id 23307)
    UPDATE TABLEAS SET SUPERCEDEDBY = 0
    WHERE
    ID IN (SELECT ID FROM TABLEAAA WHERE AAID IN(SELECT AAID FROM
      TABLEAAA WHERE AAID=:B2 ) ) AND SUPERCEDEDBY=:B1
    call     count       cpu    elapsed       disk      query    current        rows
    Parse        1      0.00       0.00          0          0          0           0
    Execute      1      0.26       0.29          0     121996          0           0
    Fetch        0      0.00       0.00          0          0          0           0
    total        2      0.26       0.29          0     121996          0           0
    Misses in library cache during parse: 0
    Optimizer mode: ALL_ROWS
    Parsing user id: 36     (recursive depth: 2)
    Rows     Row Source Operation
          0  UPDATE  TABLEAS (cr=121996 pr=0 pw=0 time=290701 us)
          0   HASH JOIN SEMI (cr=121996 pr=0 pw=0 time=290697 us)
          0    TABLE ACCESS FULL TABLEAS (cr=121996 pr=0 pw=0 time=290563 us)
          0    VIEW  VW_NSO_1 (cr=0 pr=0 pw=0 time=0 us)
          0     MERGE JOIN CARTESIAN (cr=0 pr=0 pw=0 time=0 us)
          0      TABLE ACCESS FULL TABLEAAA (cr=0 pr=0 pw=0 time=0 us)
          0      BUFFER SORT (cr=0 pr=0 pw=0 time=0 us)
          0       INDEX RANGE SCAN IDX_TABLEAAA_AA_ID (cr=0 pr=0 pw=0 time=0 us)(object id 23316)
    UPDATE TABLEAS SET SUPERCEDEDBY = 0
    WHERE
    ID=:B1
    call     count       cpu    elapsed       disk      query    current        rows
    Parse        1      0.00       0.00          0          0          0           0
    Execute      1      0.00       0.00          0          3          0           0
    Fetch        0      0.00       0.00          0          0          0           0
    total        2      0.00       0.00          0          3          0           0
    Misses in library cache during parse: 0
    Optimizer mode: ALL_ROWS
    Parsing user id: 36     (recursive depth: 2)
    Rows     Row Source Operation
          0  UPDATE  TABLEAS (cr=3 pr=0 pw=0 time=42 us)
          0   INDEX UNIQUE SCAN PK_TABLEA_ID (cr=3 pr=0 pw=0 time=40 us)(object id 23307)
    SELECT CLASSIFICATION
    FROM
    TABLEAS WHERE ID=:B1
    call     count       cpu    elapsed       disk      query    current        rows
    Parse        0      0.00       0.00          0          0          0           0
    Execute      1      0.00       0.00          0          0          0           0
    Fetch        1      0.00       0.00          0          3          0           0
    total        2      0.00       0.00          0          3          0           0
    Misses in library cache during parse: 0
    Optimizer mode: ALL_ROWS
    Parsing user id: 36     (recursive depth: 2)
    rollback
    call     count       cpu    elapsed       disk      query    current        rows
    Parse        1      0.00       0.00          0          0          0           0
    Execute      1      0.00       0.00          0          2         31           0
    Fetch        0      0.00       0.00          0          0          0           0
    total        2      0.00       0.00          0          2         31           0
    Misses in library cache during parse: 0
    Parsing user id: 36 
    Elapsed times include waiting on following events:
      Event waited on                             Times   Max. Wait  Total Waited
      ----------------------------------------   Waited  ----------  ------------
      log file sync                                   1        0.00          0.00
      SQL*Net message to client                       1        0.00          0.00
      SQL*Net message from client                     1       14.18         14.18
    alter session set events '10046 trace name context off'
    call     count       cpu    elapsed       disk      query    current        rows
    Parse        1      0.00       0.00          0          0          0           0
    Execute      1      0.00       0.00          0          0          0           0
    Fetch        0      0.00       0.00          0          0          0           0
    total        2      0.00       0.00          0          0          0           0
    Misses in library cache during parse: 0
    Parsing user id: 36 
    OVERALL TOTALS FOR ALL NON-RECURSIVE STATEMENTS
    call     count       cpu    elapsed       disk      query    current        rows
    Parse        3      0.00       0.00          0          0          0           0
    Execute      3      0.00       0.01          0          5         44           1
    Fetch        0      0.00       0.00          0          0          0           0
    total        6      0.00       0.01          0          5         44           1
    Misses in library cache during parse: 1
    Misses in library cache during execute: 1
    Elapsed times include waiting on following events:
      Event waited on                             Times   Max. Wait  Total Waited
      ----------------------------------------   Waited  ----------  ------------
      SQL*Net message to client                       3        0.00          0.00
      SQL*Net message from client                     3       14.18         24.44
      log file sync                                   1        0.00          0.00
    OVERALL TOTALS FOR ALL RECURSIVE STATEMENTS
    call     count       cpu    elapsed       disk      query    current        rows
    Parse       12      0.00       0.00          0          0          0           0
    Execute     13      0.79       0.82          0     344825         28           5
    Fetch        6      0.01       0.00          0         77          0           3
    total       31      0.81       0.82          0     344902         28           8
    Misses in library cache during parse: 0
       16  user  SQL statements in session.
        0  internal SQL statements in session.
       16  SQL statements in session.
        0  statements EXPLAINed in this session.
    Trace file: pacs_ora_5800.trc
    Trace file compatibility: 10.01.00
    Sort options: default
           1  session in tracefile.
          16  user  SQL statements in trace file.
           0  internal SQL statements in trace file.
          16  SQL statements in trace file.
          16  unique SQL statements in trace file.
        1066  lines in trace file.
          18  elapsed seconds in trace file.Edited by: OracleExplorer on Jan 21, 2011 8:40 PM

  • Having response time issues using Studio to manage 3000+ forms

    We are currently using Documaker Studio to create and maintain our forms, of which we have thousands. Once we create the form we export it to a very old version of Documerge where it is then used in our policy production. 
    The problem is that because we have so many forms/sections, everytime we click on "SECTIONS" in Studio it takes a significant amount of time to load the screen that lists all of the sections. Many of these forms/sections are old and will never change but we want to still have access to them in the future.
    What is the best way to "backup" all these forms somewhere where they are still accessible? Ideally I think I would like to have one workspace (let's call it "PRODUCTION") that has all 3000+ forms and delete the older resources from our existing workspace (called "FORMS") so that just has the forms that we are currently working on.  This way the response time in the "FORMS" workspace would be much better. Couple questions:
    1. How would I copy my existing workspace "FORMS" (and all the resources in it) to a new workspace called "PRODUCTION"?
    2. How would I delete from the "FORMS" workspace all of the older resources?
    3. Once I am satisfied with a new form/section in my "FORMS" workspace how would I move it to "PRODUCTION"?
    4. How could I move a form/section from "PRODUCTION" back into "FORMS" in order to make corrections, or use it as a base for a new form down the road?
    5. Most importantly....Is there a better way to do this?
    Again, we are only using this workspace for forms creation and not using it to generate output...we will be doing that in the future once we upgrade from the very old Documerge on the mainframe, to Documaker Studio.
    Many thanks to any of you who can help me with this!

    However, I am a little confused on the difference between extracting and promoting. Am I correct in assuming that I would go into my PROD workspace and EXTRACT the resources that I want to continue to work on. I would then go into my new, and empty, DEV workspace and IMPORT FILES (or IMPORT LIBRARY?) using the file(s) that I created with the EXTRACT? In effect, I would have two totally separate workspaces, one called DEV and one called PROD?
    Extraction is writing a copy of a resource from the library out to disk. Promotion is copying a resource from one library to another, with the option of modifying the metadata values of the source and target resources. You would use extract in a case where you don't have access to both libraries to do a promote.
    An example promotion scenario would go something like this. You have resources in the source (DEV) that you want to promote to the target (PROD). Items to be promoted are tagged with the MODE = "To Promote". When you perform the promotion, you can select the items that you want to promote with the filter MODE="To Promote". When you perform the promotion, you can also configure Studio to set the MODE of the resource(s) in the source to be MODE="To Delete", and set the MODE of the resource(s) in the target to be MODE="" (empty). Then you can go back and delete the resources from the source (DEV) where MODE=DELETE.
    Once you have the libraries configured you could bypass the whole extract/import bit and just use promote. The source would be PROD, and the target would be DEV. During promotion, set the target MODE = "To Do", and source MODE = "In Development". In this fashion you will see which resources in PROD are currently being edited in DEV (because in PROD the MODE = "In Development"). When development is completed, change the MODE in DEV to "To Promote", then proceed with the promotion scenario described above.
    I am a bit confused on the PROMOTE function and the libraries that have the  _DEV _TEST _PROD suffixes. This looks like it duplicates the entire workspace to new libraries _PROD but it is all part of the same workspace, not two separate workspaces?  Any clarification here would be helpful.
    Those suffixes are just attached by default; these suffixes don't mean anything to Documaker. You could name your library PROD and use it for DEV. It might be confusing though ;-) The usual best practice is to name the library and subsequent tablespaces/schemas according to their use. It's possible to have multiple libraries within a single tablespace or schema (but not recommended to mix PROD and non-PROD libraries).
    Getting there, I think!
    -A

Maybe you are looking for

  • Airport Express will connect wired, but can't get it to connect wireless

    I purchased three APE units specifically to connect my iTunes library to multiple house receivers.  I went through the setup for two of them, but as soon as I unplug the ethernet cable, the wireless network won't recognize them.  I tried rebooting th

  • Header and item levels in ALV_GRID

    I want to display customers info (like adress,phone number..) in one row  and its documents in other rows namely i want to display information hierarchicaly in ALV_GRID. I know it is possible to do with REUSE_ALV_HIERSEQ_LIST_DISPLAY but i want to do

  • Photo Gallery Software for Mac OS X Server

    I'm porting my Linux server infrastructure over to Mac OS X Server. I've successfully moved over my DNS, Mail and most of my web sites to Mac without a problem. However, the last website is causing me a bit of pain. In Linux I was using Gallery 2 (ht

  • Contacts fail to sync

    The iPhone help manual states: If you've set up iTunes to sync contacts automatically, iTunes keeps your contacts up-to-date-whether you make changes on your computer or on iPhone. This is not working. I have a contact (in sync) in both my PC (Vista)

  • Code will not write to text

    Hello to all: Can someone give advice to why the code below will not write to the text that I save. In other words, when I write a text file then save it then open the text file again, all of the words that I have written are not seen.      private F