Performance Issue caused by XmlResults..next().asString();

Hi all,
We there is a simple method to get related documents from Berkeley DB through XQuery, and using XmlResults..next().asString() to get the result. (Note: the result is quite large.)
If call this method in loop, it will become slower and slover, just because of using XmlResults..next().asString(), if only get result XmlResults..next(), not using asString(), there is no performance issue.
Can any one have good suggestions to reslove it?
Many thank.
Jane

Hi,
THis question should be posted on the Berkeley DB XML forum.
Berkeley DB XML
Ron

Similar Messages

  • Oracle CPU Jan 2009 cause performance issue

    I did installed Oracle CPU Jan 2009 on HP-UX machine. But once the installation is completed, users complaint saying that it takes more than 1 minutes to open a new ticket on that application. With this, it cause a backlog processess increase tremendously.
    Will the CPU Jan 2009 patch cause any of the network performance issue:
    Server: HP-UX Itanium 64 bit
    Database: Oracle 10.2.0.3.0
    Instances: 2 instances running on this server.
    Edited by: user3858134 on Oct 26, 2009 9:30 PM

    I believe the latest CPU Patch for Oracle 10.2.0.3 on HP is CPU Jan 2009 only.Don't you think your database should be on 10.2.0.4?
    Anyways, do you have any benchmark statspack/AWR report? Can you compare that with the latest one? Do you see any difference?
    Regards,
    S.K.

  • SetAttribute causing performance issue.

    Hi ,
    I am using 11.1.1.4.0
    Code:::
            DCIteratorBinding itr=ADFUtil.findIterator(iterator);
            RowSetIterator rsi=itr.getRowSetIterator();
            Row currRow=rsi.getCurrentRow();
            currRow.setAttribute(id,null);
    If i call setAttribute multiple times(like 10-20 times) ,it causes severe performance issue .
    Is there any reason for it ??
    Should we avoid using setAttribute() ??If so then what we should use?
    Any help is appreciated .
    Thanks
    Sazz

    usecase is user wld see a existing vacancy record and able to update it.
    GEVacancyFromNotificationVO1()  is a query based vo and  getGETranVacancyVO1() is a updatable VO . Now using view criretia i am pulling the record in the updatable VO , this will have only 1 record at 1 time.
    GEVacancyFromNotificationVO1() gets the details and set then in attributes of the updatable VO as this VO includes many trasient attributes which are required in my jsff . Basicallly this data are not saved in DB but required to show in the UI.
    Anyways now the thing is setAttribute as called 20-30 times you see , the performance is slow and sometimes data is not set as well.
    I used attributeListImp class to create a name value pair and create a new row for this VO using createAndInitRow() and that works super fast . That is requied for another use case and works perfectly ok . Only when i want to update a existing record i have to update the same row. cant create another row, so facing this performance issue and sometime data doesnt set properly . i get null in DCiterator binding when i fetch the data in bean class.
    So my question is why does setAttribute of AttributeListImpl is much much faster than setAttribute of Row class.??
    public void initializeFromNotification(String role, String emp) {
            ViewObjectImpl notifyVO = this.getGEVacancyFromNotificationVO1();
            ViewObjectImpl transVO = this.getGETranVacancyVO1();
            ViewObjectImpl geLoginPersonIdVO = this.getGELoginPersonIdVO1();
            ViewObjectImpl autoPopulatevo =
                this.getGEAutopopulateHireSysforCopyVacanciesVO1();
            ViewObjectImpl geNextApproverVO = this.getGENextApproverVO1();
            ViewObjectImpl transHireVo = getGEHireSystemReqTeamTransVO1();
            ViewObjectImpl gejobdesc = getGEJobDescTransVO1();
            Row row = notifyVO.first();
            if (row != null) {
                //query the trx table
                transVO.setApplyViewCriteriaName("VacancyNumberVC");
                transVO.setNamedWhereClauseParam("p_vac_num",
                                                 row.getAttribute("VacancyNumber"));
                transVO.executeQuery();
                if (transVO.first() == null) {
                    return;
                } else {
                    transVO.setCurrentRow(transVO.first());
                Row currentRow = transVO.getCurrentRow();
                List<String> transColumns =
                    Arrays.asList(currentRow.getAttributeNames());
                //setting values from notification vo to transvacancy VO
                String arr[] = row.getAttributeNames();
                if (null != transVO.getCurrentRow()) {
                    // AttributeListImpl attrList = new AttributeListImpl();
                    for (String attr : arr) {
                        if (row.getAttribute(attr) != null) {
                            if (attr.equalsIgnoreCase("VacTrxId")) {
                            } else if (transColumns.contains(attr)) {
                                if (currentRow.getAttribute(attr) == null) {
                                    currentRow.setAttribute(attr,
                                                            row.getAttribute(attr).toString());
                if (role != null && role.startsWith("ORG_MGR")) {
                    transVO.getCurrentRow().setAttribute("userRole",
                                                         "INITIATOR_HM");
                    transVO.getCurrentRow().setAttribute("userRoleDisplay",
                                                         "Hiring Manager");
                } else if (role != null && role.startsWith("HRM")) {
                    transVO.getCurrentRow().setAttribute("userRole",
                                                         "INITIATOR_HRM");
                    transVO.getCurrentRow().setAttribute("userRoleDisplay",
                                                         "HR Manager");
                } else {
                    transVO.getCurrentRow().setAttribute("userRole",
                                                         "INITIATOR_RFO");
                    transVO.getCurrentRow().setAttribute("userRoleDisplay", "RFO");
                transVO.getCurrentRow().setAttribute("EmpNumber", emp);
                geLoginPersonIdVO.setNamedWhereClauseParam("sso", emp);
                geLoginPersonIdVO.executeQuery();
                transVO.getCurrentRow().setAttribute("userPersonId",
                                                     geLoginPersonIdVO.first().getAttribute(0));

  • Load balancing issue causes performance problems!

    Hi
    Can someone tell me if there is a way to find the list of users who are logging into the system directly (without using a logon group)? They are hitting the CI which is causing performance issues. Eventhough we've setup system messages and repeated emails its hard to track who aren't following it.
    Any ideas would be really appretiated rewarded with points.
    Thanks
    Arun

    easy:
    on the CI, change rdisp/gui_auto_logout to 60.
    if they login and there is no activity for 60 seconds, they will be logged out promptly. they will stop trying to login into the CI after awhile.
    and if they keep on trying, lock their account into SU01... they will get the message.
    another possibility is a user exit on the logon screen. check the ABAP forum for this.

  • Performance issue is most probably caused by the fact that IE can not handle such amount of requests ? is it ?

    Hi,
    I'm running web applications and there is performance issue is most probably caused by the fact that IE can not handle such amount of requests that are forwarded to the server at the same time.
    Data:
    On IE(10) – search variety 11- it takes a long time (about 25 seconds for response) .
    in general, the responding time of "search variety" in IE is much longer than Chrome.
    IE - see:http://screencast.com/t/kSeT3hC2mRfV
    Chrome see:http://screencast.com/t/NYf6fskU
    Why? any solution ?
    Thanks,

    Hi,
    without a link to your site (so we can use the same tools mentioned above)... we can only guess.
    these are peer to peer support forums... your favorite web search engine will help you to learn about web development.
    display the Developer tool console in each browser to view the suppressed error messages.
    All modern browsers support the performance api or your can use the Networking tab to view the latency metrics for the site.... commonly you may be using meta directives for caching instead of server headers or you have different settings between browsers
    which determine how long resources are cached. Commonly, IE has a different security model to other browsers... ensure that you are using the Default IE security zone settings.... Actions like innerHTML go through a sanitization process to remove script content...
    Commonly AJAX calls are made during the onload event...
    Please post questions about web site development to the IE Web Developer forum. Include with your questions a link to your website or a publicly accessible mashup that shows your issue.
    Questions regarding Internet Explorer 8, 9 and 10 and Internet Explorer 11 for the IT Pro Audience. Topics covered are: Installation, Deployment, Configuration, Security, Group Policy, Management questions. If you are a consumer looking for answers or to
    raise a question, it's highly recommended you head on over to http://answers.microsoft.com
    Rob^_^

  • BPC 7.5 AfterRange Causes Extreme Performance Issue

    In a report when I apply second AfterRange the run time goes from under 2 minutes to 15 minutes. I know there are performance penalties for using Before/AfterRange but I have never seen the discrepancy to this proportion.
    Report contains 1 Column (Time) and 4 Rows in Expansion. The first range which causes no issues for Cost Centers is defined Q73:V74, the range which causes the performance issue and applies for each Employee is defined as Q71:V72
    To test this I have two versions of essentially the same report, one with only Cost Center AfterRange and this version runs in less than 2 minutes, then I have the other version in which I have added one more AfterRange to another column and that's when the hell breaks loose and this version of the report runs in 15 minutes. The number of records is roughly 2000, and I realize that the range causing the issue is applied on pretty much every rcord in result set but still I'd not expect the run time to increase by this much.
    BPC Experts, ideas on what could be causing this issue, any help would be highly appreciated. thanks!

    Hi
    1> Please try to restart your Gernal Server
    2> Check after optimization that helps
    Regards
    Vinoo

  • WEBUTIL - Does adding it to all forms cause performance issues?

    If I add the webutil library and object library to all forms in the system (as part of a standard template) despite the fact most won't use it, will this cause any performance issues???
    Thanks in advance...

    The webutil user guide has a chapter on performance considerations. Have you looked at that?
    The number one point from that chapter is:
    1. Only WebUtil Enable Forms that actually need the functionality. Each form that is WebUtil enabled will generate a certain amount of network traffic and memory
    usage simply to instantiate the utility, even if you don’t use any WebUtil
    functionality.

  • WILL BIG INDEX WILL CAUSE PERFORMANCE ISSUE?

    In an index table, if there are a lot of insert then data will grow and/or if the index is
    huge then can it really cause performance issue?
    Is there a document in metalink that says if index is 50% of data then we have to rebuild it? What are the basis and threshold of rebuilding index?

    A big index by itself won't cause a performance issue. There are other circumstances you should consider for the index.
    First of all, which kind of index are you talking about, there are several kind of indexes in Oracle. On the other hand, assuming you are talking about a regular B*Tree index, you should consider factors such as selectivity and cardinality. If the indexed column has evenly distributed values, then the index will be highly selective, and if the indexed column is highly skewed, in order for the index not to become a real bottleneck you should gather histograms, so selectivity can be calculated at execution time and in case the query retrieves a highly selective data range the index won't slow performance, otherwise a full table scan will be considered a best data access path.
    Rebuilding indexes is an operation performed when the index becomes invalid, or when migrating the index to a new tablespace, but not when you suspect the index has become 'fragmented' in this case you should use the Coalesce command. Oracle provides efficient algorithms to maintain the index balanced.
    ~ Madrid
    http://hrivera99.blogspot.com/

  • Interested by performance issue ?  Read this !  If you can explain, you're a master Jedi !

    This is the question we will try to answer...
    What si the bottle neck (hardware) of Adobe Premiere Pro CS6
    I used PPBM5 as a benchmark testing template.
    All the data and log as been collected using performance counter
    First of all, describe my computer...
    Operating System
    Microsoft Windows 8 Pro 64-bit
    CPU
    Intel Xeon E5 2687W @ 3.10GHz
    Sandy Bridge-EP/EX 32nm Technology
    RAM
    Corsair Dominator Platinum 64.0 GB DDR3
    Motherboard
    EVGA Corporation Classified SR-X
    Graphics
    PNY Nvidia Quadro 6000
    EVGA Nvidia GTX 680   // Yes, I created bench stats for both card
    Hard Drives
    16.0GB Romex RAMDISK (RAID)
    556GB LSI MegaRAID 9260-8i SATA3 6GB/s 5 disks with Fastpath Chip Installed (RAID 0)
    I have other RAID installed, but not relevant for the present post...
    PSU
    Cosair 1000 Watts
    After many days of tests, I wanna share my results with community and comment them.
    CPU Introduction
    I tested my cpu and pushed it at maximum speed to understand where is the limit, can I reach this limit and I've logged precisely all result in graph (See pictures 1).
    Intro : I tested my E5-XEON 2687W (8 Cores Hyperthread - 16 threads) to know if programs can use the maximum of it.  I used Prime 95 to get the result.  // I know this seem to be ordinary, but you will understand soon...
    The result : Yes, I can get 100% of my CPU with 1 program using 20 threads in parallel.  The CPU gives everything it can !
    Comment : I put 3 IO (cpu, disk, ram) on the graph of my computer during the test...
    (picture 1)
    Disk Introduction
    I tested my disk and pushed it at maximum speed to understand where is the limit and I've logged precisely all result in graph (See pictures 2).
    Intro : I tested my RAID 0 556GB (LSI MegaRAID 9260-8i SATA3 6GB/s 5 disks with Fastpath Chip Installed) to know if I can reach the maximum % disk usage (0% idle Time)
    The result : As you can see in picture 2, yes, I can get the max of my drive at ~ 1.2 Gb/sec read/write steady !
    Comment : I put 3 IO (cpu, disk, ram) on the graph of my computer during the test to see the impact of transfering many Go of data during ~10 sec...
    (picture 2)
    Now, I know my limits !  It's time to enter deeper in the subject !
    PPBM5 (H.264) Result
    I rendered the sequence (H.264) using Adobe Media Encoder.
    The result :
    My CPU is not used at 100%, the turn around 50%
    My Disk is totally idle !
    All the process usage are idle except process of (Adobe Media Encoder)
    The transfert rate seem to be a wave (up and down).  Probably caused by (Encrypt time....  write.... Encrypt time.... write...)  // It's ok, ~5Mb/sec during transfert rate !
    CPU Power management give 100% of clock to CPU during the encoding process (it's ok, the clock is stable during process).
    RAM, more than enough !  39 Go RAM free after the test !  // Excellent
    ~65 thread opened by Adobe Media Encoder (Good, thread is the sign that program try to using many cores !)
    GPU Load on card seem to be a wave also ! (up and down)  ~40% usage of GPU during the process of encoding.
    GPU Ram get 1.2Go of RAM (But with GTX 680, no problem and Quadro 6000 with 6 GB RAM, no problem !)
    Comment/Question : CPU is free (50%), disks are free (99%), GPU is free (60%), RAM is free (62%), my computer is not pushed at limit during the encoding process.  Why ????  Is there some time delay in the encoding process ?
    Other : Quadro 6000 & GTX 680 gives the same result !
    (picture 3)
    PPBM5 (Disk Test) Result (RAID LSI)
    I rendered the sequence (Disk Test) using Adobe Media Encoder on my RAID 0 LSI disk.
    The result :
    My CPU is not used at 100%
    My Disk wave and wave again, but far far from the limit !
    All the process usage are idle except process of (Adobe Media Encoder)
    The transfert rate wave and wave again (up and down).  Probably caused by (Buffering time....  write.... Buffering time.... write...)  // It's ok, ~375Mb/sec peak during transfert rate !  Easy !
    CPU Power management give 100% of clock to CPU during the encoding process (it's ok, the clock is stable during process).
    RAM, more than enough !  40.5 Go RAM free after the test !  // Excellent
    ~48 thread opened by Adobe Media Encoder (Good, thread is the sign that program try to using many cores !)
    GPU Load on card = 0 (This kind of encoding is GPU irrelevant)
    GPU Ram get 400Mb of RAM (No usage for encoding)
    Comment/Question : CPU is free (65%), disks are free (60%), GPU is free (100%), RAM is free (63%), my computer is not pushed at limit during the encoding process.  Why ????  Is there some time delay in the encoding process ?
    (picture 4)
    PPBM5 (Disk Test) Result (Direct in RAMDrive)
    I rendered the same sequence (Disk Test) using Adobe Media Encoder directly in my RamDrive
    Comment/Question : Look at the transfert rate under (picture 5).  It's exactly the same speed than with my RAID 0 LSI controller.  Impossible !  Look in the same picture the transfert rate I can reach with the ramdrive (> 3.0 Gb/sec steady) and I don't go under 30% of disk usage.  CPU is idle (70%), Disk is idle (100%), GPU is idle (100%) and RAM is free (63%).  // This kind of results let me REALLY confused.  It's smell bug and big problem with hardware and IO usage in CS6 !
    (picture 5)
    PPBM5 (MPEG-DVD) Result
    I rendered the sequence (MPEG-DVD) using Adobe Media Encoder.
    The result :
    My CPU is not used at 100%
    My Disk is totally idle !
    All the process usage are idle except process of (Adobe Media Encoder)
    The transfert rate wave and wave again (up and down).  Probably caused by (Encoding time....  write.... Encoding time.... write...)  // It's ok, ~2Mb/sec during transfert rate !  Real Joke !
    CPU Power management give 100% of clock to CPU during the encoding process (it's ok, the clock is stable during process).
    RAM, more than enough !  40 Go RAM free after the test !  // Excellent
    ~80 thread opened by Adobe Media Encoder (Lot of thread, but it's ok in multi-thread apps!)
    GPU Load on card = 100 (This use the maximum of my GPU)
    GPU Ram get 1Gb of RAM
    Comment/Question : CPU is free (70%), disks are free (98%), GPU is loaded (MAX), RAM is free (63%), my computer is pushed at limit during the encoding process for GPU only.  Now, for this kind of encoding, the speed limit is affected by the slower IO (Video Card GPU)
    Other : Quadro 6000 is slower than GTX 680 for this kind of encoding (~20 s slower than GTX).
    (picture 6)
    Encoding single clip FULL HD AVCHD to H.264 Result (Premiere Pro CS6)
    You can look the result in the picture.
    Comment/Question : CPU is free (55%), disks are free (99%), GPU is free (90%), RAM is free (65%), my computer is not pushed at limit during the encoding process.  Why ????   Adobe Premiere seem to have some bug with thread management.  My hardware is idle !  I understand AVCHD can be very difficult to decode, but where is the waste ?  My computer want, but the software not !
    (picture 7)
    Render composition using 3D Raytracer in After Effects CS6
    You can look the result in the picture.
    Comment : GPU seems to be the bottle neck when using After Effects.  CPU is free (99%), Disks are free (98%), Memory is free (60%) and it depend of the setting and type of project.
    Other : Quadro 6000 & GTX 680 gives the same result in time for rendering the composition.
    (picture 8)
    Conclusion
    There is nothing you can do (I thing) with CS6 to get better performance actually.  GTX 680 is the best (Consumer grade card) and the Quadro 6000 is the best (Profressional card).  Both of card give really similar result (I will probably return my GTX 680 since I not really get any better performance).  I not used Tesla card with my Quadro, but actually, both, Premiere Pro & After Effects doesn't use multi GPU.  I tried to used both card together (GTX & Quadro), but After Effects gives priority to the slower card (In this case, the GTX 680)
    Premiere Pro, I'm speechless !  Premiere Pro is not able to get max performance of my computer.  Not just 10% or 20%, but average 60%.  I'm a programmor, multi-threadling apps are difficult to manage and I can understand Adobe's programmor.  But actually, if anybody have comment about this post, tricks or any kind of solution, you can comment this post.  It's seem to be a bug...
    Thank you.

    Patrick,
    I can't explain everything, but let me give you some background as I understand it.
    The first issue is that CS6 has a far less efficient internal buffering or caching system than CS5/5.5. That is why the MPEG encoding in CS6 is roughly 2-3 times slower than the same test with CS5. There is some 'under-the-hood' processing going on that causes this significant performance loss.
    The second issue is that AME does not handle regular memory and inter-process memory very well. I have described this here: Latest News
    As to your test results, there are some other noteworthy things to mention. 3D Ray tracing in AE is not very good in using all CUDA cores. In fact it is lousy, it only uses very few cores and the threading is pretty bad and does not use the video card's capabilities effectively. Whether that is a driver issue with nVidia or an Adobe issue, I don't know, but whichever way you turn it, the end result is disappointing.
    The overhead AME carries in our tests is something we are looking into and the next test will only use direct export and no longer the AME queue, to avoid some of the problems you saw. That entails other problems for us, since we lose the capability to check encoding logs, but a solution is in the works.
    You see very low GPU usage during the H.264 test, since there are only very few accelerated parts in the timeline, in contrast to the MPEG2-DVD test, where there is rescaling going on and that is CUDA accelerated. The disk I/O test suffers from the problems mentioned above and is the reason that my own Disk I/O results are only 33 seconds with the current test, but when I extend the duration of that timeline to 3 hours, the direct export method gives me 22 seconds, although the amount of data to be written, 37,092 MB has increased threefold. An effective write speed of 1,686 MB/s.
    There are a number of performance issues with CS6 that Adobe is aware of, but whether they can be solved and in what time, I haven't the faintest idea.
    Just my $ 0.02

  • LR 4.1 RC performance issues

    I find that the 4.1 rc version is better than 4.0 but still not snappy like 3.6 was.  Here are some things I have noticed that are slow:
    • tagging images with flags, colors, labels, etc. - seems to be a two to three second delay when doing this before they register and show up with the files.  I have write .xmp off.
    • and in regards to writing to xmp off, if I select it,  It seems to want to update my whole catalog (148,000 images) and if I try to quit, it promps me that it has not finished and do I want to quit and if I do it will resume upon restarting.  At that point when I try to quit, it just hangs even if I tell it to quit writing xmp info.  I let it go for a couple hours yesterday and it seems it still didn't finish and caused problems again trying to quit.  I ended up having to force quit.  I never remember this problem in the past and in the past, if you selected write to xmp or not, it used to do it or not do it from that point forward.  Until this is fixed, I have it off and just manually save metadata using Cmd-S.
    •When in library mode, moving between images is snappy, however once you go to the develop module, it slows to a snail's pace (3-5 second delay).  I know develop module is not as fast in general for moving between images, but in ver. 3.6 there was only a slight delay (maybe 1 sec.).  Now I'm bouncing back and forth wasting time switching modules, which brings me to my next performance issue:
    * When switching from library to develop, it takes quite a few seconds to respond, then after that it loads the image with the loading... icon/message, and another few seconds are wasted.  That used to be almost instantaneous in 3.6.  Going to develop to library seems fine with only a slight delay.
    • The adjustment brush is faster in 4.1 than 4.0 but not nearly as responsive as in 3.6 which is frustrating and very easy to overdo what you're trying to do by the time the brush catches up.
    •Sliders still have a bit of jerkiness to them in develop module.  Also sometimes when I grab a scrubby slider to move it, it highlights the field as if I want to enter a number.  Wish this would not happen.  If I want to enter the number, a single click should do it in the field.  If I grab the scrubby slider, it should just activate the scrubby slider.
    • Keyword suggestions take much longer to generate.  In 3.6 as fast as I'd add keywords, the relevant suggestions would pop up right away.
    • When zooming into 100%, it seems to be reasonably fast to render, but I've noticed sometimes it just shows the blurred semi-rendered version without actually rendering 1:1.  Even unzooming and rezooming doesn't work.  Only solution is to go to another image and then back to the original image and that seems to kick it into gear and re-initialize the rendering engine.
    Overall I love the features and new process 2012, I'm just disappointed in the performance.  Especially now that I'm trying to edit a large job.  I really wish the program could be optimized to work as well as 3.6 did and I hope once the final version of 4.1 is released (hopefully soon) we will see an overall performance boost.  I have removed ver 3.6 from my system so that's too much of a hassle to go back and not an option, plus I've gotten used to the new process and like it too much to go back.
    From the posts I've read it seems others are experiencing this as well, as well as colleagues I've spoken to.  It seems to be across platforms and regardless of computer specs (as long as someone's using a fairly updated and capable computer)
    Thanks for working on a performance boost Adobe and getting ver. 4.1 dialed in and releasing a final version.

    Cannot reproduce on Win7 64Bit (LR4.1RC).
    What OS
    what external editor
    what image format is the original
    what image format is passed to the external editor
    Beat

  • DB Performance issue

    Hi DB Gurus,
    Our application is inserting 60-70K records in a table in each transaction. When multiple sessions are open on this table user face performance issues like application response is too slow.
    Regarding this table:
    1.Size = 56424 Mbytes!
    2.Count = 188,858,094 rows!
    3.Years of data stored = 4 years
    4.Average growth = 10 million records per month, 120 million each year! (has grown 60 million since end of June 2007)
    5.Storage params = 110 extents, Initial=40960, Next=524288000, Min Extents=1, Max Extents=505
    6.There are 14 indexes on this table all of which are in use.
    7. Data is inserted through bulk insert
    8. DB: Oracle 10g
    Sheer size of this table (56G) and its rate of growth may be the culprits behind performance issue. But to ascertain that, we need to dig out more facts so that we can decide conclusively how to mail this issue.
    So my questions are:
    1. What other facts can be collected to find out the root cause of bad performance?
    2. Looking at given statistics, is there a way to resolve the performance issue - by using table partition or archiving or some other better way is there?
    We've already though of dropping some indexes but it looks difficult since they are used in reports based on this table (along with other tables)
    3. Any guess what else can be causing this issue?
    4. How many records per session can be inserted in a table? Is there any limitation?
    Thanks in advance!!

    Run STATSPACK and check what it says are the issues. Try and find the particular INSERT statement in the list of all SQL. Look at all the sections of the report, including block contention, which may show you are waiting for data blocks or index blocks, etc, or even things like latch contention too. Make sure you run it when the INSERT is happening during one of your busy periods.
    Given that you are using Oracle 10g, I assume you are using all the automatic settings now:
    o Local Tablespace Management
    o Automatic Segment Space Management
    o Automatic Undo Management
    If not, you should be. Prior to all this, Oracle always inserted into the last block in a table, which could become a bottleneck point. And space allocation of new blocks was also a problem. When these settings were introduced it alleviated most of these problems, and meant that Oracle could scale far better on such INSERT intensive workloads. If you are not using these for some reason or other, then you need to look at the number of FREELISTS you have on the table, and the setting of INITRANS.
    Also, how many columns does this table have? And how big is an average row. And what is your block size? You can get these from the data dictionary:
    select count (*) from user_tab_columns where table_name = '<tablename>' ;
    select avg_row_len from user_tables where table_name = '<tablename>' ;
    show parameter db_block_size
    Replace <tablename> with the name of your table, in uppercase.
    I ask because a very large row in a small data block will always fill the block quickly and cause new blocks to be allocated. If so, you may just have to live with this.
    And I would be suspicious about all 14 indexes being needed. Are they all single column indexes, or do you have any multi-column indexes? Do any of them share the same leading columns? Again, if you need all 14 indexes, then you must suffer the overhead of maintaining these indexes. But unless you have something like 50 columns in this table, I would guess that there is some overlap between these indexes.
    John

  • Performance issue entering monitor etc

    Dear Sirs,
    Our system currently feels quite slow. When entering the monitor for a load, or managing a cube the system it takes a while.
    In SM50 I can see it reads at table  (sequential) på RSMONIPTAB (which contains 1 millions records).
    SAP help.sap.com describes a process u201CArchivation of Request Administration Datau201D, which might help? Look below. However, I do not like the paragraph marked in bold. What kind of features do they reference to.
    http://help.sap.com/saphelp_nw04s/helpdata/en/42/ead8d7b55e1bd2e10000000a11466f/frameset.htm  
    Archiving request administration data allows you to store the administration and log information for a request in an archive. In BI, tables store the totals records that the system requires so that the request is not set to red because information is missing. This function allows you to limit the size of the administration and log tables, RSDONE und RSMON. This improves performance for actions performed for the request and in the affected InfoProviders and saves memory space in the database without affecting the status set by the system.
    For certain functions and processes that access requests, you have to reload request administration data to the BI administration and log tables.
    Any hints on the issue in general; performance issue with monitor, please let me know.
    Best regards,
    Jørgen

    Guess note below describes gives more info about situation. However, not with a clear solution :=)
    SAP Note 694895 - Performance and tables RSMON, RSDONE
    Note Language: English Version: 5 Validity: Valid Since 29.11.2005
    Summary
    Symptom
    Tables RSMON* (for example, table RSMONMESS) and tables RS*DONE (for
    example, table RSSELDONE) continuously increase in size in the BW system
    with each request that is created.
    Currently there is no option to reduce these tables.
    Do not delete any entries from these tables. If you did, this would have
    the following consequences:
    When you carry out the next check for the requests for which you have
    deleted records from these tables, the check will not run properly.
    The status of the relevant requests turns RED in the monitor, and also in
    all affected data targets that contain the request.
    The affected request and all subsequent requests are no longer visible in
    the reporting - all queries on the affected data targets then only display
    old data.
    Numerous dumps will occur in various situations.
    You will not be able to repair the errors caused by the deletion.
    More Terms
    RSMONMESS; RSSELDONE; Performance; RSMON; RSDONE;
    Cause and Prerequisites
    This is caused by a program error.
    Solution
    In the next BW release (BI in SAP NetWeaver 2004s - that is BW 7.0), you
    will be able to archive entries from the RSMON* and RS*DONE tables using a
    request archiving process that archives the administrative information for
    requests.
    After archiving, these tables are then considerably reduced in size.
    Until then, there is unfortunately no option for reducing the tables.

  • Windows 7 Enterprise login performance issue RODC site

    Hi, 
    We are deploying windows 7 enterprise to all our computers in our domain and we didn't have any performance issue on any site with Full DC. Now we are facing a problem with our multiple sites with RODC but only with new deployed windows 7 (WinXp is not having
    this slow login problem). The problem is very strange. Whatever a user start a session on a desktop or laptop with Windows 7, wireless or cabled, when the user enter his credential and press the "validation arrow" next to the password nothing happen
    (well it look like nothing happen). After 5 to 10 minutes at this logon screen, we finally see the logon process start applying domain policy, preparing the desktop etc etc and this take less then 10 seconds but remember, we just wait 5 to 10 minutes until
    the computer process the "process my log in" botton. Sometime, the CTRL+ALT+DEL screen come back and when we press them we got nothing, we have to wait for the "logon" to occur before doing anything else with the computer.
    I've looked to many way to solve this but for now nothing worked.   

    Hi,
    Sorry for my dilatory reply. How about using a same user account  to log on XP computer? Will it logon fast?
    Have you tried to add a user account into Allowed RODC Password Replication Group for test? This problem probably caused by the password authentication problem In RODC host.
    For further investigation, you can try to use Network Manitor at RODC host to capture package of client machine authenticate to a domain controller.
    The blog about Understanding "Read Only Domain Controller" authentication
    may be helpful.
    After the above steps, We can narrow the cause of this problem.
    If problem persists, this problem may caused by Windows 7 Client performance. Generally speaking, the problem like this most probably caused by sercurity application problem. You can try to disable or uninstall security application temportaily for test.
    If there is any progress, please feel free let us know.
    Roger Lu
    TechNet Community Support

  • Creating datafile performance issues

    Hi guys,
    I´m using a oracle rac 10g with 3 nodes, and i needed do create a new datafile in production environment, then i had some performance issues, do you guys have any idea what the cause of this ?
    CREATE TABLESPACE "TBS_TDE_CMS_DATA"
    DATAFILE
    '+RDATA' SIZE 1000M REUSE AUTOEXTEND ON NEXT 1000M MAXSIZE 30000M
    LOGGING EXTENT MANAGEMENT LOCAL SEGMENT SPACE MANAGEMENT AUTO;
    Any hint would be helpful.
    thanks.

    BrunoSales wrote:
    Hi guys,
    I´m using a oracle rac 10g with 3 nodes, and i needed do create a new datafile in production environment, then i had some performance issues, do you guys have any idea what the cause of this ?
    CREATE TABLESPACE "TBS_TDE_CMS_DATA"
    DATAFILE
    '+RDATA' SIZE 1000M REUSE AUTOEXTEND ON NEXT 1000M MAXSIZE 30000M
    LOGGING EXTENT MANAGEMENT LOCAL SEGMENT SPACE MANAGEMENT AUTO;
    Any hint would be helpful.
    thanks.why are creating 1GB of initial extent? I think is too big. Remember whenever you add datafile to database, oracle first formats the complete datafile. So doing this can have performance issue as it depends upon the datafile size. check if you face any issue in I/O or check if there are any spikes in I/O.
    You can try adding a datafile with default values and check how much time it take

  • Performance Issue with sql query

    Hi,
    My db is 10.2.0.5 with RAC on ASM, Cluster ware version 10.2.0.5.
    With bsoa table as
    SQL> desc bsoa;
    Name                                      Null?    Type
    ID                                        NOT NULL NUMBER
    LOGIN_TIME                                         DATE
    LOGOUT_TIME                                        DATE
    SUCCESSFUL_IND                                     VARCHAR2(1)
    WORK_STATION_NAME                                  VARCHAR2(80)
    OS_USER                                            VARCHAR2(30)
    USER_NAME                                 NOT NULL VARCHAR2(30)
    FORM_ID                                            NUMBER
    AUDIT_TRAIL_NO                                     NUMBER
    CREATED_BY                                         VARCHAR2(30)
    CREATION_DATE                                      DATE
    LAST_UPDATED_BY                                    VARCHAR2(30)
    LAST_UPDATE_DATE                                   DATE
    SITE_NO                                            NUMBER
    SESSION_ID                                         NUMBER(8)
    The query
    UPDATE BSOA SET LOGOUT_TIME =SYSDATE WHERE SYS_CONTEXT('USERENV', 'SESSIONID') = SESSION_ID
    Is taking a lot of time to execute and in AWR reports also it is on top in
    1. SQL Order by elapsed time
    2. SQL order by reads
    3. SQL order by gets
    So i am trying a way to solve the performance issue as the application is slow specially during login and logout time.
    I understand that the function in the where condition cause to do FTS, but i can not think what other parts to look at.
    Also:
    SQL> SELECT COUNT(1) FROM BSOA;
      COUNT(1)
       7800373
    The explain plan for  "UPDATE BSOA SET LOGOUT_TIME =SYSDATE WHERE SYS_CONTEXT('USERENV', 'SESSIONID') = SESSION_ID" is
    {code}
    PLAN_TABLE_OUTPUT
    Plan hash value: 1184960901
    | Id  | Operation          | Name               | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | UPDATE STATEMENT   |                    |     1 |    26 | 18748   (3)| 00:03:45 |
    |   1 |  UPDATE            | BSOA |       |       |            |          |
    |*  2 |   TABLE ACCESS FULL| BSOA |     1 |    26 | 18748   (3)| 00:03:45 |
    Predicate Information (identified by operation id):
       2 - filter("SESSION_ID"=TO_NUMBER(SYS_CONTEXT('USERENV','SESSIONID')))
    {code}

    Hi,
    There are also triggers before update and AUDITS on this table.
    CREATE OR REPLACE TRIGGER B2.TRIGGER1
    BEFORE UPDATE
    ON B2.BSOA  REFERENCING OLD AS OLD NEW AS NEW
    FOR EACH ROW
    :NEW.LAST_UPDATED_BY   := USER    ;
    :NEW.LAST_UPDATE_DATE  := SYSDATE ;
    END;
    CREATE OR REPLACE TRIGGER B2.TRIGGER2
    BEFORE INSERT
    ON B2.BSOA  REFERENCING OLD AS OLD NEW AS NEW
    FOR EACH ROW
    :NEW.CREATED_BY        := USER ;
    :NEW.CREATION_DATE     := SYSDATE ;
    :NEW.LAST_UPDATED_BY   := USER    ;
    :NEW.LAST_UPDATE_DATE  := SYSDATE ;
    END;
    And also there is an audit on this table
    AUDIT UPDATE ON B2.BSOA BY ACCESS WHENEVER SUCCESSFUL;
    AUDIT UPDATE ON B2.BSOA BY ACCESS WHENEVER NOT SUCCESSFUL;
    And the sessionid column in BSOA has height balanced histogram.
    When i create an index i get the following error. As i am on 10g I can't use DDL_LOCK_TIMEOUT . I may have to wait for next down time.
    SQL> CREATE INDEX B2.BSOA_SESSID_I ON B2.BSOA(SESSION_ID) TABLESPACE B2 COMPUTE STATISTICS;
    CREATE INDEX B2.BSOA_SESSID_I ON B2.BSOA(SESSION_ID) TABLESPACE B2 COMPUTE STATISTICS
    ERROR at line 1:
    ORA-00054: resource busy and acquire with NOWAIT specified
    Thanks

Maybe you are looking for

  • IPhoto albums not showing up in screensaver

    I have just updated to iLife 09. Now when I go to system preferences - screen saver - pictures, my iPhoto albums no longer come up. I can only choose a folder called Pictures folder and it contains all of my iPhoto pictures. I cannot narrow it to one

  • HELP! newbie to itunes for mac

    hello-previously a windows user and just got a macbook. trying to delete duplicate songs instead of deleting one at a time. in windows, you would ctrl and select songs and delete. how does one do it with a mac? thanks for any help!

  • Color picker control for WD

    Hi! Is there some sort of color picker control for WD?

  • Help me - getCustomDatum() function

    I am testing sample of intermedia on Oracle 9.2.0.1.0. I cannot test the sample because getCustomDatum() function is replaced by the new function, and I do not know use it. I am using ordim817.zip packet, I think it is used for Oracle 8.1.7. Please h

  • 2600 classic - error in connection?

    When I attempt to make a phone call, I recieve a "error in connection" and stop sign icon.  I can recieve phone calls, and send messages. There is there is strange yellow number two icon on top of my screen as well, was never there before. Any help w