Oracle Database Performance With Semantic

Hello,
Is there a Developer's Guide for Semantic that specifically talks about database performance with the Semantic network/tables/indexes? We are having issues with performance the larger the semantic network becomes.
Any help or pointers would be appriciated.
Thanks
-MichaelB

Matt,
Thanks for your response. Here are the answers to the questions about our setup/environment.
1) Are you querying multiple models and/or a model + entailment? If so, are you using a virtual model and using the ALLOW_DUP=T query option?
A single model, no entailments. We attempted to use multiple models, and a virtual model (with ALLOW_DUP=T), however the UNION ALL in the explain plan made the query duration unacceptable.
2) Are you using named graphs?
No named graphs.
3) How many triples are you querying?
Approximately 85 million.
4) What semantic network and/or datatype indexes have been created?
We have PCSGM, PSCGM, PSCM, PCSM, CPSM, and SCM.
5) What is your hardware setup (number and type of disks, RAM, processor, etc.)?
We are running the 11.2.0.3 database on a Sun Solaris T2000, we have ASM managing our disks from RAID5, I believe currently we have two Disk Groups with the indexes in one and the data tables in the other. We have 32 GB of memory, and 32 CPUs. However, it is not the only thing running on the machine.
6) How much memory have you allocated to the database (pga, sga, memory_target, etc.)?
We have the memory_target set to 9GB, the db_cache_size set to 2GB, and the db_keep_cache_size set to 4.5GB. `pga_aggregate_target` is set to 0 (auto), as is `sga_target`.
(Since my initial request, we pinned the RDF_VALUE$ (~2.5GB) and C_PK_VID (~1.7GB) objects in the KEEP buffer cache, which drastically improved performance)
7) Are you using parallel query execution?
Yes, some of the more complex queries we run with the parallel hint set to 8.
8) Have you tried dynamic sampling?
Yes. We have ODS set to 3 for our more complex queries, we have not altered this much to see if there is a performance gained by changing this value.
Thanks again,
-Michael

Similar Messages

  • Oracle database performance after server reboot

    hi masters,
    this is not some kind of question, but a discussion. some statements come from our client that after weekly reboot of system, the oracle database performance is low for some time and increase after some time(say 2 days).
    i think it is but obvious, because at reboot oracle flushes all cache, and temporary space, so it need to re parse the sql statements and perform some disc I/O's so it might need some time and hence performance will degrade.
    but at the same time some people claim that after reboot their database performance is better than their normal performance for some days. it seems controversial that's why i am posting it here.
    what might be the reason behind this?? prior can have a valid reason of hard parsing, but what with second case??
    any clarification is highly appreciated...
    thank you
    regadrs
    VD

    Vikrant,
    You should wait for some time buddy, its weekend ;-) .
    this is not some kind of question, but a discussion. some statements come from our client that after weekly reboot of system, the oracle database performance is low for some time and increase after some time(say 2 days).i think it is but obvious, because at reboot oracle flushes all cache, and temporary space, so it need to re parse the sql statements and perform some disc I/O's so it might need some time and hence performance will degrade.
    >
    I would start from saying that checking the performance when the system just started, is a wrong approach. There would be lots of IOs , parsing, calculations(related to memory allocations) happening so there would be a delay/bad performance at that time. Very simple example can be parsing, another can be memory allocation. Oracle doesn't allocate the entire memory in the instance startup that is allocated to the memory areas but allocates just the bare minimum that is needed to start the instance and than after the startup, it keeps on allocating the memory. So surely enough, with the startup and after a while of it, there would be a different performance than that after the instance hsa already been started and the workload informations have started coming up.
    Its correct that Oracle would deallocate all the caches with the reboot as the instance is on the memory(physical) and with the reboot , that would be flushed including the SGA which is allocated over it. Temporary tablespace is now not freed with the reboot. I guess its a rather illogical thing to do but that's what is there now. Oracle keeps the segment allocated even after the reboot is issued, hence the reason for larger temporary tablespaces.
    >
    but at the same time some people claim that after reboot their database performance is better than their normal performance for some days. it seems controversial that's why i am posting it here.
    what might be the reason behind this?? prior can have a valid reason of hard parsing, but what with second case??
    >
    This should not come as a surprise once we understand what might be happening with this process. Assume a situation where you have undersized caches. For example, shared pool . which is very heavily used for database , if this is going to be undersized and you are not using automatic memory management, you won't be enjoying the dynamic management of this parameter. Now, if you do lots of parsing , thanks to your wrongly written queries, you would eventually end up filling up shared pool to its max thus leaving no space for incoming new hard parsed cursors. Here , if you can't manage to add more memory to add to it, the only solution left would be to flush the shared pool( as good as rebooting the db because this would do the same) and than make space for the new cursors. The performance is going to be better becausethe cursors would not be getting flushed out immediately and will be kept in the shared pool as long as its not filled up again.Once you have reached to limit of it, again there would be performance benefit. So there are always odds added to the statements like this that I rebuilt my index , I got better, I rebooted my server, my querie are much faster now. Most of the time when these kind of statements are given, they are based on what we have seen, without understading what actually might have happened. So I would siggest to hear the statement but not take them as a rule of thumb to follow.
    Hope this all makes some sense for you and would help somewhat.
    Aman....

  • General Oracle Database Performance trouble solving best practice Steps

    We use  Oracle 11g DB on Windows2008R2 as web application backend DB.
    We have peformance trouble in that DB.
    I would like to know General Oracle Database Performance trouble solving best practice Steps.
    Is there any good general best practice document for performace trouble solving in the internet ?

    @Girish Sharma:  I disagree with this. Many people say things like your phrase "..first identify the root cause and then move forward" but that is not the first step. Any such technique is nothing more than looking at some report, finding a number that you don't like, and attempting to "fix" it. Some people use that supposedly funny term "compulsive tuning disorder" (first used by Gaja Krishna Vaidyanatha) to describe this approach (also advocated in this topic by @Supriyo Dey). The first step must be to determine what the problem is. Until you know that, all those reports you mentioned (which, remember, require EE plus pack licences) are useless.
    @teradata0802, your best practice starts by finding the problem. Is it, for example, that the overnight batch jobs don't finish until lunchtime? A screen takes 10 seconds to refresh, and your target is one second? A report takes half an hour, but you need to run it every five minutes? Determine what business function is causing your client to lose money because it is too slow. Then investigate what it is doing, how, and why. You have to begin by focussing on the problem, not by running database-wide reports..

  • How does Oracle Database compare with IBM DB2?

    Hi
    thank you for reading my post
    How Does ORACLE Database compare with IBM DB2 ?
    Which one has better performance with similar workload?
    Which one perform better on System Z hardware and z/OS?
    Is there any document that compare these two product?
    Thanks.

    Hi
    The below link may useful to you.
    http://www.oracle.com/technology/deploy/availability/pdf/CWP_HA_Oracle10gR2_DB28.2.pdf
    With best regards
    Shan

  • Oracle database Performance testing

    What are the procedure to dp Oracle performance Testing???
    Is there any tools to do Oracle database Performance or load testing??

    insa wrote:
    OS- windows
    DB-Oracle11gr2
    Just i want check if 100 of users using my Application at a time what will be the performance of my database??Database performance is dependent on a number of factors - where application design, application code and data model, are critical components.
    And these components need to be designed, written, and implemented with performance and scalability as primary requirements. From design issues like using 3NF correctly, to coding issues like using bind variables correctly.
    You cannot simply slap design and code together, run that against the database, and then tune database performance. That is not the correct approach. Oracle database technology performs very well. It is the application that does not perform. It is the design that does not perform. It is the architecture that does not perform.
    When I see the comment that someone wants to test database performance, I see someone not understanding what determines performance - and incorrectly thinking that the database is a black box, and that this black box's performance can be tuned and set, irrespective of the data model, application code and architecture used.

  • NetBean, I can't establish the oracle database connection with version 9i.

    In NetBean, I can't establish the oracle database connection with version 9i.
    I have create the datasource and it is fine.
    when i try to access the tables i get a message
    "Unable to refresh children, connection is broken, unsupported feature"
    in netbeans.
    Does netbeans support oracle 9i version.
    Thanks,
    VKK

    In NetBean, I can't establish the oracle database connection with version 9i.
    I have create the datasource and it is fine.
    when i try to access the tables i get a message
    "Unable to refresh children, connection is broken, unsupported feature"
    in netbeans.
    Does netbeans support oracle 9i version.
    Thanks,
    VKK

  • SAP oracle database crash with errore:ORA-00490: PSP process terminated

    Hi ALL,
    Our oracle database crash with errore code in trace : ORA:00490; i started database again and it  working fine; but could not fine reason of down and what is this erroe all about
    can some help me
    Errors in file /oracle/SRD/saptrace/background/srd_pmon_28096.trc:
    ORA-00490: PSP process terminated with error
    Tue Nov 25 09:00:57 2008
    PMON: terminating instance due to error 490
    Instance terminated by PMON, pid = 28096
    Thanks,
    Dinesh

    Hi stefen,
    please find the trace file as below
    /oracle/SRD/saptrace/background/srd_pmon_28096.trc
    Oracle Database 10g Enterprise Edition Release 10.2.0.2.0 - 64bit Production
    With the Partitioning and Data Mining options
    ORACLE_HOME = /oracle/SRD/102_64
    System name:    SunOS
    Node name:      nzlsfn23
    Release:        5.10
    Version:        Generic_137111-01
    Machine:        sun4u
    Instance name: SRD
    Redo thread mounted by this instance: 1
    Oracle process number: 2
    Unix process pid: 28096, image: oracle@nzlsfn23 (PMON)
    2008-11-25 09:00:57.497
    SERVICE NAME:(SYS$BACKGROUND) 2008-11-25 09:00:56.210
    SESSION ID:(24.1) 2008-11-25 09:00:56.194
    Background process PSP0 found dead
    Oracle pid = 6
    OS pid (from detached process) = 28098
    OS pid (from process state) = 28098
    dtp = 38000afd8, proc = 497000860
    Dump of memory from 0x000000038000AFD8 to 0x000000038000B020
    38000AFD0                   00000005 00000000          [........]
    38000AFE0 00000004 97000860 00000000 00000000  [.......`........]
    38000AFF0 00000000 50535030 00020000 00000000  [....PSP0........]
    38000B000 00006DC2 00000000 00000000 48E50DA4  [..m.........H...]
    38000B010 00000001 000E3273 00040081 00000000  [......2s........]
    Dump of memory from 0x0000000497000860 to 0x0000000497001048
    497000860 02010000 00000000 00000000 00000000  [................]
    497000870 00000000 00000000 00000000 00000000  [................]
    497000880 00000004 97042570 00000004 97047810  [......%p......x.]
    497000890 00000004 97028E68 00000004 97045BE8  [.......h......[.]
    4970008A0 00000000 00000000 00000004 97045C70  [..............\p]
    4970008B0 00000004 97045C70 00000004 97047800  [......\p......x.]
    4970008C0 01060000 00000000 00000004 97025ED8  [..............^.]
    4970008D0 00000004 97028E68 00000006 00000000  [.......h........]
    4970008E0 00000000 00000000 00000000 00000000  [................]
    4970008F0 00000000 00000000 00000004 97042490  [..............$.]
    497000900 00000004 970425A0 00000000 00000000  [......%.........]
    497000910 00000000 00000000 00000000 00000000  [................]
            Repeat 3 times
    497000950 00000003 00000000 00000000 00000000  [................]
            Repeat 1 times
    497000970 00000000 00000000 00000000 00000000  [................]
    497000980 00000004 00000000 00000000 00000000  [................]
    497000990 00000003 00000000 00000000 00000000  [................]
            Repeat 1 times
    4970009B0 00000004 00000000 00000000 00000000  [................]
    4970009C0 00000005 00000000 00000000 00000000  [................]
    4970009D0 00000003 00000000 00000000 00000000  [................]
    4970009E0 00000000 00000000 00000000 00000000  [................]
            Repeat 8 times
    497000A70 00000000 00000000 00000004 97000A78  [...............x]
    497000A80 00000004 97000A78 00000000 00000000  [.......x........]
    497000A90 00000000 00000000 00000004 97000A98  [................]
    497000AA0 00000004 97000A98 00000000 00000000  [................]
    497000AB0 00000000 00000000 00000000 00000000  [................]
            Repeat 2 times
    497000AE0 00000000 00000000 00000018 00000030  [...............0]
    497000AF0 00000001 00000B3D 00000004 970037D0  [.......=......7.]
    497000B00 00000004 580096B0 00000001 00000000  [....X...........]
    497000B10 00000000 00000000 00000000 00000000  [................]
            Repeat 2 times
    497000B40 00006DC2 00000000 00000000 00000000  [..m.............]
    497000B50 00000000 00000000 00000000 00000000  [................]
            Repeat 2 times
    497000B80 00000004 97000860 00000000 00000000  [.......`........]
    497000B90 00000000 00000000 00000000 00000000  [................]
            Repeat 7 times
    497000C10 00000004 97000C10 00000004 97000C10  [................]
    497000C20 00000000 00000000 00010000 00000000  [................]
    497000C30 00000000 00000117 0000000A 00000000  [................]
    497000C40 00006DC2 00000000 00000000 48E50DA4  [..m.........H...]
    497000C50 00000001 00000000 00000000 00000000  [................]
    497000C60 00000000 00000000 00000000 00000000  [................]
            Repeat 2 times
    497000C90 00000000 00000000 00000003 FFFFFFFF  [................]
    497000CA0 00000000 00000000 00000000 00000000  [................]
            Repeat 13 times
    497000D80 73726461 646D0000 00000000 00000000  [srdadm..........]
    497000D90 00000000 00000000 00000000 00000000  [................]
    497000DA0 00000000 00000006 6E7A6C73 666E3233  [........nzlsfn23]
    497000DB0 00000000 00000000 00000000 00000000  [................]
            Repeat 2 times
    497000DE0 00000000 00000000 00000000 00000008  [................]
    497000DF0 554E4B4E 4F574E00 00000000 00000000  [UNKNOWN.........]
    497000E00 00000000 00000000 00000000 00000000  [................]
    497000E10 00000000 00000008 32383039 38000000  [........28098...]
    497000E20 00000000 00000000 00000000 00000000  [................]
    497000E30 00000000 00000005 6F726163 6C65406E  [........oracle@n]
    497000E40 7A6C7366 6E323320 28505350 30290000  [zlsfn23 (PSP0)..]
    497000E50 00000000 00000000 00000000 00000000  [................]
    497000E60 00000000 00000000 00000000 00000016  [................]
    497000E70 00000000 00000002 00000000 00000000  [................]
    497000E80 00000000 00000000 00000000 00000000  [................]
            Repeat 8 times
    497000F10 00000000 00000000 00000000 00020000  [................]
    497000F20 00000000 00000000 00000000 00000000  [................]
    497000F30 00000000 00000000 00000003 9E1F6748  [..............gH]
    497000F40 00000004 97001728 00000004 97000758  [.......(.......X]
    497000F50 00000000 00000000 00000003 9E26B5B0  [.............&..]
    497000F60 00000000 00000000 00000000 00000000  [................]
            Repeat 1 times
    497000F80 00000004 97000F80 00000004 97000F80  [................]
    497000F90 00000000 00040000 00000000 00000000  [................]
    497000FA0 00000000 00031A55 00000000 0004D7DD  [.......U........]
    497000FB0 00000000 00071A55 00000000 00000000  [.......U........]
    497000FC0 00000000 00000000 00000000 00000000  [................]
    497000FD0 00000000 00000828 00000000 000000E0  [.......(........]
    497000FE0 00000000 00000828 00000000 00000000  [.......(........]
    497000FF0 00000000 00000000 00000000 00000000  [................]
            Repeat 4 times
    497001040 00000002 00000000                    [........]
    error 490 detected in background process
    ORA-00490: PSP process terminated with error

  • Oracle Spatial Performance with 10-20.000 users

    Does anyone have any experience when Oracle Spatial is used with say 20.000 concurrent users. I am not interested in MapViewer response time, but lets say there is:
    - an app using 800 different tables each having an sdo_geometry column
    - the app is configured with different tables visible on different view scales
    - let's say an average of 40-50 tables is visible at any given time
    - some tables will have only a few records, while other can hold millions.
    - there is no client side caching
    - clients can zoom in/out pan.
    Anwers I am interested in:
    - What sort of server would be required
    - How can Oracle serve all that data (each Refresh renders the map and retrieves the data over the wire as there is no client side caching).
    - What sort of network infrastructure would be required.
    - Can clients connect to different servers and hence use load balancing or does Oracle have an automatic mechanism for that?
    Thanks in advance,
    Patrick

    Patrick, et al.
    There are lots of things one can do to improve performance in mapping environments because of a lot of the visualisation is based on "background" or read-only data. Here are some "tips":
    1. Spatially sort read-only data.
    This tip makes sure that data that is close to each other in space are next to each other on disk! Dan gave a good suggestion when he referenced Chapter 14, "Reorganize the Table Data to Minimize I/O" pp 580- 582, Pro Oracle Spatial. But just as easily one can create a table as select ... where sdo_filter() where the filtering object is an optimized rectangle across the whole of the dataset. (This is quite quick on 10g and above but much slower on earlier releases.)
    When implementing this make sure that the created table is created such that its blocks are next to each other in the tablespace. (Consider tablespace defragmentation beforehand.) Also, if the data is READ ONLY set the PCTFREE to 0 in order to pack the data up into as small a number of blocks as possible.
    2. Generalise data
    Rendering spatial data can be expensive where the data is geometrically detailed (many vertices) esp where the data is being visualised at smaller scales than it was captured at. So, if your "zoom thresholds" allow 1:10,000 data to be used at 1:100,000 then you are going to have problems. Consider pre-generalising the data (see sdo_util.simplify) before deployment. You can add multiple columns to your base table to hold this data. Be careful with polygon data because generalising polygons that share boundaries will create gaps etc as the data is more generalised. Often it is better to export the data to a GIS which can maintain the boundary relationships when generalising (say via topological relationships).
    Oracle's MapViewer has excellent on-the-fly generalisation but here one needs to be careful. Application tier caching (cf Bryan's comments) can help here a lot.
    3. Don't draw data that is sub-pixel.
    As one zooms out objects become smaller and smaller until they reach a point where the whole object can be drawn within a single pixel. If you have control over your map visualisation application you might want to consider setting the SDO_FILTER parameter "min_resolution" flag dynamically so that its value is the same as the number of meters / pixel (eg min_resolution=10). If this is set Oracle Spatial will only include spatial objects in the returned search set if one side of a geometry's MBR is greater than or equal to this value. Thus any geometries smaller than a pixel will not be returned. Very useful for large scale data being drawn at small scales and for which no selection (eg identify) is required. With Oracle MapViewer this behaviour can be set via the generalized_pixels parameter.
    3. SDO_TOLERANCE, Clean Data
    If you are querying data other than via MBR (eg find all land parcels that touch each other) then make sure that your sdo_tolerance values are appropriate. I have seen sites where data captured to 1cm had an sdo_tolerance value set to a millionth of a meter!
    A corollary to this is make sure that all your data passes validation at the chosen sdo_tolerance value before deploying to visualisation. Run sdo_geom.validate_geometry()/validate_layer()...
    4. Rtree Spatial Indexing
    At 10g and above lots of great work went in to the RTree indexing. So, make sure you are using RTrees and not QuadTrees. Also, many GIS applications create sub-optimal RTrees by not using the additional parameters available at 10g and above.
    4.1 If your table/column sdo_geometry data contains only points, lines or polygons then let the RTree indexer know (via layer_gtype) as it can implement certain optimizations based on this knowledge.
    4.2 With 10g you can set the RTree's spatial index data block use via sdo_pct_free. Consider setting this parameter to 0 if the table/column sdo_geometry data is read only.
    4.3 If a table/column is in high demand (eg it is the most commonly used table in all visualisations) you can consider loading (a part of) the RTree index into memory. Now, with the RTree indexing, the sdo_non_leaf_tbl=true parameter will split the RTree index into its leaf (contains actual rowid reference) and non-leaf (the tree built on the leaves) components. Most RTrees are built without this so only the MDRT*** secondary tables are built. But if sdo_non_leaf_tbl is set to true you will see the creation of an additional MDNT*** secondary table (for the non_leaf part of the rtree index). Now, if appropriate, the non_leaf table can be loaded into memory via the following:
    ALTER TABLE MDNT*** STORAGE(BUFFER_AREA KEEP);
    This is NOT a general panacea for all performance problems. One should investigate other options before embarking on this (cf Tom Kyte's books such as Expert Oracle Database Architecture, 9i and 10g Programming Techniques and Solutions.)
    4.4 Don't forget to check your spatial index data quality regularly. Because many sites use GIS package GUI tools to create tables, load data and index them, there is a real tendency to not check what they have done or regularly monitor the objects. Check the SDO_RTREE_QUALITY column in USER_SDO_INDEX_METADATA and look for indexes with an SDO_RTREE_QUALITY setting that is > 2. If > 2 consider rebuilding or recreating the index.
    5. The rendering engine.
    Whatever rendering engine one uses make sure you try and understand fully what it can and cannot do. AutoDesk's MapGuide is an excellent product but I have seen it simply cache table/column data and never dynamically access it. Also, I have been at one site which was running Deegree and MapViewer and MapViewer was so fast in comparison to Deegree that I was called in to find out why. I discovered that Deegree was using SDO_RELATE(... ANYINTERACT ...) for all MBR queries while MapViewer was using SDO_FILTER. Just this difference was causing some queries to perform at < 10% of the speed of MapViewer!!!!
    6. Consider "denormalising" data
    There is an old adage in databases that is "normalise for edit, denormalise for performance". When we load spatial data we often get it from suppliers in a fairly flat or normalised form. In consort with spatial sorting, consider denormalising the data via aggregations based on a rendering attribute and some sort of spatial unit. For example, if you have 1 million points stored as single points in SDO_GEOMETRY.SDO_POINT which you want to render by a single attribute containing 20 values, consider aggregating the data using this attribute AND some sort of spatial BUCKET or BIN. So, consider using SDO_AGGR_UNION coupled with Spatial Analysis and Mining package functions to GROUP the data BY <<column_name>> and a set of spatial extents.
    6. Tablespace use
    Finally, talk to your DBA in order to find out how the oracle database's physical and logical storage is organised. Is a SAN being used or SAME arranged disk arrays? Knowing this you can organise your spatial data and indexes using more effective and efficient methods that will ensure greater scalability.
    7. Network fetch
    If your rendering engine (app server) and database are on separate machines you need to investigate what sort of fetch sizes are being used when returning data from queries to the middle-tier. Fetch sizes for attribute only data rows and rows containing spatial data can be, and normally are, radically different. Accepting the default settings for these sizes could be killing you (as could the sort_area_size of the Oracle session the application server has created on the database). For example I have been informed that MapInfo Pro uses a fixed value of 25 records per fetch when communicating with Oracle. I have done some testing to show that this value can be too small for certain types of spatial data. SQL Developer's GeoRaptor uses 100 which is generally better (but this one can modify this). Most programmers accept defaults for network properties when programming in ADO/ODBC/OLEDB/JDBC: just be careful as to what is being set here. (This is one of the great strengths of ArcSDE: its TCP/IP network transport is well written, tuneable and very efficient.)
    8. Physical Format
    Finally, while Oracle's excellent MapViewer requires data its spatial data to be in Oracle, other commercial rendering engines do not. So, consider using alternate, physical file formats that are more optimal for your rendering engine. For example, Google Earth Enterprise "compiles" all the source data into an optimal format which the server then serves to Google Earth Enterprise clients. Similarly, a shapefile on local disk to the application server (with spatial indexing) may be faster that storing the data back in Oracle on a database server that is being shared with other business databases (eg Oracle financials). If you don't like this approach and want to use Oracle only consider using a dedicated Oracle XE on the application server for the data that is read only and used in most of your generated maps eg contour or drainage data.
    Just some things to think about.
    regards
    Simon

  • Oracle 9iFS performance with WebService

    Hello,
    We use Oracle 9iFS with Oracle iAS and Oracle Database. We have implemented WebServices that use iFS API to upload/view/modify the documents in iFS. Even for a file that is ~5Mb, the performance is poor. Do we have any Benchmark testing results so that we can compare our results? We also get some Errors every now and then, which locks up the iFS instance and can only be solved by restarting the instance, the errors are like follows
    oracle.ifs.common.IfsException: IFS-21008: Unable to connect to iFS service
    oracle.ifs.common.IfsException: IFS-12214: Unable to get item in collection by name (CLASSOBJECT)
    oracle.ifs.common.IfsException: IFS-12201: Unable to resolve collection
    oracle.ifs.common.IfsException: IFS-20008: Unable to prepare query statement
    java.sql.SQLException: Refcursor value is invalid
    or
    oracle.ifs.common.IfsException: IFS-21008: Unable to connect to iFS service
    oracle.ifs.common.IfsException: IFS-21002: Unable to construct S_LibraryObject for class oracle.ifs.server.S_LibrarySession (unexpected Exception occurred in invoked method)
    java.lang.Exception: java.lang.OutOfMemoryError
    or
    oracle.ifs.common.IfsException: IFS-21008: Unable to connect to iFS service
    oracle.ifs.common.IfsException: IFS-20001: Unable to get schema version
    oracle.ifs.common.IfsException: IFS-20000: Unable to get repository parameter (SCHEMAVERSION)
    java.sql.SQLException: Bigger type length than Maximum
    Are these errors because of - Not applying patches or some problem with concurrent iFS access. We have identified an issue with memory leak and are working to resolve it, but we feel that there maybe more issues. Any help would be appreciated, thanks

    That depends on at least three things:
    (1) what is the range of words that will appear in the document set (wide range of documents = smaller resultsets = better performance)
    (2) how precise are the user's queries likely to be (more precise = smaller resultsets = better performance)
    (3) how many milliseconds are your users willing to wait for results
    So, unfortunately, you'll probably have to experiment a bit before you'll know...

  • Oracle 10g  – Performance with BIG CONTEXT indexes

    I would like to use Oracle XE 10.2.0.1.0 only for the full-text searching of the files residing outside the database on the FTP server.
    Recently I have found out that size of the files to be indexed is 5GB.
    As I have read somewhere on this forum before size of the index should be 30-40% of the indexed text files (so with formatted documents like PDF or DOC even less).
    Lets say that the CONTEXT index size over these files will be 1.5-2GB.
    Number of the concurrent user will be max. 5.
    I can not easily test it my self yet.
    Does anybody have any experience with Oracle XE or other Oracle Database edition performance with the CONTEXT index this BIG?
    Will Oracle XE hardware resources license limitation be sufficient to handle one CONTEXT indexe this BIG?
    (Oracle XE license limitations: 1 GB RAM and 1 CPU)
    Regards.

    That depends on at least three things:
    (1) what is the range of words that will appear in the document set (wide range of documents = smaller resultsets = better performance)
    (2) how precise are the user's queries likely to be (more precise = smaller resultsets = better performance)
    (3) how many milliseconds are your users willing to wait for results
    So, unfortunately, you'll probably have to experiment a bit before you'll know...

  • Oracle database integration with SAP PI for high volume & Complex Structure

    Hi
    We have requirement for integrating oracle database to SAP PI 7.0 for sending data which is eventually transferred to multiple receivers. The involved data structure is hugely complex (around 18 child tables) with high volume processing requirement (100K+ objects need to be processed in 6-7 hours). We need to implement logic for prioritizing the object i.e. high priority objects must be processed first and then objects with normal priority.
    We could think of implementing this kind of logic in database procedures (at least it provides flexibility for implementing data selection logic as well as processed data can be marked as success in the same SP) but since PI sender adapter doesn't support calling Oracle stored procedures currently so this option is rules out. we can try implementing complex data selection using oracle table function but table function doesn't allow any SQL query which changes data (UPDATE, INSERT, DELETE etc) so it is impossible to mark selected objects in table function from PI communication channel "Update Query" option.
    Also, we need to make sure that we are not processing all the objects at once as message size for 20 objects can vary from 100 KB to 15 MB which could really lead to serious performance issues for bigger messages.
    Please share any implementation experience for handling issues:
    1 - Database Integration involving Oracle at sender side
    2 - Complex Data structures
    3 - High Volume Processing
    4 - Controlled data selection from database to contro the message size in PI
    Thanks,
    Panchdev

    Hi,
          We can call the stored procedure using receiver adapter using ccBPM, we can follow different approaches for reading the data in this case.
    a) In this  a ccBPM instance needs to be triggered using some dummy message, after receiving this message the ccBPM can make  a sync call to the Oracle database the store procedure(this can be done using the specific receiver data type strucure), on getting the response message the ccBPM  can then proceed with the further steps.The stored procedure needs to be optimized for improving the performance as the mapping complexity will largely get affected by the structure in which the stored procedure returns the message.Prioritization of the objects can be handled in the stored procedure.
    b) In this a ccBPM instance can first read data from the header level table, then it can make subsequent sync calls to Oracle tables for reading data from the child tables.This approach is less suitable for this interface as the number child tables is big.
    Pravesh.

  • How to get start to develop Oracle database applications with NetBeans?

    Dear Ladies and gentlemen,
    I am new to NetBeans. And I know nothing about Oracle. I have installed the
    NetBeans 6.8, and I want to learn Oracle database application development
    with NetBeans. I don't know how to get start. Give me some suggestions,
    please. Thank you very much.
    Best Regards
    Ma Xiaoguang

    Download Oracle Express Edition for educational purposes and start from there.
    Download at:
    http://www.oracle.com/technetwork/database/express-edition/downloads/index.html
    Documentation at:
    http://www.oracle.com/pls/xe102/homepage

  • Can oracle database interact with sysbase database in real-time?

    There are two database,one is oracle,one is sysbase,sysbase is produce database.My program only can visit oracle database,and can't directly visit sysbase database.My question is whether there is a directly link from oracle database to sysbase database,when I visit oracle,oracle database can interact with sysbase database through this link real time? How to realize above function? Please give plan for it.
    Thanks in advance!
    Best regards,
    Edward

    Yes there is. You have to create heterogeneous connection between these two databases
    Take a look at below links:
    http://www.dba-oracle.com/t_heterogeneous_database_connections_sql_server.htm
    http://asktom.oracle.com/pls/asktom/f?p=100:11:0::::P11_QUESTION_ID:13401716754850
    http://download-west.oracle.com/docs/cd/B10501_01/server.920/a96571/repmultdemo.htm

  • Oracle Database XE with Microsoft Windows Server 2008 (32-bit)

    I would like to find out if Oracle Database XE works with Microsoft Windows Server 2008 (32-bit)?
    If not, is there any other Oracle Open Source Database version can work with Microsoft Windows Server 2008 (32-bit)?
    Any help would be greatly appreciated!

    I am installing W2k8 R2 Domain Controller with Oracle Express 10g.
    This is the tip to make it work.
    Go to the installer OracleXE.exe -> Right click + Select Properties + Compatibility + Run this program in Compatibility Mode fot + Windows 2003 Sp1
    Then Select Run this Program as administrator.
    After this run the OracleXe.exe and it will install properly.
    Remember to double check that everything works fine.

  • Multiple oracle database instances with different characterset on  the same server

    Hello,
    Is it possible to have 2 database instances running with different charactersets,one with AL32UTF8 and the other with WE8MSWIN1252.?
    Are there any setup requirements to be performed prior to setting up the database instances?
    The 3rd party utility that we want to use does not support AL32UTF8 and insists on using a database with character set WE8MSWIN1252.
    Kindly help.
    Thanks,
    Ram.

    Hello Zhe,
    I guess I posted my question in a wrong forum.  I tried my best to find a suitable forum and thought this was the best and closest I found.  Apparently not.  Can you please let me know the right forum for my question?
    The below is the breif of what we are currently facing:
    We are in the process of finalizing plans to install Automic for our Retail applications to schedule jobs.  In the process came to know that Automic does not support AL32UTF8.
    Right now we have RMS and UC4 (now called Automic) run on the database server as UC4 supports AL32UTF8 and schema for UC4 is inside the RMS database.
    Going forward it is recommended to have a separate database instance on the same server as that of RMS database and with a different characterset which is WE8MSWIN1252.
    Please let me know what forum to post in, I will repost the question.
    Thanks,
    Ram.

Maybe you are looking for