Oracle Apps Database severe Performance Issue

Hi Gurus,
This is regarding a severe performance issue running in our Production E-Business Suite Instance.
its an R12.1.3 setup installed with 11.2.0.1 Database. All the servers are Solaris Sparc 64 (Solaris 10)
Let me brief you about the instance first:
2 Node Application
- Main Application Server hosting web/forms/concurrent/admin servers
- iSupplier server hosting web services (placed in DMZ, used by external suppliers via Internet)
1 Node Database Server
Database Server Specs
Memory: 144G phys mem 20G total swap
- CPUs (8Px4cores, 2Px2cores)
- I/O - fiber channel hard disk (hitachi SAN Storage) - 7 DATA_TOPs (7 drives with RAID 5) - current DB size 1.6 TB
- at peak load, around 1000 concurrent forms session and 2000 web sessions.
We have been facing some serious performance issues and we raised an SR with Oracle Support.
The Support analyzed a bunch of AWR Reports we provided them and they asked us to increase the DB_CACHE from its current usage of 27G to 40G
So, we changed SGA_TARGET from 35G to 50G and PGA was increased from 35G to 40G as v$pgastat was also suggesting some lack of memory.
We made these changes last night.
Today morning we observed the following:
1. after start of office hours, we checked in the home page of EM DB Console that ADDM was showing reduced impact due to lack of SGA memory which seemed to be a good sign. Earlier it was around 25% which was now at 12%.
However, negative aspects were:
1. lot of swapping was reported by the System Administrators on the DB Server
2. High CPU Usage
3. EM DB Console showed a lot of "Concurrency Wait Class" events ...throughout the day lot of blocking sessions were reported which were making other sessions to wait.
in the AWR Report, following foreground reports were listed:
Top 5 Timed Foreground Events
Event
Waits
Time(s)
Avg wait (ms)
% DB time
Wait Class
DB CPU
132,577
61.46
library cache lock
3,539
40,683
11496
18.86
Concurrency
library cache: mutex X
4,014,083
21,011
5
9.74
Concurrency
db file sequential read
4,138,014
20,767
5
9.63
User I/O
latch free
381,916
5,897
15
2.73
Other
This is showing "library cache lock" events as the main culprit apart from the usual suspect, the CPU.
I am attaching the AWR Report. Please let me know if  i should revert back the memory changes or is there anything else i could do.
Please help us resolving it because the performance is going worst.
Regards,
Muneer.

Pl do not post duplicates - Oracle Apps Database severe Performance Issue
For all critical production issues, pl work with Support thru SRs - using the forums to troubleshoot production issues is not wise

Similar Messages

  • Severe performance issues in production database

    Hi Experts,
    we have configured RMAN in our production database recently using some 3 party tool COMMVAULT.
    Problem is just 47 GB database taking around 6 hrs of time to complted the bakcup job. Please let me know what could be the reason.
    Further to this issues i found out some of few things, our application vendor commissioned this database server and it seems that they done some changes in TIMEZONE settings.
    when i query against some of few dictionary tables for example dba_schedular_jobs i am getting the following error.
    ORA-01882: timezone region %s not found
    when ever i connect the database , the connection itself is very slow and suffering severe performance issues in my production database.
    You help would be much appreciated.
    Regards,
    Salai

    Hi,
    also let us know if you use asm or local file system for datafiles.
    Your backup strategy will also be helpful:
    Do you make compressed backups?
    Full or incremental?
    To where do you backup the database? To the local filesystem, SAN Volume, Offsite Storage, Tape storage?
    Is there any other jobs running while the RMAN job is running?
    Please post the stats that you have gathered over the time period when the backup is running.
    Thanks.

  • Multi Select Choice on af:query has severe performance issue

    Multi-select choice used with af:query through a view criteria is causing severe performance issue on deselection of "All" checkbox, if the data in the list is around 550 rows. The same component works absolutely fine when used in a form layout.
    I can provide you a re-producible test case, if anyone needs it!
    ***: This is a customer environment issue, and customer is eager to have multi-select in this case. Appreciate any help!

    Glimpse of repetitive lines from console for the above scenario:
    <DCUtil> <findSpelObject> [2208] DCUtil, returning:oracle.jbo.uicli.binding.JUApplication, for TestSelectChoiceDefaultAMDataControl
    <ADFLogger> <begin> Attaching an iterator binding to a datasource
    <DCIteratorBinding> <getViewObject> [2209] Resolving VO:TestSelectChoiceDefaultAM._SESSION_SHARED_APPMODULE_NAME_.SessionAM.DeptReadOnly1 for iterator binding:noCtrl_oracle_adfinternal_view_faces_model_binding_FacesCtrlListBinding_59List_60
    <DCUtil> <findSpelObject> [2210] DCUtil, RETURNING: <null> for TestSelectChoiceDefaultAM._SESSION_SHARED_APPMODULE_NAME_.SessionAM.DeptReadOnly1
    <ADFLogger> <addContextData> Attaching an iterator binding to a datasource
    <ADFLogger> <addContextData> Get LOV list
    <ADFLogger> <begin> Get LOV list
    <DCUtil> <findSpelObject> [2211] DCUtil, returning:oracle.jbo.uicli.binding.JUApplication, for TestSelectChoiceDefaultAMDataControl
    <ADFLogger> <begin> Attaching an iterator binding to a datasource
    <DCIteratorBinding> <getViewObject> [2212] Resolving VO:TestSelectChoiceDefaultAM._SESSION_SHARED_APPMODULE_NAME_.SessionAM.DeptReadOnly1 for iterator binding:noCtrl_oracle_adfinternal_view_faces_model_binding_FacesCtrlListBinding_123List_124
    <DCUtil> <findSpelObject> [2213] DCUtil, RETURNING: <null> for TestSelectChoiceDefaultAM._SESSION_SHARED_APPMODULE_NAME_.SessionAM.DeptReadOnly1
    <ADFLogger> <addContextData> Attaching an iterator binding to a datasource
    <ADFLogger> <addContextData> Get LOV list
    .....many times followed by
    <ADFLogger> <addContextData> Attaching an iterator binding to a datasource
    <ADFLogger> <addContextData> Get LOV list
    <ADFLogger> <begin> Get LOV list
    <ADFLogger> <addContextData> Get LOV list
    <ADFLogger> <begin> Get LOV list
    <ADFLogger> <addContextData> Get LOV list
    <ADFLogger> <begin> Get LOV list
    <ADFLogger> <addContextData> Get LOV list
    <ADFLogger> <begin> Get LOV list
    ...many times

  • CLONE ORACLE APPS DATABASE TO OTHER MACHINE

    Hi,
    I want to know how I can clone oracle APPS database to other machine.
    Just want to know the database part.
    Thanks in advance

    You really, really want to ask this in the appropriate Oracle Apps forum. Cloning the database may or may not be sufficient-- most packaged applications have their own data dictionaries that may need to be manipulated in application-specific ways in order to properly clone the application. The folks in the Oracle Apps forum should be able to tell you what scripts, if any, you'd need to run for your particular Oracle Apps version and database.
    Justin

  • Oracle 9i reading BLOB performance issues

    Windows XP Pro SP2
    JDK 1.5.0_05
    Oracle 9i
    Oracle Thin Driver for JDK 1.4 v.10.2.0.1.0
    DBCP v.1.2.1
    Spring v1.2.7 (I am using the JDBC template for convenience)
    I have run into serious performance issues reading BLOBs from Oracle using oracle's JDBC thin driver. I am not sure if it a constraint/mis-configuration with oracle or a JDBC problem.
    I am hoping that someone has some experience accessing multi-MB BLOBs under heavy volume.
    We are considering using Oracle 8 or 9 as a document repository. It will end up storing hundreds of thousands of PDFs that can be as large as 30 MBs. We don't have access to Oracle 10.
    TESTS
    I am running tests against Oracle 8 and 9 to simulate single and multi-threaded document access. Out goal is to get a sense of KBps throughput and BLOB data access contention.
    DATA
    There is a single test table with 100 rows. Each row has a PK id and a BLOB field. The blobs range in size from a few dozen KB to 12MB. They represent a valid sample of production data. The total data size is approx. 121 MBs.
    Single Threaded Test
    The test selects a single blob object at a time and then reads the contents of the blob's binary input stream in 2 KB chunks. At the end of the test, it will have accessed all 100 blobs and streamed all 121 MBs. The test harness is JUnit.
    8i Results: On 8i it starts and terminates successfully on a steady and reliable basis. The throughput hovers around 4.8 MBps.
    9i Results: Similar reliability to 8i. The throughput is about 30% better.
    Multi-Threaded Test
    The multi-threaded test uses the same "blob reader" functionality used in the single threaded test. However, it spawns 8 threads each running a separate "blob reader".
    8i Results: The tests successfully complete on a reliable basis. The aggregate throughput of all 8 threads is a bit more than 4.8 MBps.
    9i Results: Erratic. The tests were highly erratic on 9i. Threads would intermittently lock when accessing a BLOB's output stream. Sometimes they lock accessing data from the same row, othertimes it is distinct rows. The number and the timing of the thread "locks" is indeterminate. When the test completed successfully the aggregate throughput of the 8 threads was approx. 5.4 MBps.
    I would be more than happy to post code or the data model if that would help.
    Carlos

    Hi Murphy16,
    Try investigate where are the principal issues in your RAC system.
    Check:
    * Expensive SQL's;
    * Sorts in disks;
    * Wait Events;
    * Interconnect hardware issues;
    * Applications doing unnecessary manual LOCKs (SQL);
    * If SGA is adequatly sized (take care to not use of SWAP space "DISK");
    * Backup's and unnecessary jobs running at business time (Realocate this jobs and backups to night window or a less intensive work hour at database);
    * Rebuild indexes and identify tables that must be reorganized (fragmentation);
    * Verify another software consuming resources on your server;
    Please give us more info about your environment. The steps above are general, but you can use to guide u in basic performance issues.
    Regards,
    Rodrigo Mufalani
    http://mufalani.blogspot.com

  • Migrate Oracle apps database from AIX to Linux

    Hello,
    We are planning to move oracle 11i apps database from AIX to Linux. If anyone done recently please share your knowledge.
    Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bit
    PL/SQL Release 10.2.0.4.0 - Production
    CORE 10.2.0.4.0 Production
    TNS for IBM/AIX RISC System/6000: Version 10.2.0.4.0 - Productio
    NLSRTL Version 10.2.0.4.0 - Production
    Thanks
    Prince

    user12094010 wrote:
    Hello,
    We are planning to move oracle 11i apps database from AIX to Linux. If anyone done recently please share your knowledge.
    Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bit
    PL/SQL Release 10.2.0.4.0 - Production
    CORE 10.2.0.4.0 Production
    TNS for IBM/AIX RISC System/6000: Version 10.2.0.4.0 - Productio
    NLSRTL Version 10.2.0.4.0 - Production Hi Prince,
    To move/migrate the 11i Apps Database alone means you need to follow the export/import method.
    Please refer the note: 362205.1 "10g Release 2 Export/Import Process for Oracle Applications Release 11i"
    In general steps to follow the DB migration is :
    1. create the complete export file in source database
    2.create a new DB in the target instance
    3.perform the import
    4.create the new auto-config and sync with apps tier
    -Rk

  • Oracle Apps Database

    Hello All,
    I am Working on Oracle Apps Data base and suddenly my System was Haning and Restart it self when i log in oracle apps data base it give me a instering problem when i enter this query
    Select * from Cat ;
    it gives me Error
    ORA-03113:End of File on Communication Channel
    when i use the Table it Gives me Data
    Any Body tell me how this Problem Solve .
    Thanks

    Connect as sysdba and verify that the database is open. Also, make sure you can login with any other user (Apps) and try to run some queries, check if you get the same error or not.

  • Oracle Advance Compression Deletion Performance issue in 11g R1

    Hi,
    We have implemented OAC in our datawarehouse environment to enable table and index compression. We tested in our Test machine and we gained almost 600GB due to advance compression without any issues and all the informatica loads are running fine. And hence we implemented the same in our production but unfortunately two sessions which are involving deletion of data are taking more time (3 times of actual timing) for completion which affects our production environment.
    The tables creating issue are all non partitioned tables.
    I need to know whether Oracle Advance Compression will decrease delete performance? and is there any way to disable advance compression on those particular tables?
    Our environment details:
    DB earlier version: 11.1.0.6
    DB current version : Oracle 11.1.0.7
    Applied PSU: 11.1.0.7.6
    Operating system: Solaris 5.9
    Syntax used for compression:
    ALTER TABLE TABLE_NAME MOVE COMPRESS FOR ALL OPERATIONS;
    Thanks in Advance.

    Hi,
    Thanks for your reply.
    The note is for update performance issue and also I have applied necessary patches for improving update performance.
    The update sessions are all working fine. only the deletion sessions are creating problem.
    Could someone help me out to clear this problem.
    Thanks,
    VBK

  • Oracle Apps adapter: User name issue

    Hi All,
    I am using SOA 11g to call an API in EBS using Oracle Apps Adapter .I am passing User name,Responsibility Name and other parameteres in invoke properties. Now the API is called successfully with most of the user names while for few usenames it's not able able to resolve user id based on user name. Although if the API is called from Oracle Apps/PL SQL ,it is working for all the users.
    Please suggest.
    Thanks in advance

    At what stage are you seeing this error?

  • "oracle database 11g performance issues"

    Hai everybody,
    In oracle 11g 11.2.0.1.0 we are developing business application using java, Our developers said Database performance is very poor it takes more time to retrieve values from database, they check froantend and middleware that has no problem, if the query returns less values or rows it takes little more time. How can i solve this problem? please help me
    Regards Benk

    hai,
    sorry for the delay my @$ORACLE_HOME/rdbms/admin/ashrpt.sql output shown below
    ASH Report For ORCL/orcl
    DB Name DB Id Instance Inst Num Release RAC Host
    ORCL 1295420332 orcl 1 11.2.0.1.0 NO node6.node6-
    CPUs SGA Size Buffer Cache Shared Pool ASH Buffer Size
    2 1,561M (100%) 480M (30.7%) 432M (27.7%) 4.0M (0.3%)
    Analysis Begin Time: 04-Oct-12 16:44:16
    Analysis End Time: 04-Oct-12 16:59:33
    Elapsed Time: 15.3 (mins)
    Begin Data Source: V$ACTIVE_SESSION_HISTORY
    End Data Source: V$ACTIVE_SESSION_HISTORY
    Sample Count: 3
    Average Active Sessions: 0.00
    Avg. Active Session per CPU: 0.00
    Report Target: None specified
    Top User Events DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    Avg Active
    Event Event Class % Event Sessions
    null event Other 33.33 0.00
    Top Background Events DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    Avg Active
    Event Event Class % Activity Sessions
    CPU + Wait for CPU CPU 33.33 0.00
    os thread startup Concurrency 33.33 0.00
    Top Event P1/P2/P3 Values DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    No data exists for this section of the report.
    Top Service/Module DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    Service Module % Activity Action % Action
    SYS$BACKGROUND UNNAMED 66.67 UNNAMED 66.67
    SYS$USERS UNNAMED 33.33 UNNAMED 33.33
    Top Client IDs DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    No data exists for this section of the report.
    Top SQL Command Types DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    No data exists for this section of the report.
    Top Phases of Execution DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    Avg Active
    Phase of Execution % Activity Sessions
    SQL Execution 33.33 0.00
    Top SQL with Top Events DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    No data exists for this section of the report.
    Top SQL with Top Row Sources DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    No data exists for this section of the report.
    Top SQL using literals DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    No data exists for this section of the report.
    Top Parsing Module/Action DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    No data exists for this section of the report.
    Top PL/SQL Procedures DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    No data exists for this section of the report.
    Top Java Workload DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    No data exists for this section of the report.
    Top Call Types DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    No data exists for this section of the report.
    Top Sessions DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    -> '# Samples Active' shows the number of ASH samples in which the session
    was found waiting for that particular event. The percentage shown
    in this column is calculated with respect to wall clock time
    and not total database activity.
    -> 'XIDs' shows the number of distinct transaction IDs sampled in ASH
    when the session was waiting for that particular event
    -> For sessions running Parallel Queries, this section will NOT aggregate
    the PQ slave activity into the session issuing the PQ. Refer to
    the 'Top Sessions running PQs' section for such statistics.
    Sid, Serial# % Activity Event % Event
    User Program # Samples Active XIDs
    10, 1 33.33 CPU + Wait for CPU 33.33
    SYS [email protected] (ARC2) 1/917 [  0%] 0
    19, 148 33.33 null event 33.33
    SYS [email protected] (J000) 1/917 [  0%] 0
    139, 3 33.33 os thread startup 33.33
    SYS [email protected] (CJQ0) 1/917 [  0%] 0
    Top Blocking Sessions DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    No data exists for this section of the report.
    Top Sessions running PQs DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    No data exists for this section of the report.
    Top DB Objects DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    No data exists for this section of the report.
    Top DB Files DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    No data exists for this section of the report.
    Top Latches DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    No data exists for this section of the report.
    Activity Over Time DB/Inst: ORCL/orcl (Oct 04 16:44 to 16:59)
    -> Analysis period is divided into smaller time slots
    -> Top 3 events are reported in each of those slots
    -> 'Slot Count' shows the number of ASH samples in that slot
    -> 'Event Count' shows the number of ASH samples waiting for
    that event in that slot
    -> '% Event' is 'Event Count' over all ASH samples in the analysis period
    Slot Event
    Slot Time (Duration) Count Event Count % Event
    16:48:00 (2.0 min) 1 CPU + Wait for CPU 1 33.33
    16:50:00 (2.0 min) 2 null event 1 33.33
    os thread startup 1 33.33
    End of Report
    Regards Benk

  • How to resolve most of the Oracle SQL , PL/SQL Performance issues with help of quick Checklist/guidelines ?

    Please go thru below important checklist/guidelines to identify issue in any Perforamnce issue and resolution in no time.
    Checklist for Quick Performance  problem Resolution
    ·         get trace, code and other information for given PE case
              - Latest Code from Production env
              - Trace (sql queries, statistics, row source operations with row count, explain plan, all wait events)
              - Program parameters & their frequently used values
              - Run Frequency of the program
              - existing Run-time/response time in Production
              - Business Purpose
    ·         Identify most time consuming SQL taking more than 60 % of program time using Trace & Code analysis
    ·         Check all mandatory parameters/bind variables are directly mapped to index columns of large transaction tables without any functions
    ·         Identify most time consuming operation(s) using Row Source Operation section
    ·         Study program parameter input directly mapped to SQL
    ·         Identify all Input bind parameters being used to SQL
    ·         Is SQL query returning large records for given inputs
    ·         what are the large tables and their respective columns being used to mapped with input parameters
    ·         which operation is scanning highest number of records in Row Source operation/Explain Plan
    ·         Is Oracle Cost Based Optimizer using right Driving table for given SQL ?
    ·         Check the time consuming index on large table and measure Index Selectivity
    ·         Study Where clause for input parameters mapped to tables and their columns to find the correct/optimal usage of index
    ·         Is correct index being used for all large tables?
    ·         Is there any Full Table Scan on Large tables ?
    ·         Is there any unwanted Table being used in SQL ?
    ·         Evaluate Join condition on Large tables and their columns
    ·         Is FTS on large table b'cos of usage of non index columns
    ·         Is there any implicit or explicit conversion causing index not getting used ?
    ·         Statistics of all large tables are upto date ?
    Quick Resolution tips
    1) Use Bulk Processing feature BULK COLLECT with LIMIT and FOR ALL for DML instead of row by row processing
    2) Use Data Caching Technique/Options to cache static data
    3) Use Pipe Line Table Functions whenever possible
    4) Use Global Temporary Table, Materialized view to process complex records
    5) Try avoiding multiple network trips for every row between two database using dblink, Use Global temporary table or set operator to reduce network trip
    6) Use EXTERNAL Table to build interface rather then creating custom table and program to Load and validate the data
    7) Understand Oracle's Cost based Optimizer and Tune most expensive SQL queries with help of Explain plan
    8) Follow Oracle PL/SQL Best Practices
    9) Review tables and their indexes being used in the SQL queries and avoid unnecessary Table scanning
    10) Avoid costly Full Table Scan on Big Transaction tables with Huge data volume,
    11) Use appropriate filtration condition on index columns of seeded Oracle tables directly mapped to program parameters
    12) Review Join condition on existing query explain plan
    13) Use Oracle hint to guide Oracle Cost based optimizer to choose best plan for your custom queries
    14) Avoid applying SQL functions on index columns
    15) Use appropriate hint to guide Oracle CBO to choose best plan to reduce response time
    Thanks
    Praful

    I understand you were trying to post something helpful to people, but sorry, this list is appalling.
    1) Use Bulk Processing feature BULK COLLECT with LIMIT and FOR ALL for DML instead of row by row processing
    No, use pure SQL.
    2) Use Data Caching Technique/Options to cache static data
    No, use pure SQL, and the database and operating system will handle caching.
    3) Use Pipe Line Table Functions whenever possible
    No, use pure SQL
    4) Use Global Temporary Table, Materialized view to process complex records
    No, use pure SQL
    5) Try avoiding multiple network trips for every row between two database using dblink, Use Global temporary table or set operator to reduce network trip
    No, use pure SQL
    6) Use EXTERNAL Table to build interface rather then creating custom table and program to Load and validate the data
    Makes no sense.
    7) Understand Oracle's Cost based Optimizer and Tune most expensive SQL queries with help of Explain plan
    What about using the execution trace?
    8) Follow Oracle PL/SQL Best Practices
    Which are?
    9) Review tables and their indexes being used in the SQL queries and avoid unnecessary Table scanning
    You mean design your database and queries properly?  And table scanning is not always bad.
    10) Avoid costly Full Table Scan on Big Transaction tables with Huge data volume,
    It depends if that is necessary or not.
    11) Use appropriate filtration condition on index columns of seeded Oracle tables directly mapped to program parameters
    No, consider that too many indexes can have an impact on overall performance and can prevent the CBO from picking the best plan.  There's far more to creating indexes than just picking every column that people are likely to search on; you have to consider the cardinality and selectivity of data, as well as the volumes of data being searched and the most common search requirements.
    12) Review Join condition on existing query explain plan
    Well, if you don't have your join conditions right then your query won't work, so that's obvious.
    13) Use Oracle hint to guide Oracle Cost based optimizer to choose best plan for your custom queries
    No.  Oracle recommends you do not use hints for query optimization (it says so in the documentation).  Only certain hints such as APPEND etc. which are more related to certain operations such as inserting data etc. are acceptable in general.  Oracle recommends you use the query optimization tools to help optimize your queries rather than use hints.
    14) Avoid applying SQL functions on index columns
    Why?  If there's a need for a function based index, then it should be used.
    15) Use appropriate hint to guide Oracle CBO to choose best plan to reduce response time
    See 13.
    In short, there are no silver bullets for dealing with performance.  Each situation is different and needs to be evaluated on its own merits.

  • WS Deployed on Oracle App. Server, Response issue for some requests

    I have a WebService deployed on Oracle Application Server and it's an HTTPS connection. I am facing a small issue and can't figure out
    what's going on. For each request there's a response to let know the requester that data is processed successfully or any error. The caller of the Web Service makes a call and sends in an xml data. It's processed on my end and then a return type is an Element that holds the response data. Once in a while the caller does not get a response back even though on my end I don't see any errors and the data is processed successfully.
    Here is the signature of the function:
    public Element parseData(Element data) throws RemoteException; And I can't figure out why a caller does not receive a response back for some of the requests. Any ideas what might be wrong or how to debug this issue. Is there any place I can look to see for each request is a response going back. I don't know if it's related Network or App. Server and where to start to debug this.
    Any help is appreciated.
    Thanks

    Hi there.
    If the consumer of your service is a .NET client, it will silently ignore any response that is not formated as advertize in the WSDL and will return an empty object to the caller. As you are trying to use raw XML Element, it seams to be a potential explaination.
    If you can turn off SSL in your test environment, you may want to run a TCP sniffer tool, on the host where your server is running. You can setup a filter on the IP @ of the consumer or the TCP port of the service comsumed. I'll recommend you look at ethereal as it's a cross platform utility.
    Hope it helps.
    Eric

  • Synonym required to query custom tables from Oracle Apps database

    When creating an Answers report that queries from a table owned by a custom schema (not APPS), I cannot query the table and get error. Once I add a Synonym for APPS on this table, then OBIEE works normally.
    Since OBIEE is logging in as the APPS user, are Synonyms required for all non-APPS tables/views or is there another feature of the Administrator where it can store the table alias (our typical owner of custom tables is XXFI)?

    OK, well "compare" can mean one of two things: Compare structure or compare contents. Here is a quick script to compare column structures of two tables on one database:
    (select COLUMN_NAME,
    DATA_TYPE,
    DATA_LENGTH,
    DATA_PRECISION,
    DATA_SCALE,
    NULLABLE
    from dba_Tab_columns
    where owner=:OWNR1
    and table_name = :tablename
    minus
    select COLUMN_NAME,
    DATA_TYPE,
    DATA_LENGTH,
    DATA_PRECISION,
    DATA_SCALE,
    NULLABLE
    from dba_Tab_columns
    where owner=:ownr2
    and table_name = :tablename)
    union all
    (select COLUMN_NAME,
    DATA_TYPE,
    DATA_LENGTH,
    DATA_PRECISION,
    DATA_SCALE,
    NULLABLE
    from dba_Tab_columns
    where owner=:ownr2
    and table_name = :tablename
    minus
    select COLUMN_NAME,
    DATA_TYPE,
    DATA_LENGTH,
    DATA_PRECISION,
    DATA_SCALE,
    NULLABLE
    from dba_Tab_columns
    where owner=:ownr1
    and table_name = :tablename)
    If this query returns any rows, then these indicate that there are structural differences between the tables. We do a minus in both directions to ensure that an additional column in either schema will be returned in the query.
    If you are going across dblinks to remote tables then you have to amend the "dba_tab_columns" to "sys.dba_tab_columns@yourdblink"
    If you also want to compare indexes, triggers, etc then do the same sort of thing for the associated dba_ views for those objects.
    And if you want to compare table contents, then often the fastest way is also to check minuses in both directions, if the tables are not too big:
    e.g.
    (select * from schema1.table@dblink1
    minus
    select * from schema2.table@dblink2)
    union all
    (select * from schema2.table@dblink2
    minus
    select * from schema1.table@dblink1)
    local tables, of course, simply omit the "@dblink" issues.
    There are tools to help in such things. TOAD, for example, has a pretty good schema comparison tool, and there are plenty of other options out there. But if you need to script this yourself then the logic I've shown is a good starting point.
    Cheers,
    Mike

  • Oracle 9i database software upgrade issue with standby (Dataguard) database

    All,
    I have encountered a problem in upgrading the Oracle software on 2 of our databases in-house. The databases in question here are using Dataguard, so one is set to primary and one is set to standby. The version that they are/were starting at is 9.2.0.7.
    I was able to install and complete the 9.2.0.8 and 9.2.0.8 OCT CPU patches on the primary database, however when I moved to working on the standby database, I am got an error when attempting to follow step 6 of the post installation tasks:
    Enter the following SQL*Plus commands:
    SQL> STARTUP MIGRATE
    SQL> SPOOL patch.log
    SQL> @ORACLE_BASE\ORACLE_HOME\rdbms\admin\catpatch.sql
    SQL> SPOOL OFF
    (Please note that this is my first experience with working with Dataguard, so my methods in updating might have been a bit off-base.)
    If I attempt to startup the standby database in MIGRADE mode, I get the following error (which is expected):
    ORA-01666: controlfile is for a standby database
    If I try to run the catpatch.sql script, even after mounting the database in standby mode, I receive the additional error:
    ORA-01219: database not open: queries allowed on fixed tables/views only
    My question here is, how do I get the standby database to allow me to finish the software upgrade process, so that both versions of Oracle match on the primary and standby databases? :)
    (I did try some research on Metalink, but only found an article detailing how to upgrade your dataguard database from 9i to 10g, which did not look especially helpful in this case...)
    Thank you!
    -Sabrina

    Depending on the kind of Dataguard you currently have, physical or logical, there are two useful metalinks notes:
    Upgrading to 10g with a Physical Standby in Place
         Doc ID:      Note:278521.1
    Upgrading to 10g with a Logical Standby in Place
         Doc ID:      Note:278108.1
    ~ Madrid
    http://hrivera99.blogspot.com/

  • BW database related performance issue..

    hi...
    can anyone tell me how to improve the database performance??

    Hi Nisha,
    Welcome to SDN )
    The Database performance increases by doing the following,
    1. Partioning
    2. Indexing
    3. Archive unwanted data
    4. Creating Aggregates ( Consider DB ratio and KPI ratio)
    5. Check for parallelization options
    6. deleting PSA requests
    Please the below links
    check this
    http://help.sap.com/bp_biv235/BI_EN/documentation/Multi-dimensional_modeling_EN.doc
    and
    Business Intelligence Performance Tuning [original link is broken]
    Thanks,
    Sudhakar.
    Saying thanks in SDN == Assigning Points.

Maybe you are looking for

  • EJB Client and classpath

    Dear all friends, As you probably know if EJB client wants to connect to EJB jar file that deployed in JServer we have to add that .jar file in our classpath otherwise EJB client throws exception.it seems to be owfull that for each .jar file that has

  • Edit one text and update others...

    Software involved:  Adobe InDesign CS5 and Microsoft Word 2010 I have a word document that I'm converting to an InDesign document.  It is a contract paper that's 13 pages long.  There is a place at the top to put the person's name.  It's a template f

  • Multiple chkboxes submit..help!

    Hi, I have a dynamically generated tree of checkboxes which I need to submit to a JSP for performing a search.. Can anyone pl suggest as to how I can submit the values of all the checkboxes which are checked?? TIA, sands

  • Creating a custom panel for Snapping....??

    I'm trying to create a custom panel in Flash CS3 that will allow me to turn snapping on and off (i.e. 'Snap Align', 'Snap to Grid', 'Snap to Guides', etc.) by just checking or unchecking some checkboxes, but I can't find any JSFL code that will chang

  • Bibliothèque iphoto sur mac et CS5

    Ceux qui veulent utiliser les photos de la bibliothèque iphoto avec les logiciels adobe, il faut afficher le contenue du paquet iphoto et faire un alias du dossier master et transférer l'alias hors d'iphoto en le mettant ans le dossier image par exem