Huge load on Database server

Hello all,
Good day. We are using oracle standard edtion 9.2.0. We have sun machine with dual cpu and have 8G memory. We use decicated server for our web applications. We also use SharePlex for replication. The average number of logins (users) in the system is around 25-30. Suddenly we are facing tremendous load avarage on db server - highest is 20 that means 20*100(%).
In our daily load avarage statistics, we always find some peaks telling that most of the times, the load average remain very high. We have tuned our expensive queries that we found by sp report. For your kind information, the most expensive query do not take more than 2 sec when it runs individually. During a heavy load on db server, even a simple query (usually takes 200 ms) is found as a guilty party.
We have set a large value (1024M) for our pga_aggregate_target parameter for testing some queries that were using disk sort operations. Has it an impact?
Someone of us, run a 20 minute sampling of system statistics when the system was pegged at a load average of 20, and then flushed the shared pool which forced the existing sql to get re-parsed on the next execution. The load average on the system started dropping within seconds of doing this, and then went all the way down to a level under 1 in a matter of minutes.
What more we can do? Waiting for your helping hands :)
Here I'm providing some parameters for your kind considerations.
db_block_size 8192
db_cache_size 1073741824
db_file_multiblock_read_count 16
db_keep_cache_size 117440512
db_recycle_cache_size 117440512
fast_start_mttr_target 0
hash_join_enabled TRUE
java_pool_size 33554432
job_queue_processes 10
large_pool_size 16777216
log_buffer 1048576
open_cursors 300
optimizer_index_caching 90
optimizer_index_cost_adj 50
pga_aggregate_target 1073741824
processes 400
query_rewrite_enabled FALSE
session_cached_cursors 100
sga_max_size 1897893792
shared_pool_size 218103808
sort_area_size 524288
Thanks for your time and considerations.
Regards,
Shawon

Thank you very much Pierre Forstmann! Here is more information that u asked-
1. We already used statspack. But when the load avg went high- we did not run it.
2. We found the load avg using "top"
3. The users are complaining that the system is slow for few days. But suddenly it went over 20 and nobody can use the system.
4. In db server, we just run oracle and SharePlex. We have another host for web server. No other unix process are taking the CPUs, only oracle is taking it.
5. We can not produce the same load by running queries in sqlplus.
6. We did not try SQL tracing for one user connection and analyze with tkprof.
Someone of us ran a 20 minute sampling of system statistics when the system was pegged at a load average of 20, and then flushed the shared pool which forced the existing sql to get re-parsed on the next execution. The load average on the system started dropping within seconds of doing this, and then went all the way down to a level under 1 in a matter of minutes.
How did it help to improve the load avg? Since then the load avg is below 2. But we are not sure whether the problem is solved or not. Could you please tell me what is happening with oracle?
Thanking you.
Regards,
Shawon

Similar Messages

  • How to monitor load on my database server ?

    Hi Experts,
    Could anybody please suggest me a way to monitor the load on my database server?
    What perfmon counters should I monitor on daily basis and confirm that load on the server is increasing day by day [or] by week by week [or] every month. Also, suggest me if we can achieve this using DMV's so that we can store this info in a table using a SQL
    agent job. Because a sql guy I would be more happy working with TSQL queries.
    Do all this helps me, to put for forward my argument to the management that this server can handle this much of load based on data and incase if the load is increased on the server , we might need more resources ( like cpu, better storage , more RAM etc ...)
    or move to a new server which has more power.
    Also, please suggest any load testing tools so that I can test this on testing environment.
    Thanks in Advance.

    About Perfmon counter, 
    You may refer the below perfmon counter pdf file. You may need to look at all memory counters to confirm the issues. 
    http://www.quest.com/backstage/images/promotions/SQLServer-Perfmonance-Poster.pdf
    Regarding
    your request on workload, you need to do a Capacity planning of your servers first. 
    Please refer the below link, this would explain complete info:
    http://sqlactions.com/2012/05/15/collection-and-reporting-of-perfmon-data-for-sql-server-capacity-planning-and-trend-analysis/
    To find the SQL Server Usage and monitoring server, you can even think of
    third party tools like Spotlight/Foglight etc...
    http://www.quest.com/foglight-for-sql-server/

  • How can we transfer huge amount of data from database server to xml format

    hi guru
    how can we transfer huge amount of data from database server to xml format.
    regards
    subhasis.

    Create ABAP coding
    At first we create the internal table TYPES and DATA definition, we want to fill with the XML data. I have declared the table "it_airplus" like the structure from XML file definition for a better overview, because it is a long XML Definition (see the XSD file in the sample ZIP container by airplus.com)
    *the declaration
    TYPES: BEGIN OF t_sum_vat_sum,
              a_rate(5),
              net_value(15),
              vat_value(15),
             END OF t_sum_vat_sum.
    TYPES: BEGIN OF t_sum_total_sale,
            a_currency(3),
            net_total(15),
            vat_total(15),
            vat_sum TYPE REF TO t_sum_vat_sum,
           END OF t_sum_total_sale.
    TYPES: BEGIN OF t_sum_total_bill,
            net_total(15),
            vat_total(15),
            vat_sum TYPE t_sum_vat_sum,
            add_ins_val(15),
            total_bill_amount(15),
           END OF t_sum_total_bill.TYPES: BEGIN OF t_ap_summary,
            a_num_inv_det(5),
            total_sale_values TYPE t_sum_total_sale,
            total_bill_values TYPE t_sum_total_bill,
           END OF t_ap_summary.TYPES: BEGIN OF t_ap,
            head    TYPE t_ap_head,
            details TYPE t_ap_details,
            summary TYPE t_ap_summary,
           END OF t_ap.DATA: it_airplus TYPE STANDARD TABLE OF t_ap
    *call the transformation
    CALL TRANSFORMATION ZFI_AIRPLUS
         SOURCE xml l_xml_x1
         RESULT xml_output = it_airplus
         .see the complete report: Read data from XML file via XSLT program
    Create XSLT program
    There are two options to create a XSLT program:
    Tcode: SE80 -> create/choose packet -> right click on it | Create -> Others -> XSL Transformation
    Tcode: XSLT_TOOL
    For a quick overview you can watch at the SXSLTDEMO* programs.
    In this example we already use the three XSLT options explained later.
    As you can see we define a XSL and ASX (ABAP) tags to handle the ABAP and XML variables/tags. After "

  • Load BW with just database server in R3.

    Hi All,
    We have added two new fields to our 2LIS_13_VDKON datasource in our R3 systems.  When we transport, we need a quiet system.  The LBWQ queue and RSA7 queues need to be empty or our data will become corrupt. 
    In our production systems, we have a maintenance window each weekend during which the transports are imported.  I want to use this as my quiet time, problem is that during this maintenance window, only the database server is running in R3.  The application servers are not available during this window. 
    Am I able to load data into bw (to clean out RSA7) with just a database server running in R3?
    I don't think I can, but I wanted to post the question anyways.
    We are on version 3.5.
    Thanks
    Charla

    Do they take the app server on DBCI down too? When transports are being imported, is it done via STMS in your source system or is it done via a script? If it's done manually via tcode STMS, or an automated process (scheduled standard or custom ABAP program) within the SAP application, an app server must be available.

  • High Database server load from expensive SQL statements

    Dear all,
    I am facing problem in the production  sever there is high Database server load from expensive SQL statements as per EVA report
    Buffer Load [%]     Disk Load [%]     CPU Load [%]
    55     69     0
    Analysis of DB SQL CACHE
    EXPENSIVE SQL STATEMENTS OVERVIEW
    Object Name     CPU Load [%]     I/O Load [%]     Elapsed Time [%]     Executions     Records Processed
    BSIS        1     9     0     22     90462
    CDEF$     6     6     0     2131113     2575694
    BSAD     1     3     0     21     408576
    MKPF     1     32     0     180     3899
    ICOL$     9     1     0     2575694     8703798
    OBJ$     6     1     0     3405254     3400023
    COL$     12     0     0     2138793     22919657
    MKPF     1     13     0     75     396
    MCHB     7     0     0     366543     41708
    Please suggest the step by step to reduce  the  expensive SQL  from these table.
    Regards

    Hi,
    In tx code ST05 i have find the following deatails.
    Duration |Obj. name |Op.    |Recs.|RC    |Statement  
         4 TSP03A     REOPEN             0 SELECT WHERE "NAME" = 'LOCL' AND "P" = 'S_CLIENTS'                                                                                344 TSP03A     FETCH       1      0                                                                                42 ZRIN    DECLARE            0 SELECT WHERE "MANDT" = :A0 AND "BCQ" = :A1                                                                                442 ZRIN    PREPARE            0 SELECT WHERE "MANDT" = :A0 AND "BCQ" = :A1                                                                                3 ZRIN    OPEN               0 SELECT WHERE "MANDT" = '600' AND "BCQ" = 'BCQ'                                                                               
    2,181,565 ZRIN  FETCH       4   1403                                                                               
    706,267 DBA_SEGME FETCH      99      0                                                                                8,248 DBA_SEGME FETCH      99      0                                                                                44,994 DBA_SEGME FETCH      99      0                                                                                67,713 DBA_SEGME FETCH      99      0   
    1,367,923 DBA_SEGME FETCH      99      0
      16,735|TADIR     |FETCH  |    1|  1403|                                                                                |
    5
    TADIR
    REOPEN
    0
    SELECT WHERE "PGMID" = 'R3TR' AND "OBJECT" = 'TABL' AND "OBJ_NAME" = 'CME__TEXT'
    22,415
    TADIR
    FETCH
    1
    1403
    4
    TADIR
    REOPEN
    0
    SELECT WHERE "PGMID" = 'R3TR' AND "OBJECT" = 'TABL' AND "OBJ_NAME" = 'CME_PATTERN_SIGN'
    232
    TADIR
    FETCH
    1
    1403
    4
    TADIR
    REOPEN
    0
    SELECT WHERE "PGMID" = 'R3TR' AND "OBJECT" = 'TABL' AND "OBJ_NAME" = 'CML_ARC_DEADLINE'
    19,189
    TADIR
    FETCH
    1
    1403
    4
    TADIR
    REOPEN
    706,267
    DBA_SEGME
    FETCH
    99
    0
    |
    44,994
    DBA_SEGME
    FETCH
    99
    0
    67,713
    DBA_SEGME
    FETCH
    99
    0
    Please suggest how to  reduce it.
    Regards,

  • How can we position 5 Database Server for a centralized Database ?

    1) We are providing a e governance solution for an organization,where we are providing a  centralized database,Client have provided 5 Database server for the same.how can we position the Database Server? there are 5000 Concurrent users and 25000 users,SAN
    Storage for approx. 60 TB,Database size of 2 TB and growth of 1 TB every year
    2) How many instance can we have for above said Case?
    3) How much RAM Required ?

    Thanks for sharing information!!
    DB size is going to be too huge, you need to check for data archival option along with proper compression option.
    As per requirement, you need good server setup like; CPU: 4 cpu, 12 cores and RAM will be good to start with 256 GB.
    When you say 5 db server, is that it means 5 physical server not db instances rt?
    If that is the case then you can go for SQL Server 2014 to use that in memory options along with so many good support for temppdb, query optimization due to cardinality changes, SSD support etc.
    Intances will be:
    INS1, 2 : main instance and archival instance, using first two servers out of 5 servers.
    INS3: DR server, if it will not be active all the time and then we can easily check for 1/4 load on DR server so you can configure AlwaysOn to check INS1 and DR instance to get all HA DR part too.
    In case of INS4, INS5 can be used for reporting purpose being configured as other replication to act as local DR too.
    Hopefully I have covered all the available objects you have.
    Santosh Singh

  • Performance issue with database server

    Hello,
    We have installed ECC 6.0 on HP UX 11.23 OS with Oracle as database.
    We have cluster environment in our production severs.
    One server as DB-CI installed and other as Dialogue Instance installed.
    After installation, the dialogue instance (Application server))works fine but if we login into DB server and execute any t-code It takes too much time or just stuck.
    and after that if I go to application and execute that t-code it works fine and after that if I go to database and execute that t-code now even in DB it works fine.
    Kindly guide me.
    Thanks,
    Chandresh Pranami.

    Hi Chandresh,
    Check below things for your Database server.
    SM50 - Workprocess status busy or wait
    ST04 - Data buffer hit ratio
    SM37 - Any background job running there.
    Normally even if you are login through a diffrent app server the database side execution will happen in db server only.Application server will do the execution but the data read and to be updated will happen in db server only.When the sql request will go to db server may select update..types of sql then it willalways will be processed by db server.
    So there will be lot of load already thereIg you execute your work here it will take more time because more load for the application will be there.
    Regards
    Ashok

  • ORA-12709: error while loading create database character set

    I installed Oracle 8.05 on Linux successfully: was able to login
    whith SQLPlus, start and stop the db whith svrmgrl etc.
    During this install I chose WE8ISO8859P9 as the database
    characterset when prompted.
    After that I installed Oracle Application Server 3.02, and now
    I'm getting the
    ORA-12709: error while loading create database character set
    message when I try to start up the database, and the database
    won't mount.
    Platform is RedHat Linux 5.2.
    NLS_LANG set to different settings,
    e.g. AMERICAN_AMERICA.WE8ISO8859P9
    but without success.
    Anyone any clue?
    Thanks!
    null

    Jogchum Reitsma (guest) wrote:
    : I installed Oracle 8.05 on Linux successfully: was able to
    login
    : whith SQLPlus, start and stop the db whith svrmgrl etc.
    : During this install I chose WE8ISO8859P9 as the database
    : characterset when prompted.
    : After that I installed Oracle Application Server 3.02, and now
    : I'm getting the
    : ORA-12709: error while loading create database character set
    : message when I try to start up the database, and the database
    : won't mount.
    : Platform is RedHat Linux 5.2.
    : NLS_LANG set to different settings,
    : e.g. AMERICAN_AMERICA.WE8ISO8859P9
    : but without success.
    : Anyone any clue?
    : Thanks!
    You can create the database with WE8DEC character set
    and to use the WE8ISO8859P9 on the client or even on Linux.
    The NLS_LANG setting doesn't effect the database, but the
    interface with the database. The same setting can be used in de
    windows 95/98/NT registry.
    null

  • ORA-12709: error while loading create database character set after upgrade

    Dear All
    i m getting ORA-12709: error while loading create database character set, After upgraded the database from 10.2.0.3 to 11.2.0.3 in ebusiness suit env.
    current application version 12.0.6
    please help me to resolve it.
    SQL> startup;
    ORACLE instance started.
    Total System Global Area 1.2831E+10 bytes
    Fixed Size 2171296 bytes
    Variable Size 2650807904 bytes
    Database Buffers 1.0133E+10 bytes
    Redo Buffers 44785664 bytes
    ORA-12709: error while loading create database character set
    -bash-3.00$ echo $ORA_NLS10
    /u01/oracle/PROD/db/teche_st/11.2.0/nls/data/9idata
    export ORACLE_BASE=/u01/oracle
    export ORACLE_HOME=/u01/oracle/PROD/db/tech_st/11.2.0
    export PATH=$ORACLE_HOME/bin:$ORACLE_HOME/perl/bin:$PATH
    export PERL5LIB=$ORACLE_HOME/perl/lib/5.10.0:$ORACLE_HOME/perl/site_perl/5.10.0
    export ORA_NLS10=/u01/oracle/PROD/db/teche_st/11.2.0/nls/data/9idata
    export ORACLE_SID=PROD
    -bash-3.00$ pwd
    /u01/oracle/PROD/db/tech_st/11.2.0/nls/data/9idata
    -bash-3.00$ ls -lh |more
    total 56912
    -rw-r--r-- 1 oracle oinstall 951 Jan 15 16:05 lx00001.nlb
    -rw-r--r-- 1 oracle oinstall 957 Jan 15 16:05 lx00002.nlb
    -rw-r--r-- 1 oracle oinstall 959 Jan 15 16:05 lx00003.nlb
    -rw-r--r-- 1 oracle oinstall 984 Jan 15 16:05 lx00004.nlb
    -rw-r--r-- 1 oracle oinstall 968 Jan 15 16:05 lx00005.nlb
    -rw-r--r-- 1 oracle oinstall 962 Jan 15 16:05 lx00006.nlb
    -rw-r--r-- 1 oracle oinstall 960 Jan 15 16:05 lx00007.nlb
    -rw-r--r-- 1 oracle oinstall 950 Jan 15 16:05 lx00008.nlb
    -rw-r--r-- 1 oracle oinstall 940 Jan 15 16:05 lx00009.nlb
    -rw-r--r-- 1 oracle oinstall 939 Jan 15 16:05 lx0000a.nlb
    -rw-r--r-- 1 oracle oinstall 1006 Jan 15 16:05 lx0000b.nlb
    -rw-r--r-- 1 oracle oinstall 1008 Jan 15 16:05 lx0000c.nlb
    -rw-r--r-- 1 oracle oinstall 998 Jan 15 16:05 lx0000d.nlb
    -rw-r--r-- 1 oracle oinstall 1005 Jan 15 16:05 lx0000e.nlb
    -rw-r--r-- 1 oracle oinstall 926 Jan 15 16:05 lx0000f.nlb
    -rw-r--r-- 1 oracle oinstall 1.0K Jan 15 16:05 lx00010.nlb
    -rw-r--r-- 1 oracle oinstall 958 Jan 15 16:05 lx00011.nlb
    -rw-r--r-- 1 oracle oinstall 956 Jan 15 16:05 lx00012.nlb
    -rw-r--r-- 1 oracle oinstall 1005 Jan 15 16:05 lx00013.nlb
    -rw-r--r-- 1 oracle oinstall 970 Jan 15 16:05 lx00014.nlb
    -rw-r--r-- 1 oracle oinstall 950 Jan 15 16:05 lx00015.nlb
    -rw-r--r-- 1 oracle oinstall 1.0K Jan 15 16:05 lx00016.nlb
    -rw-r--r-- 1 oracle oinstall 957 Jan 15 16:05 lx00017.nlb
    -rw-r--r-- 1 oracle oinstall 932 Jan 15 16:05 lx00018.nlb
    -rw-r--r-- 1 oracle oinstall 932 Jan 15 16:05 lx00019.nlb
    -rw-r--r-- 1 oracle oinstall 951 Jan 15 16:05 lx0001a.nlb
    -rw-r--r-- 1 oracle oinstall 944 Jan 15 16:05 lx0001b.nlb
    -rw-r--r-- 1 oracle oinstall 953 Jan 15 16:05 lx0001c.nlb
    Starting up:
    Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - 64bit Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options.
    ORACLE_HOME = /u01/oracle/PROD/db/tech_st/11.2.0
    System name: SunOS
    Node name: proddb3.zakathouse.org
    Release: 5.10
    Version: Generic_147440-19
    Machine: sun4u
    Using parameter settings in server-side spfile /u01/oracle/PROD/db/tech_st/11.2.0/dbs/spfilePROD.ora
    System parameters with non-default values:
    processes = 200
    sessions = 400
    timed_statistics = TRUE
    event = ""
    shared_pool_size = 416M
    shared_pool_reserved_size= 40M
    nls_language = "american"
    nls_territory = "america"
    nls_sort = "binary"
    nls_date_format = "DD-MON-RR"
    nls_numeric_characters = ".,"
    nls_comp = "binary"
    nls_length_semantics = "BYTE"
    memory_target = 11G
    memory_max_target = 12G
    control_files = "/u01/oracle/PROD/db/apps_st/data/cntrl01.dbf"
    control_files = "/u01/oracle/PROD/db/tech_st/10.2.0/dbs/cntrl02.dbf"
    control_files = "/u01/oracle/PROD/db/apps_st/data/cntrl03.dbf"
    db_block_checksum = "TRUE"
    db_block_size = 8192
    compatible = "11.2.0.0.0"
    log_archive_dest_1 = "LOCATION=/u01/oracle/PROD/db/apps_st/data/archive"
    log_archive_format = "%t_%s_%r.dbf"
    log_buffer = 14278656
    log_checkpoint_interval = 100000
    log_checkpoint_timeout = 1200
    db_files = 512
    db_file_multiblock_read_count= 8
    db_recovery_file_dest = "/u01/oracle/fast_recovery_area"
    db_recovery_file_dest_size= 14726M
    log_checkpoints_to_alert = TRUE
    dml_locks = 10000
    undo_management = "AUTO"
    undo_tablespace = "APPS_UNDOTS1"
    db_block_checking = "FALSE"
    session_cached_cursors = 500
    utl_file_dir = "/usr/tmp"
    utl_file_dir = "/usr/tmp"
    utl_file_dir = "/u01/oracle/PROD/db/tech_st/10.2.0/appsutil/outbound"
    utl_file_dir = "/u01/oracle/PROD/db/tech_st/10.2.0/appsutil/outbound/PROD_proddb3"
    utl_file_dir = "/usr/tmp"
    plsql_code_type = "INTERPRETED"
    plsql_optimize_level = 2
    job_queue_processes = 2
    cursor_sharing = "EXACT"
    parallel_min_servers = 0
    parallel_max_servers = 8
    core_dump_dest = "/u01/oracle/PROD/db/tech_st/10.2.0/admin/PROD_proddb3/cdump"
    audit_file_dest = "/u01/oracle/admin/PROD/adump"
    db_name = "PROD"
    open_cursors = 600
    pga_aggregate_target = 1G
    workarea_size_policy = "AUTO"
    optimizer_secure_view_merging= FALSE
    aq_tm_processes = 1
    olap_page_pool_size = 4M
    diagnostic_dest = "/u01/oracle"
    max_dump_file_size = "20480"
    Tue Jan 15 16:16:02 2013
    PMON started with pid=2, OS id=18608
    Tue Jan 15 16:16:02 2013
    PSP0 started with pid=3, OS id=18610
    Tue Jan 15 16:16:03 2013
    VKTM started with pid=4, OS id=18612 at elevated priority
    VKTM running at (10)millisec precision with DBRM quantum (100)ms
    Tue Jan 15 16:16:03 2013
    GEN0 started with pid=5, OS id=18616
    Tue Jan 15 16:16:03 2013
    DIAG started with pid=6, OS id=18618
    Tue Jan 15 16:16:03 2013
    DBRM started with pid=7, OS id=18620
    Tue Jan 15 16:16:03 2013
    DIA0 started with pid=8, OS id=18622
    Tue Jan 15 16:16:03 2013
    MMAN started with pid=9, OS id=18624
    Tue Jan 15 16:16:03 2013
    DBW0 started with pid=10, OS id=18626
    Tue Jan 15 16:16:03 2013
    LGWR started with pid=11, OS id=18628
    Tue Jan 15 16:16:03 2013
    CKPT started with pid=12, OS id=18630
    Tue Jan 15 16:16:03 2013
    SMON started with pid=13, OS id=18632
    Tue Jan 15 16:16:04 2013
    RECO started with pid=14, OS id=18634
    Tue Jan 15 16:16:04 2013
    MMON started with pid=15, OS id=18636
    Tue Jan 15 16:16:04 2013
    MMNL started with pid=16, OS id=18638
    DISM started, OS id=18640
    ORACLE_BASE from environment = /u01/oracle
    Tue Jan 15 16:16:08 2013
    ALTER DATABASE MOUNT
    ORA-12709 signalled during: ALTER DATABASE MOUNT...

    ORA-12709 signalled during: ALTER DATABASE MOUNT...Do you have any trace files generated at the time you get this error?
    Please see these docs.
    ORA-12709: WHILE STARTING THE DATABASE [ID 1076156.6]
    Upgrading from 9i to 10gR2 Fails With ORA-12709 : Error While Loading Create Database Character Set [ID 732861.1]
    Ora-12709 While Trying To Start The Database [ID 311035.1]
    ORA-12709 when Mounting the Database [ID 160478.1]
    How to Move From One Database Character Set to Another at the Database Level [ID 1059300.6]
    Thanks,
    Hussein

  • Failed to Establish Connection with SQL Database Server (EAS)

    Hi,
    We have recently installed version 11.1.2.1 of essbase. Everything is working just peachy. One of my users tried to build a load rule using an Oracle ODBC data connection and gets the error:
    Failed to Establish Connection with SQL Database Server.  See log for more information.
    When I look in the log I see this:
    [Tue Feb 28 09:14:43 2012]Local/Sample/Basic/mressler@EnCanaAD/4240/Info(1013091)
    Received Command [SQLListDsn] from user [mressler@EnCanaAD]
    [Tue Feb 28 09:14:43 2012]Local/Sample/Basic/mressler@EnCanaAD/4240/Info(1021020)
    Cannot read SQL driver name for [Backup Exec Catalogs] from [ODBC.INI]
    [Tue Feb 28 09:15:03 2012]Local/Sample/Basic/mressler@EnCanaAD/5228/Info(1013091)
    Received Command [SQLRetrieve] from user [mressler@EnCanaAD]
    [Tue Feb 28 09:15:03 2012]Local/Sample/Basic/mressler@EnCanaAD/5228/Info(1021020)
    Cannot read SQL driver name for [Backup Exec Catalogs] from [ODBC.INI]
    [Tue Feb 28 09:15:03 2012]Local/Sample/Basic/mressler@EnCanaAD/5228/Info(1021004)
    Connection String is generated
    [Tue Feb 28 09:15:03 2012]Local/Sample/Basic/mressler@EnCanaAD/5228/Info(1021041)
    Connection String is [DSN=Test;UID=...;PWD=...;]
    [Tue Feb 28 09:15:03 2012]Local/Sample/Basic/mressler@EnCanaAD/5228/Info(1021006)
    SELECT Statement [SELECT '1' from dual] is generated
    [Tue Feb 28 09:15:03 2012]Local/Sample/Basic/mressler@EnCanaAD/5228/Info(1021013)
    ODBC Layer Error: [IM004] ==> [[Microsoft][ODBC Driver Manager] Driver's SQLAllocHandle on SQL_HANDLE_ENV failed]
    [Tue Feb 28 09:15:03 2012]Local/Sample/Basic/mressler@EnCanaAD/5228/Info(1021014)
    ODBC Layer Error: Native Error code [0]
    [Tue Feb 28 09:15:03 2012]Local/Sample/Basic/mressler@EnCanaAD/5228/Error(1021001)
    Failed to Establish Connection With SQL Database Server. See log for more information
    [Tue Feb 28 09:15:03 2012]Local/Sample/Basic/mressler@EnCanaAD/5228/Warning(1080014)
    Transaction [ 0x140002( 0x4f4cfd87.0x7b890 ) ] aborted due to status [1021001].
    I have tested that the ODBC connection can connect to the database. The sql query has been dumbed down to select '1' from dual. I have not included the word "select" in the query.
    Anyone seen this before? We are in the process of uninstalling the Oracle client completely and re-installing it to see if this solves the issue. We only installed the 64 bit version of the Oracle client. I am running on Windows Enterprise 2008 R2 - 64 bit. EAS, Essbase, EIS are all 64 bit installs. I believe Studio is the only one that is 32bit. However we are not currently using this product.
    Thanks,
    Mike

    Hi Mike,
    With Oracle Support's help, we got 3 of the 5 servers working now. We were using the Datadirect drivers but were configuring them wrong! You have to use the Standard Connection portion in the setup, specifying the Host, Port Number (1521) and the SID. We were using the TNSNames Connection which worked in the past, but apparently does not now.
    So I am down to 2 servers not working. They are both Windows 2008 R2 Enterprise 64 bit, running Oracle EPM Essbase 11.1.2.1. When I Open SQL in EAS it prompts for database (Essbase Server, App, DB) and when I hit ok it pops up with "Network error: Cannot Locate Connect Information For []". We did not do any of the suggestions in the document you linked to get the other 3 servers working. When you hit ok on that error, you then get another error stating, "There are no Data Sources defined. 'SQL data sources' option will be disabled". We do have datasources defined on the server.
    Any ideas?
    Thanks,
    Mike
    Edited by: Mike on Mar 1, 2012 8:38 AM

  • Essbase Studio - Failed to Establish Connection With SQL Database Server

    Hi all,
    I am new to Hyperion and am having trouble deploying what seems a simple cube in Essbase Studio.
    My environment is Windows 2003, EPM 11.1.1.2, SQL Server 2000.
    I have the following two issues which may be related.
    1. The EPM System Diagnostic tool says that Hyperion Foundation cannot connect to the SQL database.
    Error:
    Failed: Connection to database
    Error: java.net.UnknownHostException: <server name: <server name>
    Recommended Action:
    Every other EPM application is able to connect to the database. I have tried re-configuring Foundation Services and checking the config files and nothing looks wrong. I would appreciate advice on how to fix this.
    2 In Essbase Studio, I was able to connect to the database where the source data is, build the minischema and create dimensions and measures. But when I run the cube deployment wizard I get the error:
    Message: Failed to deploy Essbase cube
    Caused By: Failed to build Essbase cube dimension: (Time)
    Caused By: Cannot incremental build. Essbase Error(1021001): Failed to Establish Connection With SQL Database Server. See log for more information
    …ODBC Layer Error: [08001] ==> [[Microsoft][ODBC SQL Server Driver][TCP/IP Sockets]SQL Server does not exist or access denied.]
    I have checked the DSN as well as the connections set up within Studio, and they are all able to connect to the database. I am using the admin user for the Essbase server and have created a different user for the databases where the data resides and the shared services database. I tried using the same user all the way through and this didn’t help.
    Can you please advise what else I can check or change to resolve this issue.
    Also, can you work in EAS with cubes that are created in Studio? I tried looking at it and I got an error that the rule file couldn’t be accessed because it was created in Studio.
    After the above failures, I tried creating a cube in EIS, with mixed success. I have been able to load the members and data without errors. But when I try to view the data in the cube, I am seeing field names for the measures instead of the data itself.
    ...Definitely more questions than answers at this stage of my learning.
    Regards
    Michelle

    Michelle,
    I don't know if you found an answer to you question after so many months but I was hoping I could be helpful.
    The issue you are experience happens often when the dimension that you are getting an error on, in this case TIME, is built from a snowflake lineage and there is a bad foreign key reference. This dimension is most like high-up in your outline build process for Essbase Studio and this prevents the build from happening usually early on.
    Check your logs also. They are in Hyperion > Products > logs > essbase > essbasestudio.
    You can also make your logs more verbose by setting a configuration variable in the essbase studio server.properties file but that should only be used for debugging as it really saps performance.
    And, yes, you can of course edit an Essbase Studio deployed cube in EAS. However, any changes you make to the cube in EAS are subject to being wiped-out upon the next Essbase Studio deployment of that Applicaion/Database combo.
    If you want to provide more detail, screenshots, etc. I would be glad to help where I can.
    Cheers,
    Christian
    http://www.artofbi.com

  • DI job causing high levels of I/O on database server

    We have a DI job that is loading a sql server 2005 database.  When the facts are loaded itu2019s causingu2019s a high level of I/O on the database server causing the DI job to slow down.  No more than 5 facts are loaded concurrently.  The fact dataflows all have a sql transform to run the select query against the DB, a few query transforms to do lookups to get dimension keys, and all do inserts to the target. The DBA says there are too many DB connections open and DI is not closing them.  My thinking was DI would manage the open connections for lookup, etc and would close then properly when the dataflow is complete. 
    Any thoughts on what else would cause high levels of DB I/O?
    Additional Info:
    - Run the DI job, source and target tables are in SQL Server, and it takes 5 hours.
    - Run the same DI job again, on the same data set, and it takes 12+ hours.  This run will have high levels on DB I/O.
    - But if SQL Server is stopped and restart, the job will again take 5 hours the first time it runs.
    Edited by: Chris Sam on Apr 15, 2009 3:43 PM

    There are a lot of areas of a DI Job that can be tuned for performance, but given the fact that your job runs fine after the database is restarted, it sounds like a problem with the database server and not the Data Integrator job.
    There are a lot of resources out there for dealing with SQL Server disk IO bottlenecks.  As a minimum first step all of them will recommend putting your .mdf and .ldf files on seperate drives and using Raid 10 for the .mdf file.

  • Load Balancing FTP Server thru CSM using a single Client IP

    Hello,
    We have a need to load balance 3 FTP servers. These servers are reached only from a single client IP which is a database server. The FTP method that is being used is currently passive. Our configuration is currently unidirectional, ie, the FTP client (the one database server) sends to the VIP and the FTP Servers then talk directly back to the FTP client and the traffic does not go back through the CSM. The problem is that because FTP negotiates another port to talk on, we have to use sticky so that the connection is sent back to the original FTP server that sent the FTP data port to talk on. But, since we only have a single client IP that is ever used we are not load balancing appropriately across the FTP servers.
    Traffic flow goes something like this, tcp port followed after colon as an example
    1. FTP Client ----> VIP:21
    2. CSM ---------> FTP Server:21
    3. FTP Server --------> FTP Client(FTP server says come talk to me on port 1700)
    4. FTP Client ---------> VIP:1700
    5. CSM ---------> FTP Server:1700
    6. FTP Server:1700 ---------> FTP Client
    repeat steps 4 thru 6
    Here's our hardware and software:
    WS-X6066-SLB-APC running 4.2(2)
    Config is as follows
    module ContentSwitchingModule 9
    ft group 101 vlan 9
    priority 10
    vlan 216 client
    ip address 10.209.16.31 255.255.252.0
    gateway 10.209.16.1
    vlan 20 server
    ip address 10.209.0.31 255.255.252.0
    alias 10.209.0.11 255.255.252.0
    probe ICMP1 icmp
    interval 3
    failed 3
    receive 3
    serverfarm FHEPRT
    no nat server
    no nat client
    real 10.209.0.72
    inservice
    real 10.209.0.73
    inservice
    real 10.209.0.71
    inservice
    probe ICMP1
    sticky 106 netmask 255.255.255.255 address source timeout 3
    policy FHEPRT_POL1
    sticky-group 106
    serverfarm FHEPRT
    vserver FHEPRT1
    virtual 10.209.16.71 any
    vlan 216
    unidirectional
    serverfarm FHEPRT
    replicate csrp connection
    no persistent rebalance
    slb-policy FHEPRT_POL1
    inservice

    You are missing "service ftp" config in the Vip definition. Try the following
    vserver FHEPRT1
    virtual 10.209.16.71 tcp ftp service ftp
    Syed Iftekhar Ahmed

  • Connection not closed in the database server end

    Hi,
    Our database server (sql server 2000) and java application are running in two different boxes. Because the java application sent to the db server quite a few queries that can not be completed (applications hangs there) due to huge amount of data (18 million records in one table and 1 million records in another table and there are joins among the three tables), we stopped the java program. However, in the db server end, the connections associated with these problem queries were not closed although the java app end already closed these connections by calling connection.close(). When this happened, we have to manually close the connection from sql server enterprise manager.
    Here are the questions.
    1)what could be the reason why db server does not close the connection?
    2)Is there any solution in the java end that could close the connection in the db server end when this situation happens? I say this situation happens, I mean in most cases when java application closes connection, db server closes connections in its end too.
    Thanks
    Mark

    mark.lin wrote:
    Thanks for your answers and questions.
    Here are the answers to your questions
    1)The jdbc driver is jturboJTurbo? Never heard of it. You mean this?
    http://www.newatlanta.com/products/jturbo/index.jsp
    Why pay for something when you have two options available that are free? One of them is from Microsoft itself. Whose idea was this?
    2)It's a desk top application.
    It seems to me that it should not be driver issue. I was this happened the Microsoft's own jdbc driver. By jdbc specification, although you don't close the connection explicitly in the program, the connection will be automatically closed in the java app end during garbage collection. Very bad - never depend on this. Always close your ResultSet, Statement, and Connection in reverse order of opening in a finally block, each one wrapped in individual try/catch blocks.
    One should never, ever depend on the GC to take care of resources like database connections or resource handles. Those should be cleaned up by your code.
    I tested this by intentionally not closing a connection. When I stops the application, the connection in the sql server end was also closed.
    But if you don't close the application and fail to close connections, you can see where eventually you'll exhaust the pool.
    It seems to me this is a sql server issue rather than a jdbc driver issue. I need way to prove this.
    Like I suggested, it's easy to swap in the jTDS driver and try it again. What do you have to lose?
    Or call Microsoft support and ask them. Or maybe JTurbo. You're paying for it - might as well use the support.
    %

  • SSMA "Loading to database new table File Name.dbo[Table] ...sql execution failed

    In this case I had already processed this table once before 'no changes' and processed the second time I received this error ...sql execution Failed. It's like the sql server 2008 R2 doesn't see or detect the process. Starting Phase #0 Synchronizing Database:
    then it shows 'Loading to database new table file name.dbo[table] ... sql execution failed (this is my issue)

    Hi TheJudge2,
    Based on my understanding, you have installed SQL Server Migration Assistant successfully then you perform migration. After migration, you can’t find the table which you want to migrate to SQL Server 2008 R2 under the database in your SQL Server Management
    Studio, then you perform migration again. Then error message “Loading to database new table file name.dbo[table] ... sql execution failed” is thrown out. Please correct me if I misunderstand your meaning.
    Firstly, I would like to know whether you perform migration from Access database to SQL Server. I have performed migration from Access database to SQL Server in my lab environment.  Based on my test, I could migrate Access database to SQL Server successfully
    and table existed under the database in SQL Server Management Studio after migration.
    Besides, “sql execution failed” is a general error message. It is hard to find out the cause from this error message. Please check the information under the Output in SQL Server Migration Assistant again. It would be better if you can provide the complete
    output information for our deep analysis.
    Best regards,
    Qiuyun Yu

Maybe you are looking for

  • Inserting an image from a CSV file

    I am trying to create a document that is a directory of members of a community.  The directory is in a CSV file.  One ot the fields is a @photos field that contains the path to the location of the images on my local compiuter,  When I try and preview

  • Websites not loading correctly, clock date wrong

    Yesterday I noticed the clock on my computer was 3 hours behind and the date 4 years ahead. Since then any attachments in Gmail come up with a warning as it not being trusted. Also gmail's buttons ( delete, spam, move etc) do not load the words. Then

  • Need Sorter java  file   for sorting the records in webdynpro java

    Hi  Gurus    I  had urgent requirement please give your inputs    In my webdynpro java i need to sort entire records in table    I now we need One sorter java class please provide that java file    please suggest the all the steps how  to do Thanks r

  • How to tell Firefox "Okay for THIS site to redirect"?

    Websites for huge, bureaucratic corporations sometimes try to redirect the browser to other pages. How do I tell Firefox that this is okay -- for a particular website?

  • Moving from trial version to purchasing a full version

    I installed a trial version of an app (Stuff) on my PlayBook. I decided I wanted to purchase it after the trial period was over. App World will only let me delete it and reinstall the trial version. Does anyone have a suggestion ?