Workaround for ORA-04030 error

hi,
we've encountered the following error:
ORA-04030: out of process memory when trying to allocate 296 bytes (callheap, koposal dvoid)
all of the research i've done points to setting up the database with a Very Large Memory configuration, which will be carried out at a later date as we're still in the development stage.
is there a possible workaround for this error?
thanks
santosh sewlal

hi,
we're inserting into tables - spatial data and images, from a C#.net application that sends data to the db in a BLOB

Similar Messages

  • ORA-04030 error

    Database 8i
    OS wndows 2000 server
    RAM 1.5 GB
    SGA 1136617500 bytes
    getting error ORA-04030: Out of process memory when running dbms_stats.gather_schema_stats procedure.
    after reducing e size of sort_area_size from 40 M to 10 M above procedure has been successfully completed but problem is this sort_area_size is not suitable
    for my normal batch processing How to decide about optimum sort_area_size?
    .

    In the init.ora
    PARALLEL_MAX_SERVERS specifies the maximum number of parallel execution processes and parallel recovery processes for an instance. As demand increases, Oracle increases the number of processes from the number created at instance startup up to this value.
    If you set this parameter too low, some queries may not have a parallel execution process available to them during query processing. If you set it too high, memory resource shortages may occur during peak periods, which can degrade performance.

  • Nested tables and memory usage (ORA-04030 error)

    Dear All,
    I have table with approximately 5,000,000 records
    and try to Bulk Collect part of it into nested table in PL/SQL, the code is bellow
    Declare
         Type TcardRec Is Record(
              serno Pls_Integer,
              numberx Char(16),
              caccserno Pls_Integer
         Type TcardList Is Table Of TcardRec;
         fcardInfo TcardList;
    Begin
    Select c.serno, substr(c.numberx,1,16), c.caccserno
    Bulk Collect Into fcardinfo
    From cardx c;
    End;
    After reading approx. 80% it fails with error
    ORA-04030: out of process memory when trying to allocate 16396 bytes (koh-kghu call ,pmucalm coll)
    I 2G memory, is it realy no enough?
    How can I tune memory usage for collection?
    How can I estimate the maximum size of the collection the will fit into memory?
    Thank you in advance for any help
    Artem

    Declare it as a cursor.
    Open the cursor.
    Use fetch bulk collect with the limit option in the loop.
    In your case, you could do like:
    Declare
    Cursor c1 is
    Select c.serno, substr(c.numberx,1,16), c.caccserno
    From cardx c;
    TcardList Is Table Of c1%rowtype;
    fcardInfo TcardList;
    Begin
    Open c1;
    Loop
    Fetch c1 Bulk Collect Into fcardInfo Limit 10000;
    Exit when c1%notfound;
    -- Do some processing here.
    End Loop;
    Close c1;
    End;
    I hope this helps.

  • Need help to avoid ORA-04030 error

    Oracle version : 11.1.0.6.0
    OS - Solaris 64bit Sparc
    Hi All,
    Today we have recieved the below error messeges in the alert log.
    ORA-04030: out of process memory when trying to allocate 56 bytes (callheap,kghscAllocDataElem:dataElem)
    ORA-04030: out of process memory when trying to allocate 184 bytes (callheap,temporary memory)
    ORA-04030: out of process memory when trying to allocate 56 bytes (callheap,temporary memory)
    Trace dumping is performing id=[cdmp_20110222134605]
    Tue Feb 22 13:46:52 2011
    Sweep Incident[104533]: completed
    Sweep Incident[104532]: completed
    Sweep Incident[104531]: completed
    Sweep Incident[104530]: completed
    Sweep Incident[104528]: completed
    Tue Feb 22 17:04:25 2011
    SYS@MYDB>!oerr ora 4030
    04030, 00000, "out of process memory when trying to allocate %s bytes (%s,%s)"
    //*Cause:  Operating system process private memory was exhausted.
    //*Action:I came to know that it was due to the operating system prvate memory exhausted. Database memory configuration is as below
    SYS@MYDB>show parameter sga
    NAME                                 TYPE        VALUE
    lock_sga                             boolean     FALSE
    pre_page_sga                         boolean     FALSE
    sga_max_size                         big integer 2000M
    sga_target                           big integer 0
    SYS@MYDB>show parameter pga
    NAME                                 TYPE        VALUE
    pga_aggregate_target                 big integer 521658240
    SYS@MYDB>OS Memory information
    load averages:  1.58,  1.82,  2.13;                    up 196+10:43:15                                                         20:55:35
    408 processes: 406 sleeping, 2 on cpu
    CPU states: 80.0% idle, 15.6% user,  4.4% kernel,  0.0% iowait,  0.0% swap
    Memory: 32G phys mem, 18G free mem, 43G total swap, 42G free swapThere are other 2 more databases are running on this server, one is 10g and another one is 9i and its memory configuration have given below respectivly
    SQL>show parameter sga
    NAME                                 TYPE        VALUE
    lock_sga                             boolean     FALSE
    pre_page_sga                         boolean     FALSE
    sga_max_size                         big integer 4G
    sga_target                           big integer 4G
    RPT>show parameter pga
    NAME                                 TYPE        VALUE
    pga_aggregate_target                 big integer 4G
    SQL> show parameter sga
    NAME                                 TYPE        VALUE
    lock_sga                             boolean     FALSE
    pre_page_sga                         boolean     FALSE
    sga_max_size                         big integer 1135171060
    SQL> show parameter pga
    NAME                                 TYPE        VALUE
    pga_aggregate_target                 big integer 2147483648
    SQL>Could you please suggest me how to avoid this error in future ? Looks like i need to fine tune the memory parameters, could you please suggest how can i do so ? I also noticed SGA Target is set to 0 (Automatic Memory managment is not enabled), If i set SGA target will resolve the issue ?
    Kindly advise ..
    Thanks in advance
    Nishant

    I just took the plan by setting set autotrace traceonly explain. Please check this will help, I will check and update you the table's statistics soon.
    Execution Plan
    Plan hash value: 1841079096
    | Id  | Operation                               | Name               | Rows  | Bytes |TempSpc| Cost (%CPU)| Time     |
    |   0 | SELECT STATEMENT                        |                    |   210M|    96G|       |    29M  (1)| 99:12:47 |
    |   1 |  SORT ORDER BY                          |                    |   210M|    96G|       |    29M  (1)| 99:12:47 |
    |   2 |   HASH UNIQUE                           |                    |   210M|    96G|  1000K| 41115   (2)| 00:08:14 |
    |   3 |    CONCATENATION                        |                    |       |       |       |            |          |
    |   4 |     NESTED LOOPS                        |                    |       |       |       |            |          |
    |   5 |      NESTED LOOPS                       |                    |   980 |   470K|       | 34365   (2)| 00:06:53 |
    |   6 |       NESTED LOOPS                      |                    |   980 |   455K|       | 33384   (2)| 00:06:41 |
    |   7 |        NESTED LOOPS                     |                    |   980 |   445K|       | 32893   (2)| 00:06:35 |
    |*  8 |         HASH JOIN                       |                    |   980 |   434K|  7968K| 31912   (2)| 00:06:23 |
    |   9 |          NESTED LOOPS                   |                    | 19606 |  7735K|       | 28804   (2)| 00:05:46 |
    |* 10 |           HASH JOIN                     |                    |     1 |   306 |    11M| 27300   (2)| 00:05:28 |
    |  11 |            TABLE ACCESS FULL            | VERSION_FILE_MAP   |   517K|  6064K|       |  2275   (3)| 00:00:28 |
    |* 12 |            HASH JOIN                    |                    |   296K|    83M|    11M| 20121   (2)| 00:04:02 |
    |* 13 |             HASH JOIN                   |                    |   296K|  8401K|  7168K|  4103   (4)| 00:00:50 |
    |  14 |              INDEX FAST FULL SCAN       | ATTACHMENT_PK      |   407K|  2389K|       |   280   (3)| 00:00:04 |
    |* 15 |              HASH JOIN                  |                    |   296K|  6663K|  8408K|  2966   (4)| 00:00:36 |
    |  16 |               TABLE ACCESS FULL         | ATTACHMENT_MAP     |   296K|  4925K|       |  1810   (5)| 00:00:22 |
    |  17 |               INDEX FAST FULL SCAN      | VERSION_PK         |   482K|  2829K|       |   317   (3)| 00:00:04 |
    |  18 |             TABLE ACCESS FULL           | FILES              |   509K|   128M|       |  8744   (1)| 00:01:45 |
    |* 19 |           TABLE ACCESS FULL             | ATTACHMENT         | 31967 |  3059K|       |  1504   (2)| 00:00:19 |
    |  20 |          TABLE ACCESS FULL              | VERSION            |   482K|    23M|       |  1290   (2)| 00:00:16 |
    |* 21 |         TABLE ACCESS BY INDEX ROWID     | VERSION            |     1 |    12 |       |     1   (0)| 00:00:01 |
    |* 22 |          INDEX UNIQUE SCAN              | VERSION_PK         |     1 |       |       |     1   (0)| 00:00:01 |
    |* 23 |        INDEX UNIQUE SCAN                | PAGE_THREE_PK      |     1 |    10 |       |     1   (0)| 00:00:01 |
    |* 24 |       INDEX UNIQUE SCAN                 | PAGE_TWO_PK        |     1 |       |       |     1   (0)| 00:00:01 |
    |  25 |      TABLE ACCESS BY INDEX ROWID        | PAGE_TWO           |     1 |    16 |       |     1   (0)| 00:00:01 |
    |  26 |     NESTED LOOPS                        |                    |       |       |       |            |          |
    |  27 |      NESTED LOOPS                       |                    |     1 |   492 |       |  6644   (2)| 00:01:20 |
    |  28 |       NESTED LOOPS                      |                    |     1 |   227 |       |  6643   (2)| 00:01:20 |
    |  29 |        NESTED LOOPS                     |                    |     1 |   215 |       |  6641   (2)| 00:01:20 |
    |  30 |         NESTED LOOPS                    |                    |     1 |   209 |       |  6640   (2)| 00:01:20 |
    |  31 |          MERGE JOIN CARTESIAN           |                    |     1 |   192 |       |  6639   (2)| 00:01:20 |
    |  32 |           NESTED LOOPS                  |                    |       |       |       |            |          |
    |  33 |            NESTED LOOPS                 |                    |     1 |   186 |       |  6359   (2)| 00:01:17 |
    |  34 |             NESTED LOOPS                |                    |     1 |   170 |       |  6358   (2)| 00:01:17 |
    |  35 |              NESTED LOOPS               |                    |     1 |   158 |       |  6357   (2)| 00:01:17 |
    |* 36 |               HASH JOIN                 |                    |     1 |   148 |    28M|  6356   (2)| 00:01:17 |
    |  37 |                TABLE ACCESS FULL        | VERSION            |   482K|    23M|       |  1290   (2)| 00:00:16 |
    |* 38 |                TABLE ACCESS FULL        | ATTACHMENT         |   407K|    38M|       |  1504   (2)| 00:00:19 |
    |* 39 |               INDEX UNIQUE SCAN         | PAGE_THREE_PK      |     1 |    10 |       |     1   (0)| 00:00:01 |
    |* 40 |              TABLE ACCESS BY INDEX ROWID| VERSION            |     1 |    12 |       |     1   (0)| 00:00:01 |
    |* 41 |               INDEX UNIQUE SCAN         | VERSION_PK         |     1 |       |       |     1   (0)| 00:00:01 |
    |* 42 |             INDEX UNIQUE SCAN           | PAGE_TWO_PK        |     1 |       |       |     1   (0)| 00:00:01 |
    |  43 |            TABLE ACCESS BY INDEX ROWID  | PAGE_TWO           |     1 |    16 |       |     1   (0)| 00:00:01 |
    |  44 |           BUFFER SORT                   |                    |   407K|  2389K|       |  6638   (2)| 00:01:20 |
    |  45 |            INDEX FAST FULL SCAN         | ATTACHMENT_PK      |   407K|  2389K|       |   280   (3)| 00:00:04 |
    |* 46 |          TABLE ACCESS BY INDEX ROWID    | ATTACHMENT_MAP     |     1 |    17 |       |     2   (0)| 00:00:01 |
    |* 47 |           INDEX RANGE SCAN              | ATTACHMENTMAP_IDX3 |     1 |       |       |     1   (0)| 00:00:01 |
    |* 48 |         INDEX UNIQUE SCAN               | VERSION_PK         |     1 |     6 |       |     1   (0)| 00:00:01 |
    |  49 |        TABLE ACCESS BY INDEX ROWID      | VERSION_FILE_MAP   |     1 |    12 |       |     2   (0)| 00:00:01 |
    |* 50 |         INDEX RANGE SCAN                | VERSION_FMAP_IDX2  |     1 |       |       |     1   (0)| 00:00:01 |
    |* 51 |       INDEX UNIQUE SCAN                 | FILES_PK           |     1 |       |       |     1   (0)| 00:00:01 |
    |* 52 |      TABLE ACCESS BY INDEX ROWID        | FILES              |     1 |   265 |       |     1   (0)| 00:00:01 |
    Predicate Information (identified by operation id):
       8 - access("A"."ID"="VERSION"."ATTACH_ID" AND "A"."LATEST_VSN"="VERSION"."ID")
           filter("A"."LATEST_VSN">="VERSION"."ID")
      10 - access("C"."ID"="D"."VERSION_ID" AND "D"."FILE_ID"="E"."ID")
      12 - access("E"."ID"="A"."FILE_ID")
      13 - access("A"."ATTACH_ID"="B"."ID")
      15 - access("A"."LATEST_VSN"="C"."ID")
      19 - filter(NVL("A"."DELETE_FLAG",0)=0 AND ("CTXSYS"."CONTAINS"("A"."DESCRIPTION",'EHS%')>0 OR
                  "CTXSYS"."CONTAINS"("E"."CONTENT_URL",'EHS%')>0 AND "E"."CONTENT_URL" IS NOT NULL))
      21 - filter("A"."ID"="D"."ATTACH_ID")
      22 - access("A"."LATEST_VSN"="D"."ID")
      23 - access("A"."ID"="C"."ID" AND "A"."CLASS"="C"."CLASS")
      24 - access("A"."ID"="B"."ID" AND "A"."CLASS"="B"."CLASS")
      36 - access("A"."ID"="VERSION"."ATTACH_ID" AND "A"."LATEST_VSN"="VERSION"."ID")
           filter("A"."LATEST_VSN">="VERSION"."ID")
      38 - filter(NVL("A"."DELETE_FLAG",0)=0)
      39 - access("A"."ID"="C"."ID" AND "A"."CLASS"="C"."CLASS")
      40 - filter("A"."ID"="D"."ATTACH_ID")
      41 - access("A"."LATEST_VSN"="D"."ID")
      42 - access("A"."ID"="B"."ID" AND "A"."CLASS"="B"."CLASS")
      46 - filter("A"."FILE_ID"=0)
      47 - access("A"."ATTACH_ID"="B"."ID")
      48 - access("A"."LATEST_VSN"="C"."ID")
      50 - access("C"."ID"="D"."VERSION_ID")
      51 - access("D"."FILE_ID"="E"."ID")
           filter(LNNVL("E"."ID"="A"."FILE_ID"))
      52 - filter("CTXSYS"."CONTAINS"("A"."DESCRIPTION",'EHS%')>0 OR
                  "CTXSYS"."CONTAINS"("E"."CONTENT_URL",'EHS%')>0 AND "E"."CONTENT_URL" IS NOT NULL)

  • ORA-04030 error on backup

    Hi all,
    I´m using oracle 9.2.0.4 with Red Hat9 (memory: 2GB).
    I´m trying to perform a backup by using RMAN, but now I get the error like this:
    kgefec: fatal error 0
    released channel: c1
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03015: error occurred in stored script saturday_backup
    RMAN-03009: failure of backup command on c1 channel at 12/09/2007 06:22:30
    RMAN-10015: error compiling PL/SQL program
    RMAN-10033: error during compilation of job step 2: ORA-00603: ORACLE server session terminated by fatal error
    ORA-04030: out of process memory when trying to allocate 200 bytes (PL/SQL DIANA,LISIMA: Add cells)Somebody nows about this error?
    It seems memory problem , but it began to occur a few days ago.
    Thanks!!!!

    There's a bug reported:
    Bug 3174292 - Proxy backup generates huge PLSQL code with many datafiles
    Doc ID: Note:3174292.8
    Should be fixed in 9.2.0.6 .
    Werner

  • Workaround for ORA-14551 WHILE CALLING FUNCTION IN SQL OVER DBLINK

    Hi,
    any idea how to workaround such issue in 9.2.0.8 (I know this is working with 11.2).
    create table mylog(id number, data date , mess clob);
    create or replace function myfunc(id in number, data in date ,mess in varchar2)
    return number is
    pragma autonomous_transaction;
    retval number;
    begin
       insert into mylog values (id , data ,mess);
    commit;
    retval := id;
    return retval;
    end;
    SQL> select schema.myfunc@dblink(2,sysdate,'bbbbbbb') from dual;
    ERROR at line 1:
    ORA-14551: cannot perform a DML operation inside a query
    ORA-06512: at "SCHEMA.MYFUNC", line 6
    ORA-06512: at line 1Regards
    GregG

    DBMS_SQL is documented in what I think 9i calls "Supplied Oracle Packages and Types" manual.
    The basic concept behind this is that it allows you to manually create a cursor and execute it (via an OCI like interface). For example, it can be used to issue remote DDLs via a database link as described in {message:id=10323373}.
    DBMS_SQL provides flexibility and power that is beyond the standard cursor constructs in PL/SQL. The system package version even enables you to execute SQL and PL/SQL code as any other schema in the database.
    You will have however to play around with DBMS_SQL (locally and remotely) to determine how to hide the fact that a local select (which Oracle assumes, correctly, does not change database state), actually changes the state of a remote database.
    Personally I would not bother with such a hack. I would use the problem to enforce an upgrade to the latest Oracle version, or kick out the crappy app that needs to use a SQL select to make database changes.

  • Possible workaround for: Launch Debugger Error "Unable to connect to the RDS server"

    I do not yet have a reproducible case for this. *Sometimes* ColdFusion Builder throws the "Unable to connect to the RDS server" error, when attempting to start a debug session against a remote server. However, "RDS Dataview" and "RDS Fileview" views correctly list the servers and their related data/file, and testing RDS connection result is "Test connection was successful". The workaround is:
    1. Login to the ColdFusion Administrator
    2. Navigate to "Debugger Settings" page (beneath "Debugging & Logging")
    3. Click on "Stop Debugger Server" -or- "Restart Debugger Server"
    4. Restart ColdFusion Application Server
    Notes:
    a) Step #3 must be done via ColdFusion Administrator, not via ColdFusion Admin API (stopLineDebugger()|restartLineDebugger())
    b) Step #4 can be done via ColdFusion Administrator -or- ColdFusion Builder
    c) Note that there is a CF9 Release Note for Bug #78488: "When a server having "Windows service option" ENABLED is added as a remote server, then Start-Stop-Restart operation cannot be done on it."  However, I've found that "Restarting" a remote server, that is running as a Windows service, is possible in ColdFusion Builder (at least this is so on standalone developer edition).  Like Bug #78488 says, I am unable to "Stop|Start" a remote server that is running as a Windows service. I'm just mentioning all of this, due to my suggestion in "b", that CF can also be restarted via ColdFusion Builder.
    It is interesting that attempts to stop|restart the debugger server, via CF Admin API, do not resolve the issue. So, it seems that the "Stop Debugger Server" and "Restart Debugger Server" buttons, in CF Admin, are doing more than calling the corresponding CF Admin API methods.
    Attached is an image of the error.
    Just posting all of this as an FYI, in case it helps anyone.
    Thanks!,
    -Aaron Neff
    (OH.. I'm unable to attach images at this time *DOH!* ..well, I'll try to remember to update this post w/ the image, once I see that attachments are allowed.)
    I'll try to describe the error message window:
    - It is 494x452 pixels
    - Title of the window is "Error occurred while launching the Debugger"
    - Body content of the window is as follows:
    Unable to connect to the RDS server 'ServerNameHere'
    Ensure that the server is currently running and that line debugging and RDS are enabled. You may also need to check your RDS connection settings in Eclipse to ensure that they are correct.
    Error Message:
    -1:Error connecting to JVM for debugging at localhost:5005. Could not attach to the VM at port 5005. You must specify this debugger port in the JVM settings of your application server, for example: -Xdebug -Xrunjdwp:transport=dt_socket,server=y,suspend=n,address=5005. You will also need to make sure that another debugger is not currently connected to the CF JVM at the same port (5005).
    Root cause ->
    Connect refused: connect

    hi,
    I have been facing the same problems while connecting to Oracle Web Conferencing and is getting error message :Unable to connect to the server. The Web COnferencing Console will now close. Please try again later. If the problem persist, contact your system administrator.
    Diagnostic report shows failed on the connectivity part
    Connectivity
    mx-direct Failed
    0:00:00.719 Connecting to web331.oracle.com:1025
    0:00:00.922 Name resolved to 148.87.108.234
    0:00:21.906 connect() failed (0x8007274C)
    0:00:21.906 Connection failed (0x8007274C)
    Pl help me to fix the problem.
    Regards
    Gobind Ahuja

  • Need your help for ORA-01242 error

    Hi,
    Data base Relaease : 9.2.0.4.0
    OS release: Suse linux 2.4.21-198-smp
    Today morning i have faced one problem in my database. Today i tried to create statspack schema in my production environment from my login (my linux login is in DBA group)... so i just tried to create a tablespace. Once issued create tablespace command my production database gets down... after that i started my database (Database is started with out any problem....). In my alery log i have seen below entries
    Thu Mar 29 11:08:31 2007
    create tablespace stats
    datafile '/sw01/oracle/oradata/cargainz/stat.dbf' size 100M
    Thu Mar 29 11:08:49 2007
    KCF: write/open error block=0x3 online=1
    file=17 /sw01/oracle/oradata/cargainz/stat.dbf
    error=27072 txt: 'Linux Error: 9: Bad file descriptor
    Additional information: 2'
    Thu Mar 29 11:08:49 2007
    Errors in file /sw01/oracle/admin/cargainz/bdump/cargainz_dbw0_1292.trc:
    ORA-01242: data file suffered media failure: database in NOARCHIVELOG mode
    ORA-01114: IO error writing block to file 17 (block # 3)
    ORA-01110: data file 17: '/sw01/oracle/oradata/cargainz/stat.dbf'
    ORA-27072: skgfdisp: I/O error
    Linux Error: 9: Bad file descriptor
    Additional information: 2
    DBW0: terminating instance due to error 1242
    Thu Mar 29 11:08:49 2007
    Errors in file /sw01/oracle/admin/cargainz/bdump/cargainz_ckpt_1296.trc:
    ORA-01242: data file suffered media failure: database in NOARCHIVELOG mode
    Thu Mar 29 11:08:49 2007
    ORA-1242 signalled during: create tablespace stats
    datafile '/sw01/oracle/ora...
    Thu Mar 29 11:08:49 2007
    Instance terminated by DBW0, pid = 1292
    Thu Mar 29 11:08:54 2007
    USER: terminating instance due to error 1092
    Instance terminated by USER, pid = 24370
    Thu Mar 29 11:10:14 2007
    Starting ORACLE instance (normal)
    =================================================
    In cargainz_dbw0_1292.trc
    Oracle9i Enterprise Edition Release 9.2.0.4.0 - Production
    With the Partitioning, OLAP and Oracle Data Mining options
    JServer Release 9.2.0.4.0 - Production
    ORACLE_HOME = /sw01/oracle/9.0.2
    System name: Linux
    Node name: ICM02006
    Release: 2.4.21-198-smp
    Version: #1 SMP Thu Mar 11 17:56:49 UTC 2004
    Machine: i686
    Instance name: cargainz
    Redo thread mounted by this instance: 1
    Oracle process number: 3
    Unix process pid: 1292, image: oracle@ICM02006 (DBW0)
    *** 2007-03-29 11:08:49.159
    *** SESSION ID:(2.1) 2007-03-29 11:08:49.146
    KCF: write/open error block=0x3 online=1
    file=17 /sw01/oracle/oradata/cargainz/stat.dbf
    error=27072 txt: 'Linux Error: 9: Bad file descriptor
    Additional information: 2'
    error 1242 detected in background process
    ORA-01242: data file suffered media failure: database in NOARCHIVELOG mode
    ORA-01114: IO error writing block to file 17 (block # 3)
    ORA-01110: data file 17: '/sw01/oracle/oradata/cargainz/stat.dbf'
    ORA-27072: skgfdisp: I/O error
    Linux Error: 9: Bad file descriptor
    Additional information: 2
    =====================================================
    in cargainz_ckpt_1296.trc
    /sw01/oracle/admin/cargainz/bdump/cargainz_ckpt_1296.trc
    Oracle9i Enterprise Edition Release 9.2.0.4.0 - Production
    With the Partitioning, OLAP and Oracle Data Mining options
    JServer Release 9.2.0.4.0 - Production
    ORACLE_HOME = /sw01/oracle/9.0.2
    System name: Linux
    Node name: ICM02006
    Release: 2.4.21-198-smp
    Version: #1 SMP Thu Mar 11 17:56:49 UTC 2004
    Machine: i686
    Instance name: cargainz
    Redo thread mounted by this instance: 1
    Oracle process number: 5
    Unix process pid: 1296, image: oracle@ICM02006 (CKPT)
    *** 2007-03-29 11:08:49.182
    *** SESSION ID:(4.1) 2007-03-29 11:08:49.182
    error 1242 detected in background process
    ORA-01242: data file suffered media failure: database in NOARCHIVELOG mode
    Can you please any one explain me the reason for this problem?.....
    Regards,
    Karthik

    Your database faced media failure.
    Oracle says...
    ORA-01242 data file suffered media failure: database in NOARCHIVELOG mode
    Cause: The database is in NOARCHIVELOG mode and a database file was detected as inaccessible due to media failure.
    Action: Restore accessibility to the file mentioned in the error stack and restart the instance.
    Oracle was unable to write to the file and instance got terminated.
    Is this your production database ? Your first step is to keep the database in archivelog mode.

  • Trace column for ORA-01438 Error

    Hi
    I encountered the following error while trying to insert into a table using bind variables in a program:
    ORA-01438: value larger than specified precision allows for this column
    I enabled trace for 1438 with errorstack level 4 in order to capture the bind variables being passed into the sql using the following sql:
    alter system set event='1438 trace name context forever, level 4' scope=spfile;
    A trace file is being generated after encountering the error again and I tried to use tkprof to read the trace file, however, I do not see any bind variable in the file. Is there something wrong? How do I trace the column which is causing the problem?
    Thanks in advance!

    Hi,
    With scope=spfile in your alter system command you tell Oracle just to put this into the server parameter file. This file just gets read while startup the database. If you want to set it for the instance now you have to define scope=memory, and if you want to use it now and after a restart you have to type scope=both or just nothing about scope, because both is the default.
    Set the parameter with scope=memory and try it again!
    Or when you running this in PL/SQL just capture the exception and give out the current value.
    Regards

  • Workaround for OCR "Unknown Error"/"Paper Capture recognition service experienced an error"

    Problem Description:
    When attempting to "Recognize Text", one of the following error messages appears:
    "Unknown Error." when using one of the "Searchable Image" or "Searchable Image (Exact)" settings, or
    "Unable to process the page because the Paper Capture recognition service experienced an error. (6)" when using the "ClearScan" setting.
    Problem Origin:
    It seems to be a compatibility issue in the PDF and not in the OCR algorithm itself, although this is speculation and not an official statement.
    Problem Workaround:
    A possible solution consists of saving the afflicted PDF pages as PNG images (lossless), opening the PNGs in Acrobat, and then performing OCR. Specifically,
    Page Thumbnails -> Right-click afflicted pages -> Extract Pages...
    File -> Open extracted pages
    File -> Save as -> Image -> PNG
    Right-click PNG files -> Convert to Adobe PDF
    Tools -> Recognize Text -> In This File
    File -> Open original PDF
    Page Thumbnails -> Right-click afflicted pages -> Replace Pages...
    This has worked for every problem instance that I've come across, and there appears to be no quality loss. If anyone else can verify this solution, I would appreciate it.

    Attempt to run Acrobat 9.5.3 OCR on MacBook Pro under Snow Leopard generates the error message:
       Acrobat could not perform recognition (OCR) because:
       Unable to process the page because the Paper Capture recognition service experienced and error. (17)
       Ignore future errors in this document.
    Login as Administrator, launch Acrobat, open one page PDF document without OCR, ..., OCR successful, close document, note additional warning message:
       Suit Product Activation Needed
       Adobe Acrobat was installed as part of a suite.  To enable Adobe Acrobat, please start another component of this suite (such as Adobe Photoshop).
    Switch back to user account, quit and relaunch Acrobat and OCR works fine.
    Looking back at my records I see that I ran into the same error on 11/11/11 probably under Acrobat 9.4.6, but I didn't record the fix.

  • My workaround for "An unknown error occured (-36)"

    I have a fairly new iPod Shuffle, running Win 2000 SP4, iTunes 4.7.1.30, and apart from the first time after (re)installing the iPod Setup software, every time (without fail) I plug my Shuffle into my computer and try to transfer songs onto/off it, I would get the error message:
    "An unknown error occured (-36)"
    And nothing would happen. I used it at work on my Mac with no problem, so was sure it wasn't the Shuffle. I don't have any space on the Shuffle allocated to filestorage (doing so might actually solve the problem - untested).
    After reading some threads here and other places, I did some experimenting, and came up with this really simple work around which has worked for me without fail!
    1. Don't have iTunes running. You need to beat it.
    2. When you plug in your Shuffle, quickly go into Windows Explorer, and you will see (might need to refresh F5) a new removable disk appears. Double click it as soon as you can. That's all you need to do. Normally iTunes would hide it again (since I guess it knows I am not using it for Filestorage). And of course normally I then just get the error message.
    4. But now, I can use my Shuffle every time with no problems. The Removable Disk remains visible in Windows Explorer.
    I am guessing it is the dismounting of the drive by iTunes that causes Windows some problems.
    Incidently, you can do the same thing on the Mac, opening the Shuffle in the Finder when it first appears, and this also prevents iTunes from dismounting it. But I had no problems on the Mac.
    Edit: I am using iPod Updater 12-10-2005, iPod Shuffle software 1.1.2

    See if the troubleshooting document helps: "-36 error when updating or restoring iPod

  • Solution for ora-28604 error?

    CREATE BITMAP INDEX IDX_NAME ON TABLE1(CODE);
    This creation is giving the following error:
    ORA-28604: table too fragmented to build bitmap index (39548144,72,72)
    How to resolve this issue?

    HI...
    U can resolve using two methods:-
    1. Export the table, drop the table and import it and then create the index.
    2. move the table into another tablespace.
    ALETR TABLE <TABLE_NAME> MOVE TABLESPACE  <TBS_NAME>;
    the user whose table u are moving, should have privileges on that tablespace.
    Go through METALINK Doc ID: Note:119674.1

  • Workaround for ORA-02060: update on view with db-link

    Hello,
    In my application I have a page that is based on a view. This view itself is based on a table in the same database and a table in another database which I pick up through a db-link.
    To update the data in the view I have created an instead-of trigger which handles all the DML (which is a direct update on the local table and a message towards a broker for the update of the remote database).
    When updating from the database with a statement like UPDATE VIEW SET COLUMN1 = VALUE WHERE COLUMN2 = PK_ID, there is no problem. Everything is triggered in the right way.
    But when I want to update a row in my APEX page I get 'ORA-02060: select for update specified a join of distributed tables'. This is due to the fact that APEX fetches the row as SELECT * FROM VIEW FOR UPDATE.
    How can I work around this? I don't need the locking mechanism that 'FOR UPDATE' provides, because I can solve that myself. So basically, can I alter APEX's statement that performs the Automatic Row Processing (DML)?

    So you answered your question yourself :) Pretty or not, it does the work. Imagine programmers beeing able to work only using wizards and have no clue about what they do. What kind of quality would you get? :)
    Denes Kubicek
    http://deneskubicek.blogspot.com/
    http://www.opal-consulting.de/training
    http://apex.oracle.com/pls/otn/f?p=31517:1
    -------------------------------------------------------------------

  • ORA-04030 error on Oracle 10g R2 and Windows Server 2003

    We are getting this messege atleast once a day, using Oracle database 10g R2 with Oracle Developer 6i forms, we have to restart the service and then it works fine.... DB Server has 4GB RAM
    SGA 1.6 GB
    PGA 800 M

    Hi
    It is due to lack of memory or due to more put more work on server processes so try to increase memory or decrease the number of users
    Shafiulla

  • Flashback : Error:ORA-04030: out of process memory when trying to allocate

    Hi All,
    I have executed this query on my db for using flash back features:
    select * from flashback_transaction_query where table_owner='USERNAME'
    However, it throws the error after being executed for 5 min:
    ORA-04030: out of process memory when trying to allocate 268 bytes (Logminer LCR c,krvxbpdl)
    Please help me in understading the error and provide some links/docs to resolve this.
    Thanks,
    Kishore

    That's the error description:
    What does an ORA-4030 mean?
    This error indicates that the oracle server process is unable to allocate more memory from the operating system.This memory consists of the PGA (Program Global Area) and its contents depend upon the server configuration.For dedicated server processes it contains the stack and the UGA (User Global Area) which holds user session data, cursor information and the sort area. In a multithreaded configuration (shared server), the UGA is allocated in the SGA (System Global Area) and will not be responsible for ORA-4030 errors.
    The ORA-4030 thus indicates the process needs more memory (stack UGA or PGA) to perform its job.
    On metalink:
    Diagnosing and Resolving ORA-4030 errors
    Doc ID: Note:233869.1
    Werner

Maybe you are looking for

  • Opening And Closing Balance Using Union ALL

    Hi, I have two txt files, once contains AccountNo, OpeningBalance, and the other file contains AccountNo, ClosingBalance. I have two varibles PeriodStart = 1 and PeriodEnd=12. I am tring to get this output. I have used two flat file source, sort it,

  • Unable to load the specified offline registry hive when rejoining the domain

    I have a Windows Server 2008 R2 Ent member server that we needed to clone. So we removed it from the domain, cloned it, and when we went back to join it to the domain it we get a error:   The Following error occurred attempting to join the domain:  

  • Editing CSS in Dreamweaver.

    Well Im trying to edit a CSS document from a template. There is a bunch of sections and I need to find out which change what. Well I opened the CSS file in Dreamweaver but it doesn't allow me to view in browser... Is there something I need to do so I

  • Can't find setting for IE10 Compatibility View in Group Policy Management Console.

    Hoping this is easy. I'm trying to use the Group Policy setting for Compatibility View under Computer Configuration/Administrative Templates/Windows Components/Internet Explorer and it is not there. The domain controllers are 2008 Standard SP2, so no

  • HT201343 Is Mountain Lion Airplay feature enabled on boot-up?

    I would like to be able to log into and access my Mac Mini from the comfort of my sofa. If Airplay is enabled, can i use the AppleTV to display the initial login screen. Also, assuming the above is possible, should the Bluetooth keyboard be paired wi