FullOffline Backup - ORA-19566: exceeded limit of 0 corrupt blocks for file

Dear SAP gurus,
I am getting an error from the DBA Planning Calendar every time the job for "Full Offline backup" is run. And it is always as you can see from the log on the same file "oracle/SHD/sapdata4/sr3_16/sr3.data16".
The oracle error is the following:
ORA-19566: exceeded limit of 0 corrupt blocks for file /oracle/SHD/sapdata4/sr3_16/sr3.data16
I found the SAP Note 969192 - RMAN Backup of SYSTEM tablespace terminates with ORA-19566
but it does no apply because this is for the tablespace SYSTEM and not PSAPSR3.
Please find below the log:
BR0051I BRBACKUP 7.00 (46)
BR0055I Start of database backup: begomwsv.ffd 2011-08-17 10.01.37
BR0484I BRBACKUP log file: /oracle/SHD/sapbackup/begomwsv.ffd
BR0477I Oracle pfile /oracle/SHD/102_64/dbs/initSHD.ora created from spfile /oracle/SHD/102_64/dbs/spfileSHD.ora
BR0101I Parameters
Name                           Value
oracle_sid                     SHD
oracle_home                    /oracle/SHD/102_64
oracle_profile                 /oracle/SHD/102_64/dbs/initSHD.ora
sapdata_home                   /oracle/SHD
sap_profile                    /oracle/SHD/102_64/dbs/initSHD.sap
backup_mode                    FULL
backup_type                    offline_force
backup_dev_type                disk
backup_root_dir                /mnt/backup/oracle/SHD
compress                       no
disk_copy_cmd                  rman
cpio_disk_flags                -pdcu
exec_parallel                  0
rman_compress                  no
system_info                    shdadm/orashd eccdev01 Linux 2.6.16.60-0.87.1-smp #1 SMP Wed May 11 11:48:12 UTC 2011 x86_64
oracle_info                    SHD 10.2.0.4.0 8192 17654 1114483454 eccdev01 UTF8 UTF8
sap_info                       700 SAPSR3 0002LK0003SHD0011Y01548735220015Maintenance_ORA
make_info                      linuxx86_64 OCI_102 Jan 29 2010
command_line                   brbackup -u / -jid FLLOF20110817100136 -c force -t offline_force -m full -p initSHD.sap
BR0116I ARCHIVE LOG LIST before backup for database instance SHD
Parameter                      Value
Database log mode              No Archive Mode
Automatic archival             Disabled
Archive destination            /oracle/SHD/oraarch/SHDarch
Archive format                 %t_%s_%r.dbf
Oldest online log sequence     17651
Next log sequence to archive   17654
Current log sequence           17654            SCN: 1114483454
Database block size            8192             Thread: 1
Current system change number   1114501246       ResetId: 664011854
BR0118I Tablespaces and data files
BR0202I Saving /oracle/SHD/sapdata3/sr3_15/sr3.data15
BR0203I to /mnt/backup/oracle/SHD/begomwsv/sr3.data15 ...
#FILE..... /oracle/SHD/sapdata3/sr3_15/sr3.data15
#SAVED.... /mnt/backup/oracle/SHD/begomwsv/sr3.data15  #1/15
BR0280I BRBACKUP time stamp: 2011-08-17 10.28.42
BR0063I 15 of 48 files processed - 44100.117 of 121180.346 MB done
BR0204I Percentage done: 36.39%, estimated end time: 11:15
BR0001I ******************________________________________
BR0202I Saving /oracle/SHD/sapdata4/sr3_16/sr3.data16
BR0203I to /mnt/backup/oracle/SHD/begomwsv/sr3.data16 ...
BR0278E Command output of 'SHELL=/bin/sh /oracle/SHD/102_64/bin/rman nocatalog':
Recovery Manager: Release 10.2.0.4.0 - Production on Wed Aug 17 10:28:42 2011
Copyright (c) 1982, 2007, Oracle.  All rights reserved.
RMAN>
RMAN> connect target *
connected to target database: SHD (DBID=1683093070, not open)
using target database control file instead of recovery catalog
RMAN> *end-of-file*
RMAN>
host command complete
RMAN> 2> 3> 4> 5> 6>
allocated channel: dsk
channel dsk: sid=223 devtype=DISK
executing command: SET NOCFAU
Starting backup at 17-AUG-11
channel dsk: starting datafile copy
input datafile fno=00019 name=/oracle/SHD/sapdata4/sr3_16/sr3.data16
released channel: dsk
RMAN-00571: ===========================================================
RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
RMAN-00571: ===========================================================
RMAN-03009: failure of backup command on dsk channel at 08/17/2011 10:30:30
ORA-19566: exceeded limit of 0 corrupt blocks for file /oracle/SHD/sapdata4/sr3_16/sr3.data16
RMAN>
Recovery Manager complete.
BR0280I BRBACKUP time stamp: 2011-08-17 10.30.30
BR0279E Return code from 'SHELL=/bin/sh /oracle/SHD/102_64/bin/rman nocatalog': 1
BR0536E RMAN call for database instance SHD failed
BR0280I BRBACKUP time stamp: 2011-08-17 10.30.30
BR0506E Full database backup (level 0) using RMAN failed
BR0222E Copying /oracle/SHD/sapdata4/sr3_16/sr3.data16 to/from /mnt/backup/oracle/SHD/begomwsv failed due to previous errors
BR0280I BRBACKUP time stamp: 2011-08-17 10.30.34
BR0307I Shutting down database instance SHD ...
BR0280I BRBACKUP time stamp: 2011-08-17 10.30.34
BR0308I Shutdown of database instance SHD successful
BR0280I BRBACKUP time stamp: 2011-08-17 10.30.34
BR0304I Starting and opening database instance SHD ...
BR0280I BRBACKUP time stamp: 2011-08-17 10.30.47
BR0305I Start and open of database instance SHD successful
Do you guys have any idea on how to solve this issue??
Thanks in advance, Marc

Hi,
I am getting an error from the DBA Planning Calendar every time the job ...
So when was your last successfull backup of this datafile. Check if still available.
If this is some time ago, and may be you are currently without any backup, try to backup without rman at once,
to have at least something to work with in case you get additional errors right now.
Then you need to find out what object is affected. You are on the right way already. You need the statement,
that goes to dba_extents to check what object the block belongs to.
Has the DB been recovered recently, so the block might possibly belong to an index created with nologging ?
(this could be the case on BW systems).
If the last good backup of that file is still available and the redologs belonging to this backup up to current time are as well, you could try to recover that file. But I'd do this only after a good backup without rman and by not destroying the original file.
If the last good backup was an rman backup, you can do a verify restore of that datafile in advance, to check if the corruption is really not inside the file to be restored.
Check out the -w (verify) option of brrestore first, to understand how it works.
(I am not sure it this is already available in version 7.00, may be you need to switch to 7.10 or 7.20)
brrestore -c -m /oracle/SHD/sapdata4/sr3_16/sr3.data16  -b xxxxxxxx.ffr -w only_rmv
You should do a dbv check of that file as well, to check if it gets more information. I.E if more blocks are
affected. rman stops right after the first corruption, but usually you have a couple of those in line, esp. if these are
zeroed ones. (This one would also work with version 7.00 brtools)
brbackup -c -u / -t online -m /oracle/SHD/sapdata4/sr3_16/sr3.data16 -w only_dbv
Good luck.
Volker

Similar Messages

  • ORA-19566: exceeded limit of 999 corrupt blocks for file

    Hi All,
    I am new to Oracle RMAN & RAC Administration. Looking for your support to solve the below issue.
    We have 2 disk groups - +ETDATA & +ETFLASH in our 3 node RAC environments in which RMAN is configured in node-2 to take backup. We do not have RMAN catalog and the RMAN is fetching information from control file.
    Recently, the backup failed with the error ORA-19566: exceeded limit of 999 corrupt blocks for file +ETFLASH/datafile/users.6187.802328091.
    We found that the datafiles are present in both disk groups and from the control file info, we got to know that the datafiles in +ETDATA are currently in use and +ETFLASH is having old datafiles.
    RMAN> show all;
    RMAN configuration parameters for database with db_unique_name LABWRKT are:
    CONFIGURE RETENTION POLICY TO RECOVERY WINDOW OF 3 DAYS;
    CONFIGURE BACKUP OPTIMIZATION ON;
    CONFIGURE DEFAULT DEVICE TYPE TO DISK; # default
    CONFIGURE CONTROLFILE AUTOBACKUP ON;
    CONFIGURE CONTROLFILE AUTOBACKUP FORMAT FOR DEVICE TYPE DISK TO '%F'; # default
    CONFIGURE DEVICE TYPE DISK PARALLELISM 4 BACKUP TYPE TO BACKUPSET;
    CONFIGURE DATAFILE BACKUP COPIES FOR DEVICE TYPE DISK TO 1; # default
    CONFIGURE ARCHIVELOG BACKUP COPIES FOR DEVICE TYPE DISK TO 1; # default
    CONFIGURE MAXSETSIZE TO UNLIMITED; # default
    CONFIGURE ENCRYPTION FOR DATABASE OFF; # default
    CONFIGURE ENCRYPTION ALGORITHM 'AES128'; # default
    CONFIGURE COMPRESSION ALGORITHM 'BASIC' AS OF RELEASE 'DEFAULT' OPTIMIZE FOR LOAD TRUE ; # default
    CONFIGURE ARCHIVELOG DELETION POLICY TO NONE; # default
    CONFIGURE SNAPSHOT CONTROLFILE NAME TO '+ETFLASH/CONTROLFILE/snapcf_LABWRKT.f';
    CONFIGURE SNAPSHOT CONTROLFILE NAME TO '+ETFLASH/controlfile/snapcf_labwrkt.f';
    Above configuration shows that SNAPSHOT CONTROLFILE is pointing to +ETFLASH. So I changed the configuration with the SNAPSHOT CONTROLFILE points to '+ETDATA/controlfile/snapcf_labwrkt.f'. At the end of backup, SNAPSHOT file was created in +ETDATA and I was expecting it to be the copy of control file being used which has dbf located in +ETDATA. But still the backup was pointing to old datafiles in +ETFLASH. Since we dont have RMAN catalog, resync also is not possible.
    When I ran it manually, it was successfull without any error and was pointing to the exisiting datafiles.
    RMAN> backup database plus archivelog all;
    I hope the issue will get resolved if the RMAN points only to the datafiles present in +ETDATA. If I am correct, please let me know how can i make it possible? Also please explain me why the newly created snapshot file does not reflect the existing control file info?

    Hi,
    I am getting an error from the DBA Planning Calendar every time the job ...
    So when was your last successfull backup of this datafile. Check if still available.
    If this is some time ago, and may be you are currently without any backup, try to backup without rman at once,
    to have at least something to work with in case you get additional errors right now.
    Then you need to find out what object is affected. You are on the right way already. You need the statement,
    that goes to dba_extents to check what object the block belongs to.
    Has the DB been recovered recently, so the block might possibly belong to an index created with nologging ?
    (this could be the case on BW systems).
    If the last good backup of that file is still available and the redologs belonging to this backup up to current time are as well, you could try to recover that file. But I'd do this only after a good backup without rman and by not destroying the original file.
    If the last good backup was an rman backup, you can do a verify restore of that datafile in advance, to check if the corruption is really not inside the file to be restored.
    Check out the -w (verify) option of brrestore first, to understand how it works.
    (I am not sure it this is already available in version 7.00, may be you need to switch to 7.10 or 7.20)
    brrestore -c -m /oracle/SHD/sapdata4/sr3_16/sr3.data16  -b xxxxxxxx.ffr -w only_rmv
    You should do a dbv check of that file as well, to check if it gets more information. I.E if more blocks are
    affected. rman stops right after the first corruption, but usually you have a couple of those in line, esp. if these are
    zeroed ones. (This one would also work with version 7.00 brtools)
    brbackup -c -u / -t online -m /oracle/SHD/sapdata4/sr3_16/sr3.data16 -w only_dbv
    Good luck.
    Volker

  • ORA-19566: exceeded limit of 0 corrupt blocks

    Hi All,
    We have been encountering some issues with RMAN backup; it has been erroring out with same errors (max corrupt blocks). As of now, I ran the db verify for affected files and found that indexes are failing. When I tried to find out the indexes from extent views, I was unable to find it. Looks like these blocks are in free space as I found it and also the V$backup corruption view shows the logical corruption.
    Waiting for you suggestion....
    BANNER
    Oracle Database 10g Enterprise Edition Release 10.2.0.3.0 - 64bi
    PL/SQL Release 10.2.0.3.0 - Production
    CORE 10.2.0.3.0 Production
    TNS for HPUX: Version 10.2.0.3.0 - Production
    NLSRTL Version 10.2.0.3.0 - Production
    RMAN LOG:
    channel a3: starting piece 1 at 14-DEC-09
    RMAN-03009: failure of backup command on a2 channel at 12/14/2009 05:43:42
    ORA-19566: exceeded limit of 0 corrupt blocks for file /ub834/oradata/TERP/applsysd142.dbf
    continuing other job steps, job failed will not be re-run
    channel a2: starting incremental level 0 datafile backupset
    channel a2: specifying datafile(s) in backupset
    including current control file in backupset
    channel a2: starting piece 1 at 14-DEC-09
    channel a1: finished piece 1 at 14-DEC-09
    piece handle=TERP_1769708180_level0_292_1_1_20091213065437.rmn tag=TAG20091213T065459 comment=API Version 2.0,MMS Version 5.0.0.0
    channel a1: backup set complete, elapsed time: 01:14:45
    channel a2: finished piece 1 at 14-DEC-09
    piece handle=TERP_1769708180_level0_296_1_1_20091213065437.rmn tag=TAG20091213T065459 comment=API Version 2.0,MMS Version 5.0.0.0
    channel a2: backup set complete, elapsed time: 00:24:54
    RMAN-03009: failure of backup command on a4 channel at 12/14/2009 06:14:33
    ORA-19566: exceeded limit of 0 corrupt blocks for file /ub834/oradata/TERP/applsysd143.dbf
    continuing other job steps, job failed will not be re-run
    released channel: a1
    released channel: a2
    released channel: a3
    released channel: a4
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03009: failure of backup command on a3 channel at 12/14/2009 06:41:00
    ORA-19566: exceeded limit of 0 corrupt blocks for file /ub806/oradata/TERP/icxd01.dbf
    Recovery Manager complete.
    Thanks,
    Vimlendu
    Edited by: Vimlendu on Dec 20, 2009 10:27 AM

    dbv file=/ora/oradata/binadb/RAT_TRANS_IDX01.dbf blocksize=8192
    The result:
    DBVERIFY: Release 10.2.0.3.0 - Production on Thu Nov 20 11:14:01 2003
    (c) Copyright 2000 Oracle Corporation. All rights reserved.
    DBVERIFY - Verification starting : FILE =
    /ora/oradata/binadb/RAT_TRANS_IDX01.dbf
    Block Checking: DBA = 75520968, Block Type = KTB-managed data block
    **** row 80: key out of order
    ---- end index block validation
    Page 23496 failed with check code 6401
    DBVERIFY - Verification complete
    Total Pages Examined : 34560
    Total Pages Processed (Data) : 1
    Total Pages Failing (Data) : 0
    Total Pages Processed (Index): 31084
    Total Pages Failing (Index): 1
    Total Pages Processed (Other): 191
    Total Pages Empty : 3284
    Total Pages Marked Corrupt : 0
    Total Pages Influx : 0
    Seems like I have 1 page failing. I try to run this script:
    select segment_type, segment_name, owner
    from sys.dba_extents
    where file_id = 18 and 23496 between block_id
    and block_id + blocks - 1;
    No rows returned.
    Then, I try to run this script:
    Select tablespace_name, file_id, block_id, bytes
    from dba_free_space
    where file_id = 18
    and 23496 between block_id and block_id + blocks - 1
    Resulting 1 row.
    Seems like I have the possible corrupt block on unused space.
    Edited by: Vimlendu on Dec 20, 2009 2:30 PM
    Edited by: Vimlendu on Dec 20, 2009 2:41 PM

  • ORA-19510 failed to set size of 12288 blocks for file "/sisdev01/dbs/log01b.dbf

    My company's harddisk has corrupted and files under /sisdev01/dbs/ cannot list and don't have any backup, then I have to create a new database again after changing the disk and recovering the data. I plan to reuse the /sisdev01/dbs/ as following.
    I fail in creating database by using the following script:
    CREATE DATABASE SISDEV01
         LOGFILE GROUP 1 '/sisdev01/dbs/log01a.dbf' SIZE 10M
         ,     GROUP 1 '/sisdev01/dbs/log01b.dbf' SIZE 10M
         ,     GROUP 2 '/sisdev01/dbs/log02a.dbf' SIZE 10M
         ,     GROUP 2 '/sisdev01/dbs/log02b.dbf' SIZE 10M
         NOARCHIVELOG
         CHARACTER SET ZHT16BIG5
         NATIONAL CHARACTER SET ZHT16BIG5
         DATAFILE '/sisdev01/dbs/system01.dbf'      
              SIZE 210M AUTOEXTEND OFF
         , '/sisdev01/dbs/rbs01.dbf'
    SIZE 105M AUTOEXTEND OFF
         , '/sisdev01/dbs/temp01.dbf'
    SIZE 105M AUTOEXTEND OFF
         , '/sisdev01/dbs/users01.dbf'
    SIZE 1050M AUTOEXTEND OFF
         , '/sisdev01/dbs/ts_index_ivr01.dbf'
    SIZE 105M AUTOEXTEND OFF
         , '/sisdev01/dbs/ts_ivr01.dbf'
    SIZE 21M AUTOEXTEND OFF
         , '/sisdev01/dbs/ts_ems01.dbf'
    SIZE 21M AUTOEXTEND OFF
         , '/sisdev01/dbs/ts_index_ems01.dbf'
    SIZE 21M AUTOEXTEND OFF
    but I receive error message:
    CREATE DATABASE SISDEV01
    ORA-01501: CREATE DATABASE failed
    ORA-19510: failed to set size of 12288 blocks for file "/sisdev01/dbs/log01b.dbf
    " (blocksize=512)
    ORA-27059: skgfrsz: could not reduce file size
    IBM AIX RISC System/6000 Error: 28: No space left on device
    Additional information: 1
    ORA-19502: write error on file "/sisdev01/dbs/log01b.dbf", blockno 8193 (blocksi
    ze=512)
    ORA-27063: skgfospo: number of bytes read/written is incorrect
    IBM AIX RISC System/6000 Error: 28: No space left on device
    Additional information: -1
    Additional information: 524288
    Here is the content of my parameter file:
    db_file_multiblock_read_count = 8 # SMALL
    # db_file_multiblock_read_count = 16 # MEDIUM
    # db_file_multiblock_read_count = 32 # LARGE
    # db_block_buffers = 60 # SMALL
    db_block_buffers = 550 # MEDIUM
    # db_block_buffers = 3200 # LARGE
    # shared_pool_size = 3500000 # SMALL
    shared_pool_size = 5000000 # MEDIUM
    # shared_pool_size = 9000000 # LARGE
    log_checkpoint_interval = 10000
    processes = 400 # SMALL
    # processes = 100 # MEDIUM
    # processes = 200 # LARGE
    dml_locks = 100 # SMALL
    # dml_locks = 200 # MEDIUM
    # dml_locks = 500 # LARGE
    # log_buffer = 8192 # SMALL
    log_buffer = 32768 # MEDIUM
    # log_buffer = 163840 # LARGE
    # audit_trail = true # if you want auditing
    # timed_statistics = true # if you want timed statistics
    max_dump_file_size = 10240 # limit trace file size to 5 Meg each
    # Uncommenting the line below will cause automatic archiving if archiving has
    # been enabled using ALTER DATABASE ARCHIVELOG.
    log_archive_start = false
    log_archive_dest = /u01/sisdev01_arch
    log_archive_format = arch%s.dbf
    # If using private rollback segments, place lines of the following
    # form in each of your instance-specific init.ora files:
    rollback_segments = (r01, r02)
    # If using public rollback segments, define how many
    # rollback segments each instance will pick up, using the formula
    # # of rollback segments = transactions / transactions_per_rollback_segment
    # In this example each instance will grab 40/10 = 4:
    transactions = 40
    transactions_per_rollback_segment = 10
    # Global Naming -- enforce that a dblink has same name as the db it connects to
    global_names = TRUE
    # Edit and uncomment the following line to provide the suffix that will be
    # appended to the db_name parameter (separated with a dot) and stored as the
    # global database name when a database is created. If your site uses
    # Internet Domain names for e-mail, then the part of your e-mail address after
    # the '@' is a good candidate for this parameter value.
    # db_domain = us.acme.com # global database name is db_name.db_domain
    # define two control files by default
    # control_files = (/sisuat02/dbs/ctl01.ora,/sisuat02/dbs/ctl02.ora)
    control_files = (/sisdev01/dbs/ctl01.ora,/sisdev01/dbs/ctl02.ora)
    DB_NAME=sisdev01
    DB_BLOCK_SIZE=8192
    # DB_BLOCK_LRU_LATCHES=4
    DB_FILES=500
    DB_DOMAIN=shk.sony.com.hk
    background_dump_dest = /u01/dba/sisdev01/log
    user_dump_dest = /u01/dba/sisdev01/log
    NLS_LANGUAGE="traditional chinese"
    NLS_TERRITORY="hong kong"
    NLS_CURRENCY="HK$ "
    NLS_DATE_FORMAT="DD/MM/YYYY"
    NLS_NUMERIC_CHARACTERS=".,"
    utl_file_dir = *
    compatible = 8.1.5
    remote_os_authent = TRUE
    sort_area_size = 65536
    sort_area_retained_size = 65536
    It always fail during creation of logfile log01b.dbf.
    I have tried to send same size logfiles to the directory but it fails too when inserting the second logfile log01b.dbf too.
    NETOUT: 10054452 Error writing file: Error 0
    Can anyone kindly help me?

    This may be a blindingly obvious question but do you have sufficient disk space on /sisdev01/dbs/ partition?

  • ORA-27047: unable to read the header block of file

    My Windows 2003 crashed which was running Oracle XE.
    I installed Oracle XE on Windows XP on another machine.
    I coped my D:\oracle\XE10g\oradata folder of Win2003 to the same location in WinXP machine.
    When I start the database in WinXP using SQLPLUS i get the following message
    SQL> startup
    ORACLE instance started.
    Total System Global Area 146800640 bytes
    Fixed Size 1286220 bytes
    Variable Size 62918580 bytes
    Database Buffers 79691776 bytes
    Redo Buffers 2904064 bytes
    ORA-00205: error in identifying control file, check alert log for more info
    I my D:\oracle\XE10g\app\oracle\admin\XE\bdump\alert_xe I found following errors
    starting up 1 dispatcher(s) for network address '(ADDRESS=(PARTIAL=YES)(PROTOCOL=TCP))'...
    starting up 4 shared server(s) ...
    Oracle Data Guard is not available in this edition of Oracle.
    Wed Apr 25 18:38:36 2007
    ALTER DATABASE MOUNT
    Wed Apr 25 18:38:36 2007
    ORA-00202: control file: 'D:\ORACLE\XE10G\ORADATA\XE\CONTROL.DBF'
    ORA-27047: unable to read the header block of file
    OSD-04001: invalid logical block size (OS 2800189884)
    Wed Apr 25 18:38:36 2007
    ORA-205 signalled during: ALTER DATABASE MOUNT...
    ORA-00202: control file: 'D:\ORACLE\XE10G\ORADATA\XE\CONTROL.DBF'
    ORA-27047: unable to read the header block of file
    OSD-04001: invalid logical block size (OS 2800189884)
    Please help.
    Regards,
    Zulqarnain

    Try to install win 2003 server software, do the fresh installation of oracle software, now copy the datafiles and controlfiles to same locations as you did on winxp.
    get back to us, if still not out of the woods. I still doubt that a simple restore would do the trick, since you doing it across different platforms, might be I can be wrong, but this is what I personally feel, you not able to start the database on winxp.
    hare krishna
    Alok

  • ORA-27044: unable to write the header block of file / IBM AIX RISC System/6

    We are getting an error like this...
    Oracle 11.2.0.2
    AIX 5.3
    Error:
    CREATE CONTROLFILE REUSE DATABASE "CLMST" RESETLOGS NOARCHIVELOG
    ERROR at line 1:
    ORA-01503: CREATE CONTROLFILE failed
    ORA-00200: control file could not be created
    ORA-00202: control file: '/clmst02/UAT/oradata/control/control01.ctl'
    ORA-27044: unable to write the header block of file
    IBM AIX RISC System/6000 Error: 89: Invalid file system control data detected
    Additional information: 3
    Additional information: -1
    Additional information: 1

    ORA-09817 IBM AIX RISC System/6000 Error: 89
    Oracle 11.2.0.2
    AIX 5.3
    Related Errors:-
    Mon Nov 05 18:33:51 2012
    ARC3 started with pid=30, OS id=954622
    ARC1: Becoming the 'no FAL' ARCH
    ARC1: Becoming the 'no SRL' ARCH
    ARC2: Becoming the heartbeat ARCH
    Errors in file /clmst01/UAT/oracode/app/oracle/diag/rdbms/clmst/clmst/trace/clmst_ora_2514960.trc (incident=17053):
    ORA-00600: internal error code, arguments: [kccugg_end], [], [], [], [], [], [], [], [], [], [], []
    Errors in file /clmst01/UAT/oracode/app/oracle/diag/rdbms/clmst/clmst/trace/clmst_m000_2105586.trc:
    ORA-00313: open failed for members of log group 3 of thread 1
    ORA-00312: online log 3 thread 1: '/clmst02/UAT/oradata/redo/redo03A.log'
    ORA-27037: unable to obtain file status
    IBM AIX RISC System/6000 Error: 2: No such file or directory
    CREATE CONTROLFILE REUSE DATABASE "CLMST" RESETLOGS NOARCHIVELOG
    ERROR at line 1:
    ORA-01503: CREATE CONTROLFILE failed
    ORA-00200: control file could not be created
    ORA-00202: control file: '/clmst02/UAT/oradata/control/control01.ctl'
    ORA-27044: unable to write the header block of file
    IBM AIX RISC System/6000 Error: 89: Invalid file system control data detected
    Additional information: 3
    Additional information: -1
    Additional information: 1
    Cause:-
    Your /clmst02/ mount point might be corrupted.
    Check:-
    fsck /clmst02
    The current volume is: /dev/fslv00
    File system is currently mounted.
    Primary superblock is valid.
    fsck: Performing read-only processing does not produce dependable results.
    *** Phase 1 - Initial inode scan
    *** Phase 2 - Process remaining directories
    *** Phase 3 - Process remaining files
    *** Phase 4 - Check inode allocation map
    File system inode map is corrupt (NOT FIXED)
    fsck: 0507-278 Cannot continue.
    File system is currently mounted.
    fsck: Performing read-only processing does not produce dependable results.
    Fix:-
    Either you can change the file location to another mount points or shutdown the database and run the fsck on /clmst02 mount point to fix the corrupted one & start the database.
    Fix:-
    Recreate the affected Redolog files
    alter database add logfile group 3 (
    '/clmst02/UAT/oradata/redo/redo03A.log',
    '/clmst03/UAT/oradata2/redo/redo03B.log',
    '/clmst04/UAT/oradata3/redo/redo03C.log') size 50m reuse;

  • Without exceeding the credit limit sales order is blocked for credit check.

    Dear Experts,
    In my production server as well as quality server sales order is blocked for dynamic credit check.But the fact is that the credit limit is not exceeded.only 18% of the credit limit is used by the cutomer.The same problem is there in quality server also.
    Kindly help.
    Best Regards
    Biju

    Dear Biju,
    Check whether the customer code has exceeded credit limit in any other credit control area.
    Run the following programs.
    RVKRED77 , RVKRED88 , F.28
    After credit re-organisation, please check again and revert back.
    Also would like to ask you whether you have made any recent changes in the credit master data [FD32].
    Thanks & Regards,
    Hegal K Charles

  • RMAN backup in RAC failing with error ORA-19566

    Hi all, i have 2 node RAC and i am trying to take rman backup of the database which is in ASM filesystem.
    There is a corruption in SYSAUX file and rman is failing to take backup of the sysaux datafile and error thrown at the end like
    piece handle=/home/oracle/orabackup/BLOOM_LVL0_20120627_1qnejueh_s58_p1 tag=TAG20120627T185538 comment=NONE
    channel ORA_DISK_1: backup set complete, elapsed time: 00:00:56
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03009: failure of backup command on ORA_DISK_1 channel at 06/27/2012 18:56:17
    *ORA-19566: exceeded limit of 0 corrupt blocks for file +DATA/bloom/datafile/sysaux.257.775475317*
    Then i tried to validate the backup in rman of the datafile sysaux nad fired the following query
    SQL> SELECT * FROM v$database_block_corruption;
    FILE# BLOCK# BLOCKS CORRUPTION_CHANGE# CORRUPTIO
    3 32011 1 0 ALL ZERO
    3 32049 40 0 ALL ZERO
    3 32034 1 0 ALL ZERO
    3 28964 1 1399837 LOGICAL
    3 32015 2 0 ALL ZERO
    and found all the blocks in the table WRH$_SQL_PLAN.
    Then dropped the table, and tried to take backup of the database and it is failing again with the same error..
    ANY SOLUTION TO TAKE BACKUP OF THE DATABASE.
    anyhow we can SET MAXCORRUPT FOR DATAFILE 3 to value, but does this works all time...
    Thanks in advance,
    TEJAS

    Both are good suggestions, but
    Osama-mustafa wrote:
    RMAN> blockrecover datafile 5 block 2684; -- Fix Eror That ship has already sailed:
    TEJAS_DBA wrote:
    And Yes this is the first time i am taking backup on the database.And:
    Osama-mustafa wrote:
    Or Skip Option
    Like
    RMAN> RUN{
    2> SET MAXCORRUPT FOR DATAFILE 1,3 to 1;
    3> BACKUP DATABASE;
    4> }The OP was aware of this option and hopefully was able to get at least some sort of a backup eventually. But as for the dropped WRH$_SQL_PLAN, TEJAS is, I will use the word "braver" than most. And as often is the case, Dan Morgan offered the only bit of sensible advise.
    Open an SR.
    Good luck.

  • RMAN backup got failed with the error ORA-19566

    Hi,
    One of database, while taking backup using RMAN, it failed with the following error:
    RMAN-03009: failure of backup command on DISK4 channel at 11/28/2012 09:47:29
    ORA-19566: exceeded limit of 0 corrupt blocks for file <faile_name>
    I used DBV to identify the number of blocks corrupted.
    I referred the Metalink iD 457422.1 & 336133.1. But it did not help since the corrupted blocks have got object in it.
    I could able to find the object as well. it is one of the table in the datafile.
    I used 'ALTER TABLE ... MOVE..' command to see if that helps. But again getting the same error while taking the backup and corrupted blocks still exists.
    DB Version : 10.2.0.5
    OS Version : RHEL 5.6
    Please help me out to get this fixed.

    Here it is:
    SQL> Select * from v$database_block_corruption ;
    FILE# BLOCK# BLOCKS CORRUPTION_CHANGE# CORRUPTION_TYPE
    13 3988352 128 0 ALL ZERO

  • SOLR Commit (exceeded limit of maxWarmingSearchers=4)

    My understanding of the following error, "exceeded limit of maxWarmingSearchers=4", when indexing collections, is that I am committing too often. I do not pass in autoCommit="no", which to me means that it committed the changes each and every time cfindex was executed. I am looping over a large collection of information, cfindexing with each loop. If I do pass in autoCommit="no", when and how do I commit those changes? Can I place a "cfindex autoCommit='yes'" after the loop has finished to commit all those changes? What is the best way to do this?

    You should use "MAXCORRUPT" option in rman.
    For more details about this error and solutiones, please check:
    http://heliosguneserol.wordpress.com/2012/07/03/ora-19566-exceeded-limit-of-0-corrupt-blocks/
    http://arjudba.blogspot.com.es/2008/05/ora-19566-exceeded-limit-of-0-corrupt.html
    http://www.runningoracle.com/product_info.php?products_id=397

  • ORA-02393 Exceeded Call Limit on CPU Usage

    I have created a Profile and attached it to a user, in this example:
    Create Profile percall
    Limit
    CPU_PER_CALL 10
    IDLE_TIME 5;
    I have attached it to one user - USER1
    When USER1 runs a SQL Statement -
    SELECT COUNT(*) FROM TABLE1 A WHERE A.EFFDT = (SELECT MAX(B.EFFDT) WHERE B.EMPLID = A.EMPLID AND B.EFFDT <= SYSDATE);
    I get an error (Which I want to receive) ORA-02393 Exceeded Call Limit on CPU Usage.
    The SQL statement shows in the table DBA_COMMON_AUDIT_TRAIL, but shows a success even though the user received an error ORA-02393.
    What I want is a way for a DBA to be able to report on those ORA-02393 errors. I don't see any entries in the Log files, and don't notice any errors in the Oracle Tables.
    I would like to be able to show the user (after a week when they bring up the issue) what the SQL statement was and why it Exceeded the CPU Usage. If the error could place the SQL statement in a table or just display it in an error log with the Statement to verify that THIS is the statement which exceeded the CPU Usage.
    Thank you
    Aaron

    can you modify the procedure in which the SELECT resides.
    If so, trap & log the error.

  • ORA-02394: exceeded session limit on IO usage

    I have one sql took much time to run and got this..
    ORA-02394: exceeded session limit on IO usage, you are being logged of
    when I checked profile options
    SQL> select PROFILE,RESOURCE_NAME,LIMIT from dba_profiles where RESOURCE_NAME='LOGICAL_READS_PER_SESSION';
    PROFILE RESOURCE_NAME
    LIMIT
    DEFAULT LOGICAL_READS_PER_SESSION
    UNLIMITED
    SQL> select PROFILE,RESOURCE_NAME,LIMIT from dba_profiles where RESOURCE_NAME='LOGICAL_READS_PER_CALL';
    PROFILE RESOURCE_NAME
    LIMIT
    DEFAULT LOGICAL_READS_PER_CALL
    UNLIMITED
    Is there anything we can do here bcaz both profiles have been set to unlimted?

    Hello,
    Oracle 8.1.7.4 is the final patchset of the last release for Oracle 8i, so it's a rather
    stable version.
    Do you have a way to tune this query so that it can run faster ?
    Do you have correct statistics on the optimizer ?
    Best regards,
    Jean-Valentin

  • Message Content Exceeds Limit

    I've had an error message: Message content exceeds limit while trying to send a 2mb video by mms.  What am I doing wrong?  I've sent them before and they've gone thru.  HELP Please!
    Solved!
    Go to Solution.

    Hi and Welcome to the Forums!
    Since MMS is a voice-level service from your carrier, you should contact them for assistance.
    Good luck and let us know!
    Occam's Razor nearly always applies when troubleshooting technology issues!
    If anyone has been helpful to you, please show your appreciation by clicking the button inside of their post. Please click here and read, along with the threads to which it links, for helpful information to guide you as you proceed. I always recommend that you treat your BlackBerry like any other computing device, including using a regular backup schedule...click here for an article with instructions.
    Join our BBM Channels
    BSCF General Channel
    PIN: C0001B7B4   Display/Scan Bar Code
    Knowledge Base Updates
    PIN: C0005A9AA   Display/Scan Bar Code

  • ORA-02399: exceeded maximum connect time, you are being logged off

    What could have been the cause of "ORA-02399: exceeded maximum connect time, you are being logged off"? I am always getting that issue even if I am continously doing some query.

    There's a resource limit active (RESOURCE_LIMIT=TRUE in pfile/spfile) and a user_profile (which limits the connection time) dedicated to the user.

  • My iTunes Match won't load past step 1; All songs in my library say "Exceeded Limit", even the ones that were uploaded before.

    My iTunes Match hasn't been working for a couple of days. It said I had exceeded the song limit, so I was going through my library and clearing out some songs. When I deleted the songs, the box that asked if I wanted to delete them from the cloud wasn't showing up. I decided to update my iTunes Match before deleting anything else. When I attempted the update, Step 1 moved incredibly slowly, then stopped working before reaching Step 2. When this has happened in the past, I have been advised to try several troubleshooting tactics from the support forum and from writing to Apple support in the past. The first one is to turn off iTunes Match, restart the computer, and turn on iTunes Match again. I did that. The same problem occurred. I tried the second trouble shooting tactic, which was to turn off iTunes Match, close iTunes, find the iTunes Library Genius.itdb and iTunes Library Extras.itdb and move them to the trash, then run a Repair Disk Permissions using Disk Utility, then turn iTunes Match back on. I tried that. The same problem occurred. I went to my iTunes library to see the iCloud status of my songs, and every single song was listed as not being in the cloud yet and had the "Exceeded Limit" status, even the songs that have been in my cloud for over a year. I checked a few of the forum topics here on the support site. One suggested opening up iTunes while holding down the Option button, creating a new library, then turning on iTunes Match and deleting songs from the cloud that way. Then open up the original library and turn on iTunes Match. I did that. The status of my songs hasn't changed, nor can iTunes Match load past Step 1. Does anyone know what I can do to fix this?

    Maybe I didn't explain it properly, but that's one of the things I described trying in my first post. It helped get songs out of my cloud that I didn't need there, but iTunes Match still couldn't get past the first step. Last night I deleted the iTunes Library Extras.itdb, iTunes library Genius.itdb, iTunes Library.it, and iTunes Library.xml files from my computer, entered Time Machine, and recovered those same files that had been backed up after the last time I added music (the time the problem began.) After recovering those files I opened iTunes and turned Match back on. For some reason, it worked, even though those were the files that existed when the problem initially occurred.The combination of having songs removed from the cloud as well as restoring those original files must have done something, but I have no idea why that fixed the problem.

Maybe you are looking for

  • Corrupted files on External HD - crashing iPhoto6

    Hi I've been using a Maxtor One Touch II external FW HD for my iPhoto library. It had been working fine until I tried to download photos yesterday onto it. I tried to rebuild the library, but it kept hanging. The only way to shut down was to hold dow

  • Not a valid SOAP Content-Type: text/html; charset=iso-8859-1

    Friends JDEV and SOA suite 10134 I have multiple domains on my BPEL Server. In one of the domain since I deployed the new process, all the processes of that domain are now failing on execution with following error in opmn/soa_instance/*.err log files

  • How do I get video to IMovie from my dvd camcorder?

    How do I get video from my Canon DVD  camcorder to Imovie?

  • Field for invoice number

    hi all    can any body say wht is the relevent R/3 field  and corresponding infoobject for "INVOICE NUMBER"     thank you regards Pradeep

  • Hierachy Load error from R/3

    Hi All, I got an error while loading hierarchy from R/3 to BW. error is <i><b>Diagnosis</i>     The hierarchy level of the parent node must be one larger than the     hierarchy level of the actual node. This is not the case for the node     with ID 0