SQL Logs Size Increasing automatically

Hi,
I am facing a very strange issue in my SQL Server 2008 R2 Logs folder.
Actually in every 10 seconds SQLDump0000,SQLDump0001,.....     and so on named files are being created automatically in the Logs folder where SQL stores it's logs.I don't know why is this happening.
Due to this issue logs folder size is increased to 160GB and  my c drive where windows is installed is keep showing Low space message in fact 0MB space is showing.
Please help urgently.

Hi,
I am facing a very strange issue in my SQL Server 2008 R2 Logs folder.
Actually in every 10 seconds SQLDump0000,SQLDump0001,.....     and so on named files are being created automatically in the Logs folder where SQL stores it's logs.I don't know why is this happening.
Due to this issue logs folder size is increased to 160GB and  my c drive where windows is installed is keep showing Low space message in fact 0MB space is showing.
Please help urgently.
Hello Zubair,
This dumps are getting created due to some issue SQL server is facing.Its not a SQL Server transaction log file dump.I guess your system is not updated to latest Service pack.
Latest Service Pack for SQL server 2008 R2 is SP2.Apply this SP and see if this dump generation subsides.If not you need to raise a case with Microsoft to get these dumps analyzed.
Please mark this reply as the answer or vote as helpful, as appropriate, to make it useful for other readers

Similar Messages

  • Weblogic 8.1 Server log size increase in Production environment

    Hi,
    Issue:: One of the log file is increasing in size and exceeding beyond the size mentioned in the configuration file resulting in application outage.
    Issue description:
    We are having problems with the log size in the Weblogic 8.1 server. The fileminsize has been mentioned in the config.xml.
    New log files like MYsvr.log00001,MYsvr.log00002, MYsvr.log00003, MYsvr.log00004 etc are also being generated appropriately when the max file size has been reached. But simultaneously, one of the files is growing in size, exceeding the limit mentioned in the configuration file. Eg.. the MYsvr.log00001 file is 800MB in size while the other files(MYsvr.log00002, MYsvr.log00003 etc are 10MB in size)
    This increase in size of the log has been resulting in an application outage.
    More Details:
    1. Server: BEA Weblogic 8.1 server
    2. Log size is fine in other environements. This is a problem only in the production environment.
    3. The entry in the config.xml is as follows:
    <Server ListenPort="6313" Name="MYsvr" NativeIOEnabled="true" TransactionLogFilePrefix="./">
    <ServerStart Name="MYsvr"/>
    <Log FileMinSize="10000" FileName="MYsvr.log" Name="MYsvr"
    NumberOfFilesLimited="true" RotationType="bySize"/>
    <SSL Name="MYsvr"/>
    <ServerDebug Name="MYsvr"/>
    <WebServer Name="MYsvr"/>
    <ExecuteQueue Name="default" ThreadCount="15"/>
    <KernelDebug Name="MYsvr"/>
    </Server>
    Could you please help with this issue ?
    Thank you.

    Can someone please provide a solution for the issue

  • The font size increases automatically and the window will not let me scroll

    When I launch 3.6.12, two things happen: The font size increases by itself to the largest font available, and the scroll bar doesn't work. I can scroll down the page, but as soon as I stop, the window resets to the top of the page.
    Running Win XP SP 3.

    Check that you do not run Firefox in compatibility mode or with a reduced screen resolution.
    Make sure that you do not run Firefox in compatibility mode.
    You can check that in the "Compatibility" tab in the Properties via the right-click context menu of the Firefox desktop shortcut.
    Make sure that all items are deselected in the "Compatibility" tab of the Properties window.
    http://www.microsoft.com/windowsxp/using/helpandsupport/learnmore/appcompat.mspx

  • Slow inserts after log size increase

    I have an app that does 10k-50k consecutive SELECTs and INSERTs as part of a loading operation. Recently, I enlarged the redo logs from 10M to 100M, which appears about right, based on the frequency of log switches (half a dozen/day). Trouble is, now the load takes 5 times longer (20-25 min vs. 5). Any thoughts? I thought redo logs that were too small caused poor performance, b/c of increased checkpoint activity, but I can't imagine why larger redo logs would slow inserts down.
    Rob

    Logically the only difference a larger online redo log should make is that the number of checkpoints invoked due to log switches should decrease. But a larger online redo log will take longer to archive.
    Where did you place the larger online redo log and where does the archive redo log go to? Is this performance decrease consistent or does it appear when certain redo logs are in use?
    Even if your logical disk are different it could be the physical disk for some of your files are now the same where before this was not true.
    The suggested statspack is probably a good place to start but you might also want to take some OS IO stats during the run also.
    HTH -- Mark D Powell --

  • SQLSERVER 2012 LOG SIZE INCREASE EXTENSIVELY

    Hi Friends,
    I want to inquire, as our sqlserver 2012 in full  recovery mode , I notices whenever we are executing rebuild indexes maintenancejob the logfile grown up extensively every time.  h
    thank you.
    regards,
    asad

    Hi Friends,
    I want to inquire, as our sqlserver 2012 in full  recovery mode , I notices whenever we are executing rebuild indexes maintenancejob the logfile grown up extensively every time.  h
    Asad,
    Why are you posting same thread about SQL Server log file every time. I can see all your previous threads on same Log file issue.
    This clearly shows that you don't spend event single moment reading links posted as answer. Why dont you go and read about SQL Server before asking *almost* same question every time.
    Yes log file will grow that is default behavior why are you executing index rebuild for all index. Did you searched about index rebuild and how it works and what gets logged.
    You can read here
    Curious case of logging in Online and offline index rebuild
    Use Ola Hallengren solution for index rebuild it will only rebuild index that is fragmented.
    Don't just blindly post question spend some time on net searching I am sure you would find lot of articles
    Please mark this reply as answer if it solved your issue or vote as helpful if it helped so that other forum members can benefit from it
    My Technet Wiki Article
    MVP

  • SQL LOG FILE SIZE INCREASING

    Hi DBA's
    SQL Log file size occupies more disk space on the server, the overall database size is 8GB
    How to decrease the SQL LDF file size on the server, please explain the suitable steps to perform
    Thanks
    DBA

    use master
    go
    dump transaction <YourDBName>
    with no_log
    go
    use <YourDBName>
    go
    DBCC SHRINKFILE (<YourDBNameLogFileName>,
    100) -- where 100 is the size you may want to shrink it to in MB, change it to your needs
    go
    -- then you can call to check that all went fine
    dbcc checkdb(<YourDBName>)
    Andy ,
    what point in asking user to use No_log and you did not even motioned about what this eveil command will do. Actually its
    seriously not required reason being initial size of log file set to 8 G.
    Plus what is point in running checkdb ?
    I don't agree to any part you pointed
    Please mark this reply as answer if it solved your issue or vote as helpful if it helped so that other forum members can benefit from it
    My Technet Wiki Article
    MVP

  • SQL server error log size is too big to handle

    I am working with a large database on windows sql server 2008 R2 such that it has to run continuously 24x7 because of that it is not possible
    to restart the server time to time. It is kind of monitoring system for big machines. Because of this SQL server error logs are growing too big even some times up to 60-70 GB at a limited sized hard drive. I can't delete them time to time manually. Can someone
    please suggest a way using which I can stop creation of such error logs or recycle them after sometime. Most of the errors are of this kind --
    Setting database option RECOVERY to simple for database db_name
    P.S.- I have read limiting error logs to 6 etc. But that didn't help. It will be best if you could suggest some method to disable these logs.

    Hi Mohit11,
    According to your description, your SQL Server error logs are growing too big to handle at a limited sized hard drive, and you want to know how to stop the generation of such error logs or recycle them after sometime automatically without restarting the
    SQL Server, right?
    As others mentioned above, we may not be able to disable SQL server error log generation. However we can recycle the error logs automatically by running the
    sp_cycle_errorlog on a fixed schedule (i.e. every two weeks) using SQL agent jobs so that the error logs will be recycled
    automatically without restarting SQL Server.
    And it is also very important for us to keep the error log files more readable. So we can increase the number of error logs a little more and run the sp_cycle_errorlog more frequently (i.e. daily), then each file will in a smaller size to be more readable
    and we can recycle the log files automatically.
    In addition, in order to avoid the size of all the log files growing into a too big size unexpected (sometime it may happen), we can run the following query in SQL Agent job to automatically delete all the old log files when the size of log files is larger
    than some value we want to keep (i.e. 30GB):
    --create a tample table to gather the information of error log files
    CREATE TABLE #ErrorLog
    Archieve INT,
    Dt DATETIME,
    FileSize INT
    GO
    INSERT INTO #ErrorLog
    EXEC xp_enumerrorlogs
    GO
    --delete all the old log files if the size of all the log files is larger than 30GB
    DECLARE @i int = 1;
    DECLARE @Log_number int;
    DECLARE @Log_Max_Size int = 30*1024; --here is the max size (M) of all the error log files we want to keep, change the value according to your requirement
    DECLARE @SQLSTR VARCHAR(1000);
    SET @Log_number = (SELECT COUNT(*) FROM #ErrorLog);
    IF (SELECT COUNT(FileSize/1024/1024) FROM #ErrorLog) >= @Log_Max_Size
    BEGIN
    WHILE @i <= @Log_number
    BEGIN
    SET @SQLSTR = 'DEL C:\Program Files\Microsoft SQL Server\MSSQL11.MSSQLSERVER\MSSQL\Log\ERRORLOG.' + CONVERT(VARCHAR,@i);
    EXEC xp_cmdshell @SQLSTR;
    SET @i =@i + 1;
    END
    END
    DROP TABLE #ErrorLog
    For more information about How to manage the SQL Server error log, please refer to the following article:
    http://support.microsoft.com/kb/2199578
    If you have any question, please feel free to let me know.
    Regards,
    Jerry Li

  • How can U decrease the Log Size in MS-SQL server?

    Hello friends,
                    I am in Big Trouble. I have a Problem.we are Using MS-SQL server 8.00.924. The Database Log size is almost Full.But BackUps are not Done successfully ,,How to Shink the Log file,, Any body pls help me I will give more Points.
    Regards,
    Balaram

    Hi Balaram,
    Please see the link and scripts which is given hereunder.
    http://support.microsoft.com/kb/873235
    You can check this method also.
    Method 1 – Delete the Log (LDF) File
    Although this method is simple, it is not supported by Microsoft®.  Additionally, it may not work for all systems.  Follow these steps to try the simple method on your system.
    Ensure that no users are logged into the database.
    Open Query Analyzer or another program to execute SQL scripts.  Ensure that “dsdbase” is not selected in the drop-down “DB” list.
    Open and run the script DETACH.sql.  This script detaches the DataLyzer® database from the SQL server database.
    Make a backup copy of the DSdbase_Data.MDF and DSdbase_Log files.
    Rename the DSdbase_log file to another name, for instance backup_DSdbase_log.
    Open and run the script ATTACHMDF.sql.  This script re-attaches the DataLyzer® database to the SQL database without attaching the log file.  NOTE: If your database is not stored in the location c:mssql7data change the SQL script to reflect the location of your DSdbase file.
    If the script runs successfully, a new DSdbase_Log file will automatically be created to the original size specified when creating the database.  It is suggested that users keep the backup copies made of their database and log files for several days to ensure an available backup if anything happens to the new log file.
    If the script does not run successfully, method 1 will not work for your current DataLyzer® Spectrum database.  This is because your database’s structure is dependent on the logfile.  In this case, re-name the original log file back to DSdbase_log.  Run the script ATTACH.sql and again change the file location to the location of your DataLyzer® database, if necessary.  When the database is successfully re-attached, try using method 2 to shrink the log file.
    Method 2 – Truncate and Shrink the Log (LDF) File
    Although this method of shrinking the log file is more complicated, it is supported by Microsoft®, and it will work for almost all systems.  Follow these steps:
    Ensure no users are logged into the database.
    Open Query Analyzer or another program to execute SQL scripts.  Ensure that “dsdbase” is not selected in the drop-down “DB” list.
    If you have already tried method 1, or if you have a current backup of you database and log file, skip to step 7
    Open and run the script DETACH.sql.  This script detaches the DataLyzer® database from the SQL server database.
    Make a backup copy of the DSdbase_Data.MDF and DSdbase_Log files.
    Open and run the script ATTACHMDF.sql.  This script re-attaches the DataLyzer® database to the SQL database.  NOTE: If your database is not stored in the location c:mssql7data, change the SQL script to reflect the location of your DSdbase file.
    Select “dsdbase” from the drop down “DB” list in the program you are using to execute the SQL scripts.
    8.      Run VIEW2LDF.sql.  Below is an excerpt similar to what will be displayed on your screen: FileId   FileSize    StartOffset    FSeqNo      Status   Parity  CreateTime
    2          1245184   819              1292   0          64         1999-12-15 10:25:13.597
    2          1245184   1253376       1225    0          128        1999-11-19 09:12:57.743
    2           1245184   2498560       1224    0          128     1999-11-19 09:12:23.82
    2          1499136   3743744       1318    2          64         1999-12-15 10:25:36.880
    Locate the “2” in the status column.  The location of the “2” determines where the active portion of the log is.
    If the 2 is in the bottom half of the log, open and run the SQL script MOVE2LOOP.sql.  This script is an infinite loop used to move the active portion of the log.
    Re-Run the VIEW2LDF.sql script to make sure that the 2 is now located in the first half of the script. 
    Run the script SHRINKLOG.sql to reduce the size of the database.
    Run the script VIEW2LDF.sql to view the new status of the log file.  You should notice that the number of lines is significantly smaller than it originally was.
    If you wish to further reduce the size of the database, repeat steps 8 through 13.
    Text Context of SQL Scripts
    ATTACHMDF.sql
    sp_attach_db @dbname=N'dsdbase', @filename1=N'c:mssql7datadsdbase_data.mdf'
    ATTACH.sql
    sp_attach_db @dbname=N'dsdbase', @filename1=N'c:mssql7datadsdbase_data.mdf',
    @filename2=N'c:mssql7datadsdbase_log.ldf'
    Detach.sql
    sp_detach_db dsdbase
    MOVE2LOOP.sql
    USE DSDBASE
    GO
    CREATE TABLE VIRTUAL_LOG_FILE_WRAPPER (CHAR1 CHAR(4000))
    GO
    DECLARE @I INT
    SELECT @I = 0
    WHILE (1=1)
    BEGIN
    WHILE (@I < 100)
    BEGIN
    INSERT VIRTUAL_LOG_FILE_WRAPPER VALUES ('A')
    SELECT @I = @I + 1
    END
    TRUNCATE TABLE VIRTUAL_LOG_FILE_WRAPPER
    BACKUP LOG DSDBASE WITH TRUNCATE_ONLY
    END
    SHRINKLOG.sql
    DBCC SHRINKFILE (DSDBASE_LOG, TRUNCATEONLY)
    BACKUP LOG DSDBASE WITH TRUNCATE_ONLY
    VIEW2LDF.sql
    DBCC LOGINFO(DSDBASE)
    Regards,
    Anil

  • How to Increase database / log size of mirrored database

    Hi All,
    I have been asked to increase the default database and log size of two databases, both of which are mirrored.
    I have been trying to find out what else I need to do in this case other than the standard resizing.
    Are there extra steps I need to perform as these dbs are mirrored (SQL Mirroring)?
    Farren

    Hi Farren,
    Based on my test, if you choose to expand the mirrored database by increasing the size of an existing data or log file, you can directly execute the ALTER DATABASE statement on the Principal server as Uri’s post, but before that please make sure the mirroring
    status is synchronized. You can check the mirroring configuration and partner status with the following query.
    SELECT DB_NAME(DATABASE_ID) 'DBNAME',mirroring_state,
    mirroring_state_desc,mirroring_role,mirroring_role_desc,
    mirroring_partner_name,mirroring_partner_instance
    FROM sys.database_mirroring
    WHERE mirroring_guid IS NOT NULL
    However, if you choose to expand the mirrored database by adding a new file to the database, you need to perform extra steps as follows.
       1. Check the mirroring configuration and partner status with above query.
       2. Run the below command on the principal server to break the mirror.  
    ALTER DATABASE DatabaseName SET PARTNER OFF
       3. Create your database file on the principal server.
       4. Run a log backup on the principal server.
       5. Restore this log backup on the mirrored server using the NORECOVERY and MOVE options.
       6. Re-establish mirroring between the servers for this database.
    Reference:
    Increase the Size of a Database
    How to add a database file to a mirrored SQL Server database
    Thanks,
    Lydia Zhang
    Lydia Zhang
    TechNet Community Support

  • The size of folders increase automatically bcause of which I'm unable to work on iPhone 5

    The size of folders increase automatically bcause of which I'm unable to work on iPhone 5.

    We will need more information in order to help you. What folders are you talking about since the phone does not support a standard computer file system? Each app holds its individual files. What can't you do on your phone?

  • Why size of archive log file increasing in merge clause

    my database is running in archive log mode.
    someone is running oracle merge statement. still it is running.
    He will issue commit after the operation.
    in that period redolog file increasing now.
    my question is why size of archive log file increasing with redolog file.
    i know that after commit archive log file should generate.(may be it is wrong).
    please suggest........
    Edited by: 855516 on Mar 13, 2012 11:18 AM

    855516 wrote:
    my database is running in archive log mode.
    someone is running oracle merge statement. still it is running.
    He will issue commit after the operation.
    in that period redolog file increasing now.
    my question is why size of archive log file increasing with redolog file.
    i know that after commit archive log file should generate.(may be it is wrong). No this is not correct that after commit archive log will generate....You know merge statement causes the insert (if data not present already) or update if database is present.. Obviously these operation will generate lots of redo if the amount of data been processed is high.
    If you feel that this operation is causing excessive of redo then root cause analysis should be done...
    For that use Logminer (excellent tool to provide segment level breakdown of redo size). V$logmnr_contens has columns redo block and redo byte address associated with the current redo
    change
    There are some gudlines in order to reduce redos( which may vary in any environment)
    1) check if there are unwanted indexes being used in tables which are refereed in merge. If yes then remove those could bring down the redo
    2) Use global temporary tables to reduce redo (if there is a need to keep data only temporarily in a session)
    3) Use nologging if possible (but see its implications)
    Hope this helps

  • SSIS 2008 Rebuild Index Task increasing Log Size

    I am testing SSIS 2008 Rebuild Index Task on a single database (db1). I shrank the
    db1's log file to initial size. I also checked the box Sort results in tempdb on the SSIS package.
    However, when I run the package, db1's log file size increased about 55 times the original size.
    When I run a rebuild of all the index on db1
    with (sort_in_tempdb=ON) ; there was only a slight increase in the log file size ( did not even double the initial size).
    Is this a SSIS bug? The check box is not actually sorting in tempdb?

    I am testing SSIS 2008 Rebuild Index Task on a single database (db1). I shrank the
    db1's log file to initial size. I also checked the box Sort results in tempdb on the SSIS package.
    However, when I run the package, db1's log file size increased about 55 times the original size.
    When I run a rebuild of all the index on db1
    with (sort_in_tempdb=ON) ; there was only a slight increase in the log file size ( did not even double the initial size).
    Is this a SSIS bug? The check box is not actually sorting in tempdb?
    Arthur can you please move this thread in Database engine forum. IMO its what changed from 2008 .Index rebuild is fully logged from 2008 onwards it was previously (in 2005) minimally logged .Refer below link for information regarding same
    http://support.microsoft.com/kb/2407439/en-gb
    Now about when sort intempdb is used.The intermediate sort results that are used to build the index are stored in tempdb .When you rebuild without sort_in_tempdb ( i guess your data and log file are on same drive) index will utilize disk space on which it
    resides so it might seem to you log has increased.Is it so ,am i correct.
    What query you used to measure log file size,are you absolutely sure it was log file that increased
    Please mark this reply as the answer or vote as helpful, as appropriate, to make it useful for other readers

  • I am handling logistics department in a company, i am handling more than 100 calls in a day. But my iPhone 4 have only 100 number call history. How to increase my call log size or kindly suggest a better app for storing 1 month call history..

    I am handling logistics department in a company, i am handling more than 100 calls in a day. But my iPhone 4 have only 100 number call history. How to increase my call log size or kindly suggest a better app for storing 1 month call history..

    Here's one:
    https://itunes.apple.com/us/app/callog/id327883585?mt=8

  • Update Multiple Columns when concerned about redo/undo log sizes.

    Hi ,
    I have update statements that updates multiple columns at once if any of them is changed. What I see that even though the value of column is not changed it still increases the redo size.
    Below is a sample code similar to the ones in my code. Basically I check whether there is a difference in any of the columns to be updated and update all of them.
    Is there a way to improve redo log size without splitting the update statement for every column that I will be updating. Redo/Undo log size is a concern for us..
      For i In 1.rec.Count Loop
        Update employees e
           Set e.first_name = rec(i).first_name, e.last_name = rec(i).last_name
         Where e.first_name != rec(i).first_name
            Or e.last_name != rec(i).last_name;
      End Loop;My database is 10g.

    Muhammed Soyer wrote:
    Redo/Undo log size is a concern for us..You are worried about the wrong thing.
    If you are concerned about the amount of undo and redo, you should be less concerned about the small diffrence between updating 1 or 3 columns and remove the loop that is contributing to a massive increase in both undo and redo.
    Re: global temporary table row order
    Name                                  Run1        Run2        Diff
    STAT...undo change vector size     240,500   6,802,736   6,562,236
    STAT...redo size                 1,566,136  24,504,020  22,937,884Run2 shows what adding a loop to a regular SQL statement will do to undo and redo. It made the redo used 15 times greater and the undo almost 30 times greater.

  • Increse of redo log size

    Hi,
    Please help me to increase the redo log size.
    As i am in DB - oracle 10G and OS - Suse linux 10SP2
    SQL> SELECT * FROM v$log;
        GROUP#    THREAD#  SEQUENCE#      BYTES    MEMBERS ARC STATUS
    FIRST_CHANGE# FIRST_TIME
             1          1        358  157286400          2 YES INACTIVE
       2972903289 28-NOV-11
             2          1        359  157286400          2 YES INACTIVE
       2972957401 28-NOV-11
             3          1        357  157286400          2 YES INACTIVE
       2972839164 27-NOV-11
        GROUP#    THREAD#  SEQUENCE#      BYTES    MEMBERS ARC STATUS
    FIRST_CHANGE# FIRST_TIME
             4          1        360  157286400          2 NO  CURRENT
       2973005629 29-NOV-11
    SQL> SELECT * FROM v$logfile;
        GROUP# STATUS  TYPE
    MEMBER
    IS_
             4         ONLINE
    /oracle/JID/origlogB/log_g14m1.dbf
    NO
             4         ONLINE
    /oracle/JID/mirrlogB/log_g14m2.dbf
    NO
        GROUP# STATUS  TYPE
    MEMBER
    IS_
             3         ONLINE
    /oracle/JID/origlogA/log_g13m1.dbf
    NO
             3         ONLINE
    /oracle/JID/mirrlogA/log_g13m2.dbf
        GROUP# STATUS  TYPE
    MEMBER
    IS_
    NO
             2         ONLINE
    /oracle/JID/origlogB/log_g12m1.dbf
    NO
             2         ONLINE
        GROUP# STATUS  TYPE
    MEMBER
    IS_
    /oracle/JID/mirrlogB/log_g12m2.dbf
    NO
             1         ONLINE
    /oracle/JID/origlogA/log_g11m1.dbf
    NO
        GROUP# STATUS  TYPE
    MEMBER
    IS_
             1         ONLINE
    /oracle/JID/mirrlogA/log_g11m2.dbf
    NO
    8 rows selected.
    Please help me how to execute for the above query.
    Thanks,
    Hariharan

    Hello
    Complete step:
    Step 1 SQL> select a.group#, a.member, b.bytes/1024/1024 mb from v$logfile a, v$log b where a.group# = b.group#;
    This query will show current group with redo log members and their size.
    Step 2 Make the last redo log CURRENT one
    To find which group is current at this moment use following query
    SQL> select group#, status from v$log;
    GROUP#         STATUS
    1                    CURRENT
    2                    INACTIVE
    3                    INACTIVE
    4                    INACTIVE
    Now as you can see that the first group is marked as current but we need to make group 4 as current. So force group 4 to become current one by switching log file. To switch log file use following query.
    SQL> alter system switch logfile;
    GROUP#         STATUS
    1                     INACTIVE
    2                    CURRENT
    3                    INACTIVE
    4                  INACTIVE
    SQL> alter system switch logfile;
    GROUP#        STATUS
    1                     INACTIVE
    2                     INACTIVE
    3                     INACTIVE
    4                    CURRENT
    Step 3 Drop the first online redo log
    After making the last online redo log file the CURRENT one, drop the first online redo log:
    SQL> alter database drop logfile group 1;
    Database altered.
    Note:
    Be aware that if you are going to drop a logfile group, it cannot be the current logfile group. However, where attempting to drop the logfile group resulted in the following error as a result of the logfile group having an active status:
    SQL> ALTER DATABASE DROP LOGFILE GROUP 1;
    ALTER DATABASE DROP LOGFILE GROUP 1
    ERROR at line 1:
    ORA-01624: log 1 needed for crash recovery of instance ORA920 (thread 1)
    ORA-00312: online log 1 thread 1: ''
    Easy problem to resolve. Simply perform a checkpoint on the database:
    SQL> ALTER SYSTEM CHECKPOINT GLOBAL;
    System altered.
    SQL> ALTER DATABASE DROP LOGFILE GROUP 1;
    Database altered.
    Step 4 You need to re-create dropped online redo log group with different size. Use the following query to achieve this.
    SQL> alter database add logfile group 1 ('<path>/origlogA/log_g11m1.dbf','<path>/mirrlogA/log_g11m2.dbf') size 200M reuse;
    Database altered.
    Step 5 Force another log switch
    After re-creating the online redo log group, force a log switch. The online redo log group just created should become the "CURRENT" group:
    SQL> select group#, status from v$log;
    GROUP#         STATUS
    1                     UNUSED
    2                     INACTIVE
    3                     INACTIVE
    4                    CURRENT
    SQL> alter system switch logfile;
    SQL> select group#, status from v$log;
    GROUP#         STATUS
    1                      CURRENT
    2                      INACTIVE
    3                      ACTIVE
    Step 6 # Loop back to Step 3 until all logs are rebuilt
    After re-creating an online redo log group, continue to re-create (or resize) all online redo log groups until all of them are rebuilt.
    Regards,
    Rajan

Maybe you are looking for