ORA-00349: failure obtaining block size for '+Z'  in Oracle XE

Hello,
I am attempting to move the online redo log files to a new flash recovery area location created on network drive "Z" ( Oracle Database 10g Express Edition Release 10.2.0.1.0).
When I run @?/sqlplus/admin/movelogs; in SQL*Plus as a local sysdba, I get the following errors:
ERROR at line 1:
ORA-00349: failure obtaining block size for '+Z'
ORA-06512: at line 14
Please let me know how to go about resolving this issue.
Thank you.
See below for detail:
Connected.
SQL> @?/sqlplus/admin/movelogs;
SQL> Rem
SQL> Rem $Header: movelogs.sql 19-jan-2006.00:23:11 banand Exp $
SQL> Rem
SQL> Rem movelogs.sql
SQL> Rem
SQL> Rem Copyright (c) 2006, Oracle. All rights reserved.
SQL> Rem
SQL> Rem NAME
SQL> Rem movelogs.sql - move online logs to new Flash Recovery Area
SQL> Rem
SQL> Rem DESCRIPTION
SQL> Rem This script can be used to move online logs from old online
log
SQL> Rem location to Flash Recovery Area. It assumes that the database
SQL> Rem instance is started with new Flash Recovery Area location.
SQL> Rem
SQL> Rem NOTES
SQL> Rem For use to rename online logs after moving Flash Recovery
Area.
SQL> Rem The script can be executed using following command
SQL> Rem sqlplus '/ as sysdba' @movelogs.sql
SQL> Rem
SQL> Rem MODIFIED (MM/DD/YY)
SQL> Rem banand 01/19/06 - Created
SQL> Rem
SQL>
SQL> SET ECHO ON
SQL> SET FEEDBACK 1
SQL> SET NUMWIDTH 10
SQL> SET LINESIZE 80
SQL> SET TRIMSPOOL ON
SQL> SET TAB OFF
SQL> SET PAGESIZE 100
SQL> declare
2 cursor rlc is
3 select group# grp, thread# thr, bytes/1024 bytes_k
4 from v$log
5 order by 1;
6 stmt varchar2(2048);
7 swtstmt varchar2(1024) := 'alter system switch logfile';
8 ckpstmt varchar2(1024) := 'alter system checkpoint global';
9 begin
10 for rlcRec in rlc loop
11 stmt := 'alter database add logfile thread ' ||
12 rlcRec.thr || ' size ' ||
13 rlcRec.bytes_k || 'K';
14 execute immediate stmt;
15 begin
16 stmt := 'alter database drop logfile group ' || rlcRec.grp;
17 execute immediate stmt;
18 exception
19 when others then
20 execute immediate swtstmt;
21 execute immediate ckpstmt;
22 execute immediate stmt;
23 end;
24 execute immediate swtstmt;
25 end loop;
26 end;
27 /
declare
ERROR at line 1:
ORA-00349: failure obtaining block size for '+Z'
ORA-06512: at line 14
Can someone point me in the right direction as to what I may be doing wrong here - Thank you!

888442 wrote:
I am trying to drop and recreate ONLINE redo logs on my STANDB DATABASE (11.1.0.7)., but i am getting the below error.
On primary, we have done the changes., ie we added new logfile with bigger size and 3 members. When trying to do the same on Standby we are getting this error.
Our database is in Active DG Read only mode and the oracle version is 11.1.0.7.
I have deffered the log apply and cancelled the managed recovery, and dg is in manual mode.
SQL> alter database Add LOGFILE GROUP 4 ('+DT_DG1','+DT_DG2','+DT_DG3') SIZE 1024M;
alter database Add LOGFILE GROUP 4 ('+DT_DG1','+DT_DG2','+DT_DG3') SIZE 1024M
ERROR at line 1:
ORA-00349: failure obtaining block size for '+DT_DG1'First why you are dropping & recreating online redo log files on standby.
On standby only standby redo log files will be used. Not sure what you are trying to do.
here is example how to create online redo log files, Check that diskgroup is mounted and have sufficient space to create.
sys@ORCL> select member from v$logfile;
MEMBER
C:\ORACLE\ORADATA\ORCL\REDO03.LOG
C:\ORACLE\ORADATA\ORCL\REDO02.LOG
C:\ORACLE\ORADATA\ORCL\REDO01.LOG
sys@ORCL> alter database add logfile group 4 (
  2     'C:\ORACLE\ORADATA\ORCL\redo_g01a.log',
  3     'C:\ORACLE\ORADATA\ORCL\redo_g01b.log',
  4     'C:\ORACLE\ORADATA\ORCL\redo_g01c.log') size 10m;
Database altered.
sys@ORCL> select member from v$logfile;
MEMBER
C:\ORACLE\ORADATA\ORCL\REDO03.LOG
C:\ORACLE\ORADATA\ORCL\REDO02.LOG
C:\ORACLE\ORADATA\ORCL\REDO01.LOG
C:\ORACLE\ORADATA\ORCL\REDO_G01A.LOG
C:\ORACLE\ORADATA\ORCL\REDO_G01B.LOG
C:\ORACLE\ORADATA\ORCL\REDO_G01C.LOG
6 rows selected.
sys@ORCL>
Your profile:-
888442      
     Newbie
Handle:      888442
Status Level:      Newbie
Registered:      Sep 29, 2011
Total Posts:      12
Total Questions:      8 (7 unresolved)
Close the threads if answered, Keep the forum clean.

Similar Messages

  • ORA-00349: failure obtaining block size

    I am trying to drop and recreate ONLINE redo logs on my STANDB DATABASE (11.1.0.7)., but i am getting the below error.
    On primary, we have done the changes., ie we added new logfile with bigger size and 3 members. When trying to do the same on Standby we are getting this error.
    Our database is in Active DG Read only mode and the oracle version is 11.1.0.7.
    I have deffered the log apply and cancelled the managed recovery, and dg is in manual mode.
    SQL> alter database Add LOGFILE GROUP 1 ('+DT_DG1','+DT_DG2','+DT_DG3') SIZE 1024M;
    alter database Add LOGFILE GROUP 1 ('+DT_DG1','+DT_DG2','+DT_DG3') SIZE 1024M
    ERROR at line 1:
    ORA-00349: failure obtaining block size for '+DT_DG1'

    888442 wrote:
    I am trying to drop and recreate ONLINE redo logs on my STANDB DATABASE (11.1.0.7)., but i am getting the below error.
    On primary, we have done the changes., ie we added new logfile with bigger size and 3 members. When trying to do the same on Standby we are getting this error.
    Our database is in Active DG Read only mode and the oracle version is 11.1.0.7.
    I have deffered the log apply and cancelled the managed recovery, and dg is in manual mode.
    SQL> alter database Add LOGFILE GROUP 4 ('+DT_DG1','+DT_DG2','+DT_DG3') SIZE 1024M;
    alter database Add LOGFILE GROUP 4 ('+DT_DG1','+DT_DG2','+DT_DG3') SIZE 1024M
    ERROR at line 1:
    ORA-00349: failure obtaining block size for '+DT_DG1'First why you are dropping & recreating online redo log files on standby.
    On standby only standby redo log files will be used. Not sure what you are trying to do.
    here is example how to create online redo log files, Check that diskgroup is mounted and have sufficient space to create.
    sys@ORCL> select member from v$logfile;
    MEMBER
    C:\ORACLE\ORADATA\ORCL\REDO03.LOG
    C:\ORACLE\ORADATA\ORCL\REDO02.LOG
    C:\ORACLE\ORADATA\ORCL\REDO01.LOG
    sys@ORCL> alter database add logfile group 4 (
      2     'C:\ORACLE\ORADATA\ORCL\redo_g01a.log',
      3     'C:\ORACLE\ORADATA\ORCL\redo_g01b.log',
      4     'C:\ORACLE\ORADATA\ORCL\redo_g01c.log') size 10m;
    Database altered.
    sys@ORCL> select member from v$logfile;
    MEMBER
    C:\ORACLE\ORADATA\ORCL\REDO03.LOG
    C:\ORACLE\ORADATA\ORCL\REDO02.LOG
    C:\ORACLE\ORADATA\ORCL\REDO01.LOG
    C:\ORACLE\ORADATA\ORCL\REDO_G01A.LOG
    C:\ORACLE\ORADATA\ORCL\REDO_G01B.LOG
    C:\ORACLE\ORADATA\ORCL\REDO_G01C.LOG
    6 rows selected.
    sys@ORCL>
    Your profile:-
    888442      
         Newbie
    Handle:      888442
    Status Level:      Newbie
    Registered:      Sep 29, 2011
    Total Posts:      12
    Total Questions:      8 (7 unresolved)
    Close the threads if answered, Keep the forum clean.

  • Choosing block size for RAID 0 & Final Cut

    Hi.
    I now have 3 500GB internal Seagate drives in bays 2/3/4 and want to make a striped 1.5TB RAID to use with Final Cut Studio 2. The help page talks about choosing a "large" data block size for use with video, but makes no specific size suggestion. What value would you recommend that I select for the block size? I haven't been in there yet so I don't know what the choices are.
    Any other settings I should be aware of that will optimize the RAID performance for video capture and editing? Thanks!
    Fred
    Message was edited by: FredGarvin
    Message was edited by: FredGarvin

    If you're using Disc Utility to set up your RAID, when you go to the RAID tab, you'll see an options button near the bottom of the window... clicking this will open a small menu where you can set the data block size... the largest is 256K, which is what you'd want to use.
    As for you're other question... have a look at this website: http://bytepile.com/raid_class.php
    note that disc utility can only set up RAID 0 & RAID 1 (if i remember rightly).

  • Default block size for UFS format in OSX?

    Hi,
    I formatted an external drive as "unix format" (UFS) using Disk Utility, which subsequently became unrecognizable by the firewire controller. Now in a Unix box, the superblock of this drive is coming up as corrupted or nonexistent. Does anyone know where I can find the default block size for UFS in os x? I need to specify a backup superblock. Thanks!
    -Dan

    If it is just scratch, run some benchmarks with it set to 128k and 256k and see how it feels with each. The default is too small, though some find it acceptable for small images. For larger files you want larger - and for PS scratch you definitely want 128 or 256k.

  • Using large block sizes for index and table spaces

    " You are not using large blocksizes for your index tablespaces. Oracle research proves that indexes will build flatter tree structures in larger blocksizes.
    Is this a generic statement that I can use for all tables or indexes? I also have batch and online activity. My primary target is batch and it should not impact online. Not sure if both have common tables.
    How to find the current block size used for tables and index? is there a v$parameter query?
    What is an optimal block size value for batch?
    How do I know when flatter tree str has been achieved using above changes? Is there a query to determine this?
    What about tables, what is the success criterion for tables. can we use the same flat tree str criterion? Is there a query for this?

    user3390467 wrote:
    " You are not using large blocksizes for your index tablespaces. Oracle research proves that indexes will build flatter tree structures in larger blocksizes.
    Is this a generic statement that I can use for all tables or indexes? This is a generic statement used by some consultants. Unfortunately, it is riddled with exceptions and other considerations.
    One consultant in particular seems to have anecdotal evidence that using different block sizes for index (big) and data (small) can yield almost miraculous improvements. However, that can not be backed up due to NDA. Many of the rest of us can not duplicate the improvements, and indeed some find situations where that results in a degradation (esp with high insert/update rates from separated transactions).
    I also have batch and online activity. My primary target is batch and it should not impact online. Not sure if both have common tables.
    How to find the current block size used for tables and index? is there a v$parameter query?
    What is an optimal block size value for batch?
    How do I know when flatter tree str has been achieved using above changes? Is there a query to determine this?
    What about tables, what is the success criterion for tables. can we use the same flat tree str criterion? Is there a query for this?I'd strongly recommend that you
    1) stop using generic tools to analyze specific problems
    2) define you problem in detail ()what are you really trying to accomplish - seems like performance tuning, but you never really state that)
    3) define the OS and DB version - in detail. Give rev levels and patch levels.
    If you are having a serious performance issue, I strongly recommend you look at some performance tuning specialists like "http://www.method-r.com/", "http://www.miracleas.dk/", "http://www.hotsos.com/", "http://www.pythian.com/", or even Oracle's Performance Tuning consultants. Definitely worth the price of admission.

  • Optimal Block Size for Xserve's RAID hosting Final Cut Server

    What would be the optimal block size for the software RAID on the machine that will be hosting Final Cut Server? The default is 36K. Since FCS is essentially a database, would be the optimal settings? Any glimpse what data size chunks FCS write to the disk?

    Actually I meant the block size for the internal startup volume where FCS is installed, not Xsan volumes. As to optimal settings for Xsan volumes it really depends on the type of the data you store on Xsan, and if it is primarily video, what format: SD, HD.

  • Optimal NTFS block size for Oracle 11G on Windows 2008 R2 (OLTP)

    Hi All,
    We are currently setting up an Oracle 11G instance on a Windows 2008 R2 server and were looking to see if there was an optimal NTFS block size. I've read the following: http://docs.oracle.com/cd/E11882_01/win.112/e10845/specs.htm
    But it only mentioned the block sizes that can be used (2k - 16k). And basically what i got out of it, was the different block szes affect the max # of database files possible for each database.
    Is there an optimal NTFS block size for Oracle 11G OLTP system on Windows?
    Thanks in advance

    Is there an optimal NTFS block size for Oracle 11G OLTP system on Windows?ideally FS block size should be equal to Oracle tablespace block size.
    or at least be N times less than Oracle block size.
    For example - if Oracle BS=8K then NTFS BS better to be 8K but also can be 4K or 2K.
    Also both must be 1 to N times of Disk sector size. Older disks had sectors 512 bytes.
    Contemporary HDDs have internal sector size 4K. Usually.

  • CANNOT SET BLOCK SIZE TO 16K IN ORACLE 10.2

    Has anyone tried to configure the block size to 16k in oracle 10.2? I installed 10.2 with the datawarehouse option and it set the block size to 8k. Any ideas?

    Yes, but the feature is not selected by default, if I remember correctly.
    Werner

  • Change block size for several log-files simultaneously?

    Hi,
    I'm using SignalExpress to record and analyze data.
    Sometimes I want to analyze the recorded data both for a short period of time and for longer time.
    (Imagine creating an average of every second first and then an average of every 10 seconds)
    Then I need to change all the log-files, and also the specific parts of the log-file. See attachment.
    I have sometimes up to 1000 log-files containing signals from 4 different modules, that makes 4000 adjustments to change from block size 10000 to block size 1000.
    Is there any way to adjust all the log-files block size at once?
    Many thanks!
    Anders Hansson
    Engineer
    Attachments:
    NI.JPG ‏95 KB

    Hi,
    Is't anyone else interested in a solution for this operation?
    I reported this to the NI-feedback service and they adviced me to report/request advice here to get a quicker reply.
    So...
    Best regards
    Ingenjör Hansson

  • RAID block size for final cut pro x

    Just got one of the new late 2012 27" iMacs and a 6 TB LaCie Thunderbolt drive. Can finally edit the video I took last spring. I'll be using Final Cut Pro X, and doing a lot of multicam stuff with 4 or 5 views and a separate audio track. The LaCie came formatted as a mirrored RAID. I'm going to change that to 0 (Striped RAID set), but am wondering what block size to set. The default is 32k, but I have read that this ought to be increased to the max (256k) for video editing. I have also read it should NOT be increased. And the posts I have read have all been at least 3 years old. So let me ask you all--what block size would you recommend for my situation?
    Thanks in advance!

    Hi Eddie...
    This depends on what kind of source footage you are editing....
    For compressed Video, Audio and Uncompressed audio 128k
    I have only had BAD results with 256k. 64 is also weird. Whereas 32 is fine.
    All my RAIDs have 128k for audio/video editing
    you can go further if you editing Image Sequences.. but according to my own findings and I have been dealing with raid since years.... 128k does the job the best.
    Rule of thumb.... The smaller the file sizes you are putting the RAID the smaller the block size. And vice versa.
    I.e. You would cripple the raid performance if storing a database on it, having a block size of 256. In case of servers and OS 32k would be a good choice, perhaps even 16k if supported.

  • Recommended Block Size For RAID 0

    I am setting up a RAID configuration (Striping, no Parity, Mac G5, OS-X) and was curious what the recommended Block Size should be. Content is primarily (but not limited to) Images created with Adobe Photoshop CS2 and range in size from 1.5MB to >20MB. The default for OS-X is 32K chunks of data.
    Drives are External FW-400.
    Many thanks, and Happy Holidays to all!

    If it is just scratch, run some benchmarks with it set to 128k and 256k and see how it feels with each. The default is too small, though some find it acceptable for small images. For larger files you want larger - and for PS scratch you definitely want 128 or 256k.

  • Best RAID block size for media drive?

    What block size give you best performance when it comes to pushing data?
    For a striped RAID setup.
    32 is standard but since most of the media files are big and consistent would a higher value like 128 or even 256 KB be better?

    "fools step in where angels fear to tred"
    Well I'm not volunteering to be one of those.
    Jerry, if your fiber channel raid is giving you the throughput that you need, don't be concerned.
    (I had a quick look through your manual and I'm also confused. But I can't afford that kind of setup so...?)
    I have a simple two 250GB LaCie d2 raid 0 set via SoftRaid and firewire 800 using the G5 port and a LaCie card firewire 800 for dual channel setup.
    This houses my media for FCP.
    My stripe size is set to 128K simply because that's what the SoftRaid manual recommended for video applications.
    This two drive setup is fine for multiple SD streams of DV, but can only manage a single 8 bit uncompressed HD 1080i stream without dropped frames.

  • Force Logical Block size for HDD pulled from external enclourse

    So I pulled a 3tb hitachi drive from an external enclourse, and the usb-to-sata board in the enclourse forced a logical sector size of 4K for the drive. But connecting directly to the drive exposes the logical sector size to 512 bytes. Resulting in me being unable to mount the partition. So what I am asking is, is there a way to force the logical sector size for a drive to be different than the drive exposes or am I going to have to move everything off the drive and reformat it?
    The enclourse in question:
    http://www.newegg.com/Product/Product.a … 6822145510
    Direct:
    [73828.076447] sd 0:0:1:2: [sdl] Synchronizing SCSI cache
    [73838.912744] scsi 0:0:1:2: Direct-Access Hitachi HDS723030ALA640 R001 PQ: 0 ANSI: 5
    [73838.912889] sd 0:0:1:2: [sdo] 5860533168 512-byte logical blocks: (3.00 TB/2.72 TiB)
    [73838.912892] sd 0:0:1:2: Attached scsi generic sg12 type 0
    [73838.912995] sd 0:0:1:2: [sdo] Write Protect is off
    [73838.912999] sd 0:0:1:2: [sdo] Mode Sense: cb 00 00 08
    [73838.913042] sd 0:0:1:2: [sdo] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA
    [73838.949343] sdo: sdo1
    [73838.949614] sd 0:0:1:2: [sdo] Attached SCSI disk
    USB:
    Mar 5 18:52:35 localhost kernel: [ 4.711438] sd 8:0:0:0: [sdl] 732566016 4096-byte logical blocks: (3.00 TB/2.72 TiB)
    Mar 5 18:52:35 localhost kernel: [ 4.712191] sd 8:0:0:0: [sdl] Write Protect is off
    Mar 5 18:52:35 localhost kernel: [ 4.712193] sd 8:0:0:0: [sdl] Mode Sense: 4b 00 10 08
    Mar 5 18:52:35 localhost kernel: [ 4.712938] sd 8:0:0:0: [sdl] No Caching mode page present
    Mar 5 18:52:35 localhost kernel: [ 4.713978] sd 8:0:0:0: [sdl] Assuming drive cache: write through
    Mar 5 18:52:35 localhost kernel: [ 4.714720] sd 8:0:0:0: [sdl] 732566016 4096-byte logical blocks: (3.00 TB/2.72 TiB)
    Mar 5 18:52:35 localhost kernel: [ 4.716186] sd 8:0:0:0: [sdl] No Caching mode page present
    Mar 5 18:52:35 localhost kernel: [ 4.717215] sd 8:0:0:0: [sdl] Assuming drive cache: write through
    Mar 5 18:52:35 localhost kernel: [ 4.735573] sdl: sdl1
    Mar 5 18:52:35 localhost kernel: [ 4.736557] sd 8:0:0:0: [sdl] 732566016 4096-byte logical blocks: (3.00 TB/2.72 TiB)
    Mar 5 18:52:35 localhost kernel: [ 4.738058] sd 8:0:0:0: [sdl] No Caching mode page present
    Mar 5 18:52:35 localhost kernel: [ 4.739085] sd 8:0:0:0: [sdl] Assuming drive cache: write through
    Mar 5 18:52:35 localhost kernel: [ 4.740238] sd 8:0:0:0: [sdl] Attached SCSI disk

    Hi ebrian,
    You can't perform a migration from Linux to Windows in that manner with 9i.The plan can be changed to migrate from 9.2.0.4 to 10g and than move from 32-bit Linux to 32-bit Windows.
    It should not matter the end result for the OP is the same 10g on 32 bit windows.
    It just depends on what technoly the OP wants to use (export/import or rman convert).
    Regards,
    Tycho

  • RAID Block size for video

    Can anyone share their experience on what block size in a RAID tends to be better for video, or what size you are using?  (i.e. 128K, 256K, 512K)

    Thanks for the reply Harm. I later noticed that the Areca Controller grays out the block size if RAID 3 is selected.

  • Really still need EBS (Erase Block Size) for partitioning a SSD?

    The Arch wiki mentions that when partitioning an SSD you need to align the partitions to the erase block size (see https://wiki.archlinux.org/index.php/Pa … ate_drives ). There are a lot of how-to-partition-an-ssd-under-linux which also mention this value, even an online calculater.
    Those value are known for a lot of the serious models, among them also the Samsung 840 Pro and EVO. As I already asked in an older thread, I would like to know the value for the current model Samsung 850 Pro. Unfortunately Samsung Germany's support refuses to provide me with those values.
    Therefore I ask myself, is it still necessary to align partititions to a multiple of the EBS, or is this unnecessary if the device supports TRIM?
    Thanks!

    Did you read the section immediately after the one you linked? https://wiki.archlinux.org/index.php/Pa … ng_tools_2
    In past, proper alignment required manual calculation and intervention when partitioning. Many of the common partition tools now handle partition alignment automatically:
    To verify a partition is aligned, query it using /usr/bin/blockdev as shown below, if a '0' is returned, the partition is aligned
    Is that not sufficient for your purposes?
    Last edited by 2ManyDogs (2014-11-04 18:55:37)

Maybe you are looking for