LVM  or not?

Hi,
Which one is the preferred method for partitioning a Linux for Oracle DB, LVM or not?

OSSGeek wrote:
What about the OS files and Oracle files?LVM gives the most flexibility combined with ease of use.

Similar Messages

  • LVM Volumes not available after update

    Hi All!
    I haven't updated my system for about two months and today I updated it. Now I have the problem that I cannot boot properly. I have my root partition in an LVM volume and on boot I get the message
    ERROR: device 'UUID=xxx' not found. Skipping fs
    ERROR: Unable to find root device 'UUID=xxx'
    After that I land in the recovery shell. After some research I found, that "lvm lvdisplay" showed that my volumes where not available and I had to reenable them with "lvm vgchange -a y".
    Issuing any lvm command also produced the following warning:
    WARNING: lvmetad is running but disabled. Restart lvmetad before enabling it!
    Anyway, after issuing the commands and exiting the recovery shell, the system booted again. However, I would prefer being able to boot without manual actions.
    Thanks in advance!
    Further information:
    vgdisplay
    --- Volume group ---
    VG Name ArchLVM
    System ID
    Format lvm2
    Metadata Areas 1
    Metadata Sequence No 3
    VG Access read/write
    VG Status resizable
    MAX LV 0
    Cur LV 2
    Open LV 1
    Max PV 0
    Cur PV 1
    Act PV 1
    VG Size 232.69 GiB
    PE Size 4.00 MiB
    Total PE 59568
    Alloc PE / Size 59568 / 232.69 GiB
    Free PE / Size 0 / 0
    VG UUID SoB3M1-v1fD-1abI-PNJ3-6IOn-FfdI-0RoLK5
    lvdisplay (LV Status was 'not available' right after booting)
    --- Logical volume ---
    LV Path /dev/ArchLVM/Swap
    LV Name Swap
    VG Name ArchLVM
    LV UUID XRYBrz-LojR-k6SD-XIxV-wHnY-f3VG-giKL6V
    LV Write Access read/write
    LV Creation host, time archiso, 2014-05-16 14:43:06 +0200
    LV Status available
    # open 0
    LV Size 8.00 GiB
    Current LE 2048
    Segments 1
    Allocation inherit
    Read ahead sectors auto
    - currently set to 256
    Block device 254:0
    --- Logical volume ---
    LV Path /dev/ArchLVM/Root
    LV Name Root
    VG Name ArchLVM
    LV UUID lpjDl4-Jqzu-ZWkq-Uphc-IaOo-6Rzd-cIh5yv
    LV Write Access read/write
    LV Creation host, time archiso, 2014-05-16 14:43:27 +0200
    LV Status available
    # open 1
    LV Size 224.69 GiB
    Current LE 57520
    Segments 1
    Allocation inherit
    Read ahead sectors auto
    - currently set to 256
    Block device 254:1
    /etc/fstab
    # /etc/fstab: static file system information
    # <file system> <dir> <type> <options> <dump> <pass>
    # /dev/mapper/ArchLVM-Root
    UUID=2db82d1a-47a4-4e30-a819-143e8fb75199 / ext4 rw,relatime,data=ordered 0 1
    #/dev/mapper/ArchLVM-Root / ext4 rw,relatime,data=ordered 0 1
    # /dev/sda1
    UUID=72691888-a781-4cdd-a98e-2613d87925d0 /boot ext2 rw,relatime 0 2
    /etc/mkinitcpio.conf
    # vim:set ft=sh
    # MODULES
    # The following modules are loaded before any boot hooks are
    # run. Advanced users may wish to specify all system modules
    # in this array. For instance:
    # MODULES="piix ide_disk reiserfs"
    MODULES=""
    # BINARIES
    # This setting includes any additional binaries a given user may
    # wish into the CPIO image. This is run last, so it may be used to
    # override the actual binaries included by a given hook
    # BINARIES are dependency parsed, so you may safely ignore libraries
    BINARIES=""
    # FILES
    # This setting is similar to BINARIES above, however, files are added
    # as-is and are not parsed in any way. This is useful for config files.
    FILES=""
    # HOOKS
    # This is the most important setting in this file. The HOOKS control the
    # modules and scripts added to the image, and what happens at boot time.
    # Order is important, and it is recommended that you do not change the
    # order in which HOOKS are added. Run 'mkinitcpio -H <hook name>' for
    # help on a given hook.
    # 'base' is _required_ unless you know precisely what you are doing.
    # 'udev' is _required_ in order to automatically load modules
    # 'filesystems' is _required_ unless you specify your fs modules in MODULES
    # Examples:
    ## This setup specifies all modules in the MODULES setting above.
    ## No raid, lvm2, or encrypted root is needed.
    # HOOKS="base"
    ## This setup will autodetect all modules for your system and should
    ## work as a sane default
    # HOOKS="base udev autodetect block filesystems"
    ## This setup will generate a 'full' image which supports most systems.
    ## No autodetection is done.
    # HOOKS="base udev block filesystems"
    ## This setup assembles a pata mdadm array with an encrypted root FS.
    ## Note: See 'mkinitcpio -H mdadm' for more information on raid devices.
    # HOOKS="base udev block mdadm encrypt filesystems"
    ## This setup loads an lvm2 volume group on a usb device.
    # HOOKS="base udev block lvm2 filesystems"
    ## NOTE: If you have /usr on a separate partition, you MUST include the
    # usr, fsck and shutdown hooks.
    HOOKS="base udev autodetect modconf block lvm2 filesystems keyboard fsck"
    # COMPRESSION
    # Use this to compress the initramfs image. By default, gzip compression
    # is used. Use 'cat' to create an uncompressed image.
    #COMPRESSION="gzip"
    #COMPRESSION="bzip2"
    #COMPRESSION="lzma"
    #COMPRESSION="xz"
    #COMPRESSION="lzop"
    #COMPRESSION="lz4"
    # COMPRESSION_OPTIONS
    # Additional options for the compressor
    #COMPRESSION_OPTIONS=""
    /boot/grub/grub.cfg
    # DO NOT EDIT THIS FILE
    # It is automatically generated by grub-mkconfig using templates
    # from /etc/grub.d and settings from /etc/default/grub
    ### BEGIN /etc/grub.d/00_header ###
    insmod part_gpt
    insmod part_msdos
    if [ -s $prefix/grubenv ]; then
    load_env
    fi
    if [ "${next_entry}" ] ; then
    set default="${next_entry}"
    set next_entry=
    save_env next_entry
    set boot_once=true
    else
    set default="0"
    fi
    if [ x"${feature_menuentry_id}" = xy ]; then
    menuentry_id_option="--id"
    else
    menuentry_id_option=""
    fi
    export menuentry_id_option
    if [ "${prev_saved_entry}" ]; then
    set saved_entry="${prev_saved_entry}"
    save_env saved_entry
    set prev_saved_entry=
    save_env prev_saved_entry
    set boot_once=true
    fi
    function savedefault {
    if [ -z "${boot_once}" ]; then
    saved_entry="${chosen}"
    save_env saved_entry
    fi
    function load_video {
    if [ x$feature_all_video_module = xy ]; then
    insmod all_video
    else
    insmod efi_gop
    insmod efi_uga
    insmod ieee1275_fb
    insmod vbe
    insmod vga
    insmod video_bochs
    insmod video_cirrus
    fi
    if [ x$feature_default_font_path = xy ] ; then
    font=unicode
    else
    insmod part_msdos
    insmod lvm
    insmod ext2
    set root='lvmid/SoB3M1-v1fD-1abI-PNJ3-6IOn-FfdI-0RoLK5/lpjDl4-Jqzu-ZWkq-Uphc-IaOo-6Rzd-cIh5yv'
    if [ x$feature_platform_search_hint = xy ]; then
    search --no-floppy --fs-uuid --set=root --hint='lvmid/SoB3M1-v1fD-1abI-PNJ3-6IOn-FfdI-0RoLK5/lpjDl4-Jqzu-ZWkq-Uphc-IaOo-6Rzd-cIh5yv' 2db82d1a-47a4-4e30-a819-143e8fb75199
    else
    search --no-floppy --fs-uuid --set=root 2db82d1a-47a4-4e30-a819-143e8fb75199
    fi
    font="/usr/share/grub/unicode.pf2"
    fi
    if loadfont $font ; then
    set gfxmode=auto
    load_video
    insmod gfxterm
    fi
    terminal_input console
    terminal_output gfxterm
    if [ x$feature_timeout_style = xy ] ; then
    set timeout_style=menu
    set timeout=5
    # Fallback normal timeout code in case the timeout_style feature is
    # unavailable.
    else
    set timeout=5
    fi
    ### END /etc/grub.d/00_header ###
    ### BEGIN /etc/grub.d/10_linux ###
    menuentry 'Arch Linux' --class arch --class gnu-linux --class gnu --class os $menuentry_id_option 'gnulinux-simple-2db82d1a-47a4-4e30-a819-143e8fb75199' {
    load_video
    set gfxpayload=keep
    insmod gzio
    insmod part_msdos
    insmod ext2
    set root='hd0,msdos1'
    if [ x$feature_platform_search_hint = xy ]; then
    search --no-floppy --fs-uuid --set=root --hint-bios=hd0,msdos1 --hint-efi=hd0,msdos1 --hint-baremetal=ahci0,msdos1 72691888-a781-4cdd-a98e-2613d87925d0
    else
    search --no-floppy --fs-uuid --set=root 72691888-a781-4cdd-a98e-2613d87925d0
    fi
    echo 'Loading Linux linux ...'
    linux /vmlinuz-linux root=UUID=2db82d1a-47a4-4e30-a819-143e8fb75199 rw quiet
    echo 'Loading initial ramdisk ...'
    initrd /initramfs-linux.img
    submenu 'Advanced options for Arch Linux' $menuentry_id_option 'gnulinux-advanced-2db82d1a-47a4-4e30-a819-143e8fb75199' {
    menuentry 'Arch Linux, with Linux linux' --class arch --class gnu-linux --class gnu --class os $menuentry_id_option 'gnulinux-linux-advanced-2db82d1a-47a4-4e30-a819-143e8fb75199' {
    load_video
    set gfxpayload=keep
    insmod gzio
    insmod part_msdos
    insmod ext2
    set root='hd0,msdos1'
    if [ x$feature_platform_search_hint = xy ]; then
    search --no-floppy --fs-uuid --set=root --hint-bios=hd0,msdos1 --hint-efi=hd0,msdos1 --hint-baremetal=ahci0,msdos1 72691888-a781-4cdd-a98e-2613d87925d0
    else
    search --no-floppy --fs-uuid --set=root 72691888-a781-4cdd-a98e-2613d87925d0
    fi
    echo 'Loading Linux linux ...'
    linux /vmlinuz-linux root=UUID=2db82d1a-47a4-4e30-a819-143e8fb75199 rw quiet
    echo 'Loading initial ramdisk ...'
    initrd /initramfs-linux.img
    menuentry 'Arch Linux, with Linux linux (fallback initramfs)' --class arch --class gnu-linux --class gnu --class os $menuentry_id_option 'gnulinux-linux-fallback-2db82d1a-47a4-4e30-a819-143e8fb75199' {
    load_video
    set gfxpayload=keep
    insmod gzio
    insmod part_msdos
    insmod ext2
    set root='hd0,msdos1'
    if [ x$feature_platform_search_hint = xy ]; then
    search --no-floppy --fs-uuid --set=root --hint-bios=hd0,msdos1 --hint-efi=hd0,msdos1 --hint-baremetal=ahci0,msdos1 72691888-a781-4cdd-a98e-2613d87925d0
    else
    search --no-floppy --fs-uuid --set=root 72691888-a781-4cdd-a98e-2613d87925d0
    fi
    echo 'Loading Linux linux ...'
    linux /vmlinuz-linux root=UUID=2db82d1a-47a4-4e30-a819-143e8fb75199 rw quiet
    echo 'Loading initial ramdisk ...'
    initrd /initramfs-linux-fallback.img
    ### END /etc/grub.d/10_linux ###
    ### BEGIN /etc/grub.d/20_linux_xen ###
    ### END /etc/grub.d/20_linux_xen ###
    ### BEGIN /etc/grub.d/30_os-prober ###
    ### END /etc/grub.d/30_os-prober ###
    ### BEGIN /etc/grub.d/40_custom ###
    # This file provides an easy way to add custom menu entries. Simply type the
    # menu entries you want to add after this comment. Be careful not to change
    # the 'exec tail' line above.
    ### END /etc/grub.d/40_custom ###
    ### BEGIN /etc/grub.d/41_custom ###
    if [ -f ${config_directory}/custom.cfg ]; then
    source ${config_directory}/custom.cfg
    elif [ -z "${config_directory}" -a -f $prefix/custom.cfg ]; then
    source $prefix/custom.cfg;
    fi
    ### END /etc/grub.d/41_custom ###
    ### BEGIN /etc/grub.d/60_memtest86+ ###
    ### END /etc/grub.d/60_memtest86+ ###
    Last edited by Kirodema (2014-07-16 07:31:34)

    use_lvmetad = 0
    lvm2-lvmetad is not enabled or running on my system. Shall I activate it?
    # This is an example configuration file for the LVM2 system.
    # It contains the default settings that would be used if there was no
    # /etc/lvm/lvm.conf file.
    # Refer to 'man lvm.conf' for further information including the file layout.
    # To put this file in a different directory and override /etc/lvm set
    # the environment variable LVM_SYSTEM_DIR before running the tools.
    # N.B. Take care that each setting only appears once if uncommenting
    # example settings in this file.
    # This section allows you to set the way the configuration settings are handled.
    config {
    # If enabled, any LVM2 configuration mismatch is reported.
    # This implies checking that the configuration key is understood
    # by LVM2 and that the value of the key is of a proper type.
    # If disabled, any configuration mismatch is ignored and default
    # value is used instead without any warning (a message about the
    # configuration key not being found is issued in verbose mode only).
    checks = 1
    # If enabled, any configuration mismatch aborts the LVM2 process.
    abort_on_errors = 0
    # Directory where LVM looks for configuration profiles.
    profile_dir = "/etc/lvm/profile"
    # This section allows you to configure which block devices should
    # be used by the LVM system.
    devices {
    # Where do you want your volume groups to appear ?
    dir = "/dev"
    # An array of directories that contain the device nodes you wish
    # to use with LVM2.
    scan = [ "/dev" ]
    # If set, the cache of block device nodes with all associated symlinks
    # will be constructed out of the existing udev database content.
    # This avoids using and opening any inapplicable non-block devices or
    # subdirectories found in the device directory. This setting is applied
    # to udev-managed device directory only, other directories will be scanned
    # fully. LVM2 needs to be compiled with udev support for this setting to
    # take effect. N.B. Any device node or symlink not managed by udev in
    # udev directory will be ignored with this setting on.
    obtain_device_list_from_udev = 1
    # If several entries in the scanned directories correspond to the
    # same block device and the tools need to display a name for device,
    # all the pathnames are matched against each item in the following
    # list of regular expressions in turn and the first match is used.
    preferred_names = [ ]
    # Try to avoid using undescriptive /dev/dm-N names, if present.
    # preferred_names = [ "^/dev/mpath/", "^/dev/mapper/mpath", "^/dev/[hs]d" ]
    # A filter that tells LVM2 to only use a restricted set of devices.
    # The filter consists of an array of regular expressions. These
    # expressions can be delimited by a character of your choice, and
    # prefixed with either an 'a' (for accept) or 'r' (for reject).
    # The first expression found to match a device name determines if
    # the device will be accepted or rejected (ignored). Devices that
    # don't match any patterns are accepted.
    # Be careful if there there are symbolic links or multiple filesystem
    # entries for the same device as each name is checked separately against
    # the list of patterns. The effect is that if the first pattern in the
    # list to match a name is an 'a' pattern for any of the names, the device
    # is accepted; otherwise if the first pattern in the list to match a name
    # is an 'r' pattern for any of the names it is rejected; otherwise it is
    # accepted.
    # Don't have more than one filter line active at once: only one gets used.
    # Run vgscan after you change this parameter to ensure that
    # the cache file gets regenerated (see below).
    # If it doesn't do what you expect, check the output of 'vgscan -vvvv'.
    # If lvmetad is used, then see "A note about device filtering while
    # lvmetad is used" comment that is attached to global/use_lvmetad setting.
    # By default we accept every block device:
    filter = [ "a/.*/" ]
    # Exclude the cdrom drive
    # filter = [ "r|/dev/cdrom|" ]
    # When testing I like to work with just loopback devices:
    # filter = [ "a/loop/", "r/.*/" ]
    # Or maybe all loops and ide drives except hdc:
    # filter =[ "a|loop|", "r|/dev/hdc|", "a|/dev/ide|", "r|.*|" ]
    # Use anchors if you want to be really specific
    # filter = [ "a|^/dev/hda8$|", "r/.*/" ]
    # Since "filter" is often overridden from command line, it is not suitable
    # for system-wide device filtering (udev rules, lvmetad). To hide devices
    # from LVM-specific udev processing and/or from lvmetad, you need to set
    # global_filter. The syntax is the same as for normal "filter"
    # above. Devices that fail the global_filter are not even opened by LVM.
    # global_filter = []
    # The results of the filtering are cached on disk to avoid
    # rescanning dud devices (which can take a very long time).
    # By default this cache is stored in the /etc/lvm/cache directory
    # in a file called '.cache'.
    # It is safe to delete the contents: the tools regenerate it.
    # (The old setting 'cache' is still respected if neither of
    # these new ones is present.)
    # N.B. If obtain_device_list_from_udev is set to 1 the list of
    # devices is instead obtained from udev and any existing .cache
    # file is removed.
    cache_dir = "/etc/lvm/cache"
    cache_file_prefix = ""
    # You can turn off writing this cache file by setting this to 0.
    write_cache_state = 1
    # Advanced settings.
    # List of pairs of additional acceptable block device types found
    # in /proc/devices with maximum (non-zero) number of partitions.
    # types = [ "fd", 16 ]
    # If sysfs is mounted (2.6 kernels) restrict device scanning to
    # the block devices it believes are valid.
    # 1 enables; 0 disables.
    sysfs_scan = 1
    # By default, LVM2 will ignore devices used as component paths
    # of device-mapper multipath devices.
    # 1 enables; 0 disables.
    multipath_component_detection = 1
    # By default, LVM2 will ignore devices used as components of
    # software RAID (md) devices by looking for md superblocks.
    # 1 enables; 0 disables.
    md_component_detection = 1
    # By default, if a PV is placed directly upon an md device, LVM2
    # will align its data blocks with the md device's stripe-width.
    # 1 enables; 0 disables.
    md_chunk_alignment = 1
    # Default alignment of the start of a data area in MB. If set to 0,
    # a value of 64KB will be used. Set to 1 for 1MiB, 2 for 2MiB, etc.
    # default_data_alignment = 1
    # By default, the start of a PV's data area will be a multiple of
    # the 'minimum_io_size' or 'optimal_io_size' exposed in sysfs.
    # - minimum_io_size - the smallest request the device can perform
    # w/o incurring a read-modify-write penalty (e.g. MD's chunk size)
    # - optimal_io_size - the device's preferred unit of receiving I/O
    # (e.g. MD's stripe width)
    # minimum_io_size is used if optimal_io_size is undefined (0).
    # If md_chunk_alignment is enabled, that detects the optimal_io_size.
    # This setting takes precedence over md_chunk_alignment.
    # 1 enables; 0 disables.
    data_alignment_detection = 1
    # Alignment (in KB) of start of data area when creating a new PV.
    # md_chunk_alignment and data_alignment_detection are disabled if set.
    # Set to 0 for the default alignment (see: data_alignment_default)
    # or page size, if larger.
    data_alignment = 0
    # By default, the start of the PV's aligned data area will be shifted by
    # the 'alignment_offset' exposed in sysfs. This offset is often 0 but
    # may be non-zero; e.g.: certain 4KB sector drives that compensate for
    # windows partitioning will have an alignment_offset of 3584 bytes
    # (sector 7 is the lowest aligned logical block, the 4KB sectors start
    # at LBA -1, and consequently sector 63 is aligned on a 4KB boundary).
    # But note that pvcreate --dataalignmentoffset will skip this detection.
    # 1 enables; 0 disables.
    data_alignment_offset_detection = 1
    # If, while scanning the system for PVs, LVM2 encounters a device-mapper
    # device that has its I/O suspended, it waits for it to become accessible.
    # Set this to 1 to skip such devices. This should only be needed
    # in recovery situations.
    ignore_suspended_devices = 0
    # ignore_lvm_mirrors: Introduced in version 2.02.104
    # This setting determines whether logical volumes of "mirror" segment
    # type are scanned for LVM labels. This affects the ability of
    # mirrors to be used as physical volumes. If 'ignore_lvm_mirrors'
    # is set to '1', it becomes impossible to create volume groups on top
    # of mirror logical volumes - i.e. to stack volume groups on mirrors.
    # Allowing mirror logical volumes to be scanned (setting the value to '0')
    # can potentially cause LVM processes and I/O to the mirror to become
    # blocked. This is due to the way that the "mirror" segment type handles
    # failures. In order for the hang to manifest itself, an LVM command must
    # be run just after a failure and before the automatic LVM repair process
    # takes place OR there must be failures in multiple mirrors in the same
    # volume group at the same time with write failures occurring moments
    # before a scan of the mirror's labels.
    # Note that these scanning limitations do not apply to the LVM RAID
    # types, like "raid1". The RAID segment types handle failures in a
    # different way and are not subject to possible process or I/O blocking.
    # It is encouraged that users set 'ignore_lvm_mirrors' to 1 if they
    # are using the "mirror" segment type. Users that require volume group
    # stacking on mirrored logical volumes should consider using the "raid1"
    # segment type. The "raid1" segment type is not available for
    # active/active clustered volume groups.
    # Set to 1 to disallow stacking and thereby avoid a possible deadlock.
    ignore_lvm_mirrors = 1
    # During each LVM operation errors received from each device are counted.
    # If the counter of a particular device exceeds the limit set here, no
    # further I/O is sent to that device for the remainder of the respective
    # operation. Setting the parameter to 0 disables the counters altogether.
    disable_after_error_count = 0
    # Allow use of pvcreate --uuid without requiring --restorefile.
    require_restorefile_with_uuid = 1
    # Minimum size (in KB) of block devices which can be used as PVs.
    # In a clustered environment all nodes must use the same value.
    # Any value smaller than 512KB is ignored.
    # Ignore devices smaller than 2MB such as floppy drives.
    pv_min_size = 2048
    # The original built-in setting was 512 up to and including version 2.02.84.
    # pv_min_size = 512
    # Issue discards to a logical volumes's underlying physical volume(s) when
    # the logical volume is no longer using the physical volumes' space (e.g.
    # lvremove, lvreduce, etc). Discards inform the storage that a region is
    # no longer in use. Storage that supports discards advertise the protocol
    # specific way discards should be issued by the kernel (TRIM, UNMAP, or
    # WRITE SAME with UNMAP bit set). Not all storage will support or benefit
    # from discards but SSDs and thinly provisioned LUNs generally do. If set
    # to 1, discards will only be issued if both the storage and kernel provide
    # support.
    # 1 enables; 0 disables.
    issue_discards = 0
    # This section allows you to configure the way in which LVM selects
    # free space for its Logical Volumes.
    allocation {
    # When searching for free space to extend an LV, the "cling"
    # allocation policy will choose space on the same PVs as the last
    # segment of the existing LV. If there is insufficient space and a
    # list of tags is defined here, it will check whether any of them are
    # attached to the PVs concerned and then seek to match those PV tags
    # between existing extents and new extents.
    # Use the special tag "@*" as a wildcard to match any PV tag.
    # Example: LVs are mirrored between two sites within a single VG.
    # PVs are tagged with either @site1 or @site2 to indicate where
    # they are situated.
    # cling_tag_list = [ "@site1", "@site2" ]
    # cling_tag_list = [ "@*" ]
    # Changes made in version 2.02.85 extended the reach of the 'cling'
    # policies to detect more situations where data can be grouped
    # onto the same disks. Set this to 0 to revert to the previous
    # algorithm.
    maximise_cling = 1
    # Whether to use blkid library instead of native LVM2 code to detect
    # any existing signatures while creating new Physical Volumes and
    # Logical Volumes. LVM2 needs to be compiled with blkid wiping support
    # for this setting to take effect.
    # LVM2 native detection code is currently able to recognize these signatures:
    # - MD device signature
    # - swap signature
    # - LUKS signature
    # To see the list of signatures recognized by blkid, check the output
    # of 'blkid -k' command. The blkid can recognize more signatures than
    # LVM2 native detection code, but due to this higher number of signatures
    # to be recognized, it can take more time to complete the signature scan.
    use_blkid_wiping = 1
    # Set to 1 to wipe any signatures found on newly-created Logical Volumes
    # automatically in addition to zeroing of the first KB on the LV
    # (controlled by the -Z/--zero y option).
    # The command line option -W/--wipesignatures takes precedence over this
    # setting.
    # The default is to wipe signatures when zeroing.
    wipe_signatures_when_zeroing_new_lvs = 1
    # Set to 1 to guarantee that mirror logs will always be placed on
    # different PVs from the mirror images. This was the default
    # until version 2.02.85.
    mirror_logs_require_separate_pvs = 0
    # Set to 1 to guarantee that cache_pool metadata will always be
    # placed on different PVs from the cache_pool data.
    cache_pool_metadata_require_separate_pvs = 0
    # Specify the minimal chunk size (in kiB) for cache pool volumes.
    # Using a chunk_size that is too large can result in wasteful use of
    # the cache, where small reads and writes can cause large sections of
    # an LV to be mapped into the cache. However, choosing a chunk_size
    # that is too small can result in more overhead trying to manage the
    # numerous chunks that become mapped into the cache. The former is
    # more of a problem than the latter in most cases, so we default to
    # a value that is on the smaller end of the spectrum. Supported values
    # range from 32(kiB) to 1048576 in multiples of 32.
    # cache_pool_chunk_size = 64
    # Set to 1 to guarantee that thin pool metadata will always
    # be placed on different PVs from the pool data.
    thin_pool_metadata_require_separate_pvs = 0
    # Specify chunk size calculation policy for thin pool volumes.
    # Possible options are:
    # "generic" - if thin_pool_chunk_size is defined, use it.
    # Otherwise, calculate the chunk size based on
    # estimation and device hints exposed in sysfs:
    # the minimum_io_size. The chunk size is always
    # at least 64KiB.
    # "performance" - if thin_pool_chunk_size is defined, use it.
    # Otherwise, calculate the chunk size for
    # performance based on device hints exposed in
    # sysfs: the optimal_io_size. The chunk size is
    # always at least 512KiB.
    # thin_pool_chunk_size_policy = "generic"
    # Specify the minimal chunk size (in KB) for thin pool volumes.
    # Use of the larger chunk size may improve performance for plain
    # thin volumes, however using them for snapshot volumes is less efficient,
    # as it consumes more space and takes extra time for copying.
    # When unset, lvm tries to estimate chunk size starting from 64KB
    # Supported values are in range from 64 to 1048576.
    # thin_pool_chunk_size = 64
    # Specify discards behaviour of the thin pool volume.
    # Select one of "ignore", "nopassdown", "passdown"
    # thin_pool_discards = "passdown"
    # Set to 0, to disable zeroing of thin pool data chunks before their
    # first use.
    # N.B. zeroing larger thin pool chunk size degrades performance.
    # thin_pool_zero = 1
    # This section that allows you to configure the nature of the
    # information that LVM2 reports.
    log {
    # Controls the messages sent to stdout or stderr.
    # There are three levels of verbosity, 3 being the most verbose.
    verbose = 0
    # Set to 1 to suppress all non-essential messages from stdout.
    # This has the same effect as -qq.
    # When this is set, the following commands still produce output:
    # dumpconfig, lvdisplay, lvmdiskscan, lvs, pvck, pvdisplay,
    # pvs, version, vgcfgrestore -l, vgdisplay, vgs.
    # Non-essential messages are shifted from log level 4 to log level 5
    # for syslog and lvm2_log_fn purposes.
    # Any 'yes' or 'no' questions not overridden by other arguments
    # are suppressed and default to 'no'.
    silent = 0
    # Should we send log messages through syslog?
    # 1 is yes; 0 is no.
    syslog = 1
    # Should we log error and debug messages to a file?
    # By default there is no log file.
    #file = "/var/log/lvm2.log"
    # Should we overwrite the log file each time the program is run?
    # By default we append.
    overwrite = 0
    # What level of log messages should we send to the log file and/or syslog?
    # There are 6 syslog-like log levels currently in use - 2 to 7 inclusive.
    # 7 is the most verbose (LOG_DEBUG).
    level = 0
    # Format of output messages
    # Whether or not (1 or 0) to indent messages according to their severity
    indent = 1
    # Whether or not (1 or 0) to display the command name on each line output
    command_names = 0
    # A prefix to use before the message text (but after the command name,
    # if selected). Default is two spaces, so you can see/grep the severity
    # of each message.
    prefix = " "
    # To make the messages look similar to the original LVM tools use:
    # indent = 0
    # command_names = 1
    # prefix = " -- "
    # Set this if you want log messages during activation.
    # Don't use this in low memory situations (can deadlock).
    # activation = 0
    # Some debugging messages are assigned to a class and only appear
    # in debug output if the class is listed here.
    # Classes currently available:
    # memory, devices, activation, allocation, lvmetad, metadata, cache,
    # locking
    # Use "all" to see everything.
    debug_classes = [ "memory", "devices", "activation", "allocation",
    "lvmetad", "metadata", "cache", "locking" ]
    # Configuration of metadata backups and archiving. In LVM2 when we
    # talk about a 'backup' we mean making a copy of the metadata for the
    # *current* system. The 'archive' contains old metadata configurations.
    # Backups are stored in a human readable text format.
    backup {
    # Should we maintain a backup of the current metadata configuration ?
    # Use 1 for Yes; 0 for No.
    # Think very hard before turning this off!
    backup = 1
    # Where shall we keep it ?
    # Remember to back up this directory regularly!
    backup_dir = "/etc/lvm/backup"
    # Should we maintain an archive of old metadata configurations.
    # Use 1 for Yes; 0 for No.
    # On by default. Think very hard before turning this off.
    archive = 1
    # Where should archived files go ?
    # Remember to back up this directory regularly!
    archive_dir = "/etc/lvm/archive"
    # What is the minimum number of archive files you wish to keep ?
    retain_min = 10
    # What is the minimum time you wish to keep an archive file for ?
    retain_days = 30
    # Settings for the running LVM2 in shell (readline) mode.
    shell {
    # Number of lines of history to store in ~/.lvm_history
    history_size = 100
    # Miscellaneous global LVM2 settings
    global {
    # The file creation mask for any files and directories created.
    # Interpreted as octal if the first digit is zero.
    umask = 077
    # Allow other users to read the files
    #umask = 022
    # Enabling test mode means that no changes to the on disk metadata
    # will be made. Equivalent to having the -t option on every
    # command. Defaults to off.
    test = 0
    # Default value for --units argument
    units = "h"
    # Since version 2.02.54, the tools distinguish between powers of
    # 1024 bytes (e.g. KiB, MiB, GiB) and powers of 1000 bytes (e.g.
    # KB, MB, GB).
    # If you have scripts that depend on the old behaviour, set this to 0
    # temporarily until you update them.
    si_unit_consistency = 1
    # Whether or not to display unit suffix for sizes. This setting has
    # no effect if the units are in human-readable form (global/units="h")
    # in which case the suffix is always displayed.
    suffix = 1
    # Whether or not to communicate with the kernel device-mapper.
    # Set to 0 if you want to use the tools to manipulate LVM metadata
    # without activating any logical volumes.
    # If the device-mapper kernel driver is not present in your kernel
    # setting this to 0 should suppress the error messages.
    activation = 1
    # If we can't communicate with device-mapper, should we try running
    # the LVM1 tools?
    # This option only applies to 2.4 kernels and is provided to help you
    # switch between device-mapper kernels and LVM1 kernels.
    # The LVM1 tools need to be installed with .lvm1 suffices
    # e.g. vgscan.lvm1 and they will stop working after you start using
    # the new lvm2 on-disk metadata format.
    # The default value is set when the tools are built.
    # fallback_to_lvm1 = 0
    # The default metadata format that commands should use - "lvm1" or "lvm2".
    # The command line override is -M1 or -M2.
    # Defaults to "lvm2".
    # format = "lvm2"
    # Location of proc filesystem
    proc = "/proc"
    # Type of locking to use. Defaults to local file-based locking (1).
    # Turn locking off by setting to 0 (dangerous: risks metadata corruption
    # if LVM2 commands get run concurrently).
    # Type 2 uses the external shared library locking_library.
    # Type 3 uses built-in clustered locking.
    # Type 4 uses read-only locking which forbids any operations that might
    # change metadata.
    # N.B. Don't use lvmetad with locking type 3 as lvmetad is not yet
    # supported in clustered environment. If use_lvmetad=1 and locking_type=3
    # is set at the same time, LVM always issues a warning message about this
    # and then it automatically disables lvmetad use.
    locking_type = 1
    # Set to 0 to fail when a lock request cannot be satisfied immediately.
    wait_for_locks = 1
    # If using external locking (type 2) and initialisation fails,
    # with this set to 1 an attempt will be made to use the built-in
    # clustered locking.
    # If you are using a customised locking_library you should set this to 0.
    fallback_to_clustered_locking = 1
    # If an attempt to initialise type 2 or type 3 locking failed, perhaps
    # because cluster components such as clvmd are not running, with this set
    # to 1 an attempt will be made to use local file-based locking (type 1).
    # If this succeeds, only commands against local volume groups will proceed.
    # Volume Groups marked as clustered will be ignored.
    fallback_to_local_locking = 1
    # Local non-LV directory that holds file-based locks while commands are
    # in progress. A directory like /tmp that may get wiped on reboot is OK.
    locking_dir = "/run/lock/lvm"
    # Whenever there are competing read-only and read-write access requests for
    # a volume group's metadata, instead of always granting the read-only
    # requests immediately, delay them to allow the read-write requests to be
    # serviced. Without this setting, write access may be stalled by a high
    # volume of read-only requests.
    # NB. This option only affects locking_type = 1 viz. local file-based
    # locking.
    prioritise_write_locks = 1
    # Other entries can go here to allow you to load shared libraries
    # e.g. if support for LVM1 metadata was compiled as a shared library use
    # format_libraries = "liblvm2format1.so"
    # Full pathnames can be given.
    # Search this directory first for shared libraries.
    # library_dir = "/lib"
    # The external locking library to load if locking_type is set to 2.
    # locking_library = "liblvm2clusterlock.so"
    # Treat any internal errors as fatal errors, aborting the process that
    # encountered the internal error. Please only enable for debugging.
    abort_on_internal_errors = 0
    # Check whether CRC is matching when parsed VG is used multiple times.
    # This is useful to catch unexpected internal cached volume group
    # structure modification. Please only enable for debugging.
    detect_internal_vg_cache_corruption = 0
    # If set to 1, no operations that change on-disk metadata will be permitted.
    # Additionally, read-only commands that encounter metadata in need of repair
    # will still be allowed to proceed exactly as if the repair had been
    # performed (except for the unchanged vg_seqno).
    # Inappropriate use could mess up your system, so seek advice first!
    metadata_read_only = 0
    # 'mirror_segtype_default' defines which segtype will be used when the
    # shorthand '-m' option is used for mirroring. The possible options are:
    # "mirror" - The original RAID1 implementation provided by LVM2/DM. It is
    # characterized by a flexible log solution (core, disk, mirrored)
    # and by the necessity to block I/O while reconfiguring in the
    # event of a failure.
    # There is an inherent race in the dmeventd failure handling
    # logic with snapshots of devices using this type of RAID1 that
    # in the worst case could cause a deadlock.
    # Ref: https://bugzilla.redhat.com/show_bug.cgi?id=817130#c10
    # "raid1" - This implementation leverages MD's RAID1 personality through
    # device-mapper. It is characterized by a lack of log options.
    # (A log is always allocated for every device and they are placed
    # on the same device as the image - no separate devices are
    # required.) This mirror implementation does not require I/O
    # to be blocked in the kernel in the event of a failure.
    # This mirror implementation is not cluster-aware and cannot be
    # used in a shared (active/active) fashion in a cluster.
    # Specify the '--type <mirror|raid1>' option to override this default
    # setting.
    mirror_segtype_default = "raid1"
    # 'raid10_segtype_default' determines the segment types used by default
    # when the '--stripes/-i' and '--mirrors/-m' arguments are both specified
    # during the creation of a logical volume.
    # Possible settings include:
    # "raid10" - This implementation leverages MD's RAID10 personality through
    # device-mapper.
    # "mirror" - LVM will layer the 'mirror' and 'stripe' segment types. It
    # will do this by creating a mirror on top of striped sub-LVs;
    # effectively creating a RAID 0+1 array. This is suboptimal
    # in terms of providing redundancy and performance. Changing to
    # this setting is not advised.
    # Specify the '--type <raid10|mirror>' option to override this default
    # setting.
    raid10_segtype_default = "raid10"
    # The default format for displaying LV names in lvdisplay was changed
    # in version 2.02.89 to show the LV name and path separately.
    # Previously this was always shown as /dev/vgname/lvname even when that
    # was never a valid path in the /dev filesystem.
    # Set to 1 to reinstate the previous format.
    # lvdisplay_shows_full_device_path = 0
    # Whether to use (trust) a running instance of lvmetad. If this is set to
    # 0, all commands fall back to the usual scanning mechanisms. When set to 1
    # *and* when lvmetad is running (automatically instantiated by making use of
    # systemd's socket-based service activation or run as an initscripts service
    # or run manually), the volume group metadata and PV state flags are obtained
    # from the lvmetad instance and no scanning is done by the individual
    # commands. In a setup with lvmetad, lvmetad udev rules *must* be set up for
    # LVM to work correctly. Without proper udev rules, all changes in block
    # device configuration will be *ignored* until a manual 'pvscan --cache'
    # is performed. These rules are installed by default.
    # If lvmetad has been running while use_lvmetad was 0, it MUST be stopped
    # before changing use_lvmetad to 1 and started again afterwards.
    # If using lvmetad, the volume activation is also switched to automatic
    # event-based mode. In this mode, the volumes are activated based on
    # incoming udev events that automatically inform lvmetad about new PVs
    # that appear in the system. Once the VG is complete (all the PVs are
    # present), it is auto-activated. The activation/auto_activation_volume_list
    # setting controls which volumes are auto-activated (all by default).
    # A note about device filtering while lvmetad is used:
    # When lvmetad is updated (either automatically based on udev events
    # or directly by pvscan --cache <device> call), the devices/filter
    # is ignored and all devices are scanned by default. The lvmetad always
    # keeps unfiltered information which is then provided to LVM commands
    # and then each LVM command does the filtering based on devices/filter
    # setting itself.
    # To prevent scanning devices completely, even when using lvmetad,
    # the devices/global_filter must be used.
    # N.B. Don't use lvmetad with locking type 3 as lvmetad is not yet
    # supported in clustered environment. If use_lvmetad=1 and locking_type=3
    # is set at the same time, LVM always issues a warning message about this
    # and then it automatically disables lvmetad use.
    use_lvmetad = 0
    # Full path of the utility called to check that a thin metadata device
    # is in a state that allows it to be used.
    # Each time a thin pool needs to be activated or after it is deactivated
    # this utility is executed. The activation will only proceed if the utility
    # has an exit status of 0.
    # Set to "" to skip this check. (Not recommended.)
    # The thin tools are available as part of the device-mapper-persistent-data
    # package from https://github.com/jthornber/thin-provisioning-tools.
    # thin_check_executable = "/usr/bin/thin_check"
    # Array of string options passed with thin_check command. By default,
    # option "-q" is for quiet output.
    # With thin_check version 2.1 or newer you can add "--ignore-non-fatal-errors"
    # to let it pass through ignorable errors and fix them later.
    # thin_check_options = [ "-q" ]
    # Full path of the utility called to repair a thin metadata device
    # is in a state that allows it to be used.
    # Each time a thin pool needs repair this utility is executed.
    # See thin_check_executable how to obtain binaries.
    # thin_repair_executable = "/usr/bin/thin_repair"
    # Array of extra string options passed with thin_repair command.
    # thin_repair_options = [ "" ]
    # Full path of the utility called to dump thin metadata content.
    # See thin_check_executable how to obtain binaries.
    # thin_dump_executable = "/usr/bin/thin_dump"
    # If set, given features are not used by thin driver.
    # This can be helpful not just for testing, but i.e. allows to avoid
    # using problematic implementation of some thin feature.
    # Features:
    # block_size
    # discards
    # discards_non_power_2
    # external_origin
    # metadata_resize
    # external_origin_extend
    # thin_disabled_features = [ "discards", "block_size" ]
    activation {
    # Set to 1 to perform internal checks on the operations issued to
    # libdevmapper. Useful for debugging problems with activation.
    # Some of the checks may be expensive, so it's best to use this
    # only when there seems to be a problem.
    checks = 0
    # Set to 0 to disable udev synchronisation (if compiled into the binaries).
    # Processes will not wait for notification from udev.
    # They will continue irrespective of any possible udev processing
    # in the background. You should only use this if udev is not running
    # or has rules that ignore the devices LVM2 creates.
    # The command line argument --nodevsync takes precedence over this setting.
    # If set to 1 when udev is not running, and there are LVM2 processes
    # waiting for udev, run 'dmsetup udevcomplete_all' manually to wake them up.
    udev_sync = 1
    # Set to 0 to disable the udev rules installed by LVM2 (if built with
    # --enable-udev_rules). LVM2 will then manage the /dev nodes and symlinks
    # for active logical volumes directly itself.
    # N.B. Manual intervention may be required if this setting is changed
    # while any logical volumes are active.
    udev_rules = 1
    # Set to 1 for LVM2 to verify operations performed by udev. This turns on
    # additional checks (and if necessary, repairs) on entries in the device
    # directory after udev has completed processing its events.
    # Useful for diagnosing problems with LVM2/udev interactions.
    verify_udev_operations = 0
    # If set to 1 and if deactivation of an LV fails, perhaps because
    # a process run from a quick udev rule temporarily opened the device,
    # retry the operation for a few seconds before failing.
    retry_deactivation = 1
    # How to fill in missing stripes if activating an incomplete volume.
    # Using "error" will make inaccessible parts of the device return
    # I/O errors on access. You can instead use a device path, in which
    # case, that device will be used to in place of missing stripes.
    # But note that using anything other than "error" with mirrored
    # or snapshotted volumes is likely to result in data corruption.
    missing_stripe_filler = "error"
    # The linear target is an optimised version of the striped target
    # that only handles a single stripe. Set this to 0 to disable this
    # optimisation and always use the striped target.
    use_linear_target = 1
    # How much stack (in KB) to reserve for use while devices suspended
    # Prior to version 2.02.89 this used to be set to 256KB
    reserved_stack = 64
    # How much memory (in KB) to reserve for use while devices suspended
    reserved_memory = 8192
    # Nice value used while devices suspended
    process_priority = -18
    # If volume_list is defined, each LV is only activated if there is a
    # match against the list.
    # "vgname" and "vgname/lvname" are matched exactly.
    # "@tag" matches any tag set in the LV or VG.
    # "@*" matches if any tag defined on the host is also set in the LV or VG
    # If any host tags exist but volume_list is not defined, a default
    # single-entry list containing "@*" is assumed.
    # volume_list = [ "vg1", "vg2/lvol1", "@tag1", "@*" ]
    # If auto_activation_volume_list is defined, each LV that is to be
    # activated with the autoactivation option (--activate ay/-a ay) is
    # first checked against the list. There are two scenarios in which
    # the autoactivation option is used:
    # - automatic activation of volumes based on incoming PVs. If all the
    # PVs making up a VG are present in the system, the autoactivation
    # is triggered. This requires lvmetad (global/use_lvmetad=1) and udev
    # to be running. In this case, "pvscan --cache -aay" is called
    # automatically without any user intervention while processing
    # udev events. Please, make sure you define auto_activation_volume_list
    # properly so only the volumes you want and expect are autoactivated.
    # - direct activation on command line with the autoactivation option.
    # In this case, the user calls "vgchange --activate ay/-a ay" or
    # "lvchange --activate ay/-a ay" directly.
    # By default, the auto_activation_volume_list is not defined and all
    # volumes will be activated either automatically or by using --activate ay/-a ay.
    # N.B. The "activation/volume_list" is still honoured in all cases so even
    # if the VG/LV passes the auto_activation_volume_list, it still needs to
    # pass the volume_list for it to be activated in the end.
    # If auto_activation_volume_list is defined but empty, no volumes will be
    # activated automatically and --activate ay/-a ay will do nothing.
    # auto_activation_volume_list = []
    # If auto_activation_volume_list is defined and it's not empty, only matching
    # volumes will be activated either automatically or by using --activate ay/-a ay.
    # "vgname" and "vgname/lvname" are matched exactly.
    # "@tag" matches any tag set in the LV or VG.
    # "@*" matches if any tag defined on the host is also set in the LV or VG
    # auto_activation_volume_list = [ "vg1", "vg2/lvol1", "@tag1", "@*" ]
    # If read_only_volume_list is defined, each LV that is to be activated
    # is checked against the list, and if it matches, it as activated
    # in read-only mode. (This overrides '--permission rw' stored in the
    # metadata.)
    # "vgname" and "vgname/lvname" are matched exactly.
    # "@tag" matches any tag set in the LV or VG.
    # "@*" matches if any tag defined on the host is also set in the LV or VG
    # read_only_volume_list = [ "vg1", "vg2/lvol1", "@tag1", "@*" ]
    # Each LV can have an 'activation skip' flag stored persistently against it.
    # During activation, this flag is used to decide whether such an LV is skipped.
    # The 'activation skip' flag can be set during LV creation and by default it
    # is automatically set for thin snapshot LVs. The 'auto_set_activation_skip'
    # enables or disables this automatic setting of the flag while LVs are created.
    # auto_set_activation_skip = 1
    # For RAID or 'mirror' segment types, 'raid_region_size' is the
    # size (in KiB) of each:
    # - synchronization operation when initializing
    # - each copy operation when performing a 'pvmove' (using 'mirror' segtype)
    # This setting has replaced 'mirror_region_size' since version 2.02.99
    raid_region_size = 512
    # Setting to use when there is no readahead value stored in the metadata.
    # "none" - Disable readahead.
    # "auto" - Use default value chosen by kernel.
    readahead = "auto"
    # 'raid_fault_policy' defines how a device failure in a RAID logical
    # volume is handled. This includes logical volumes that have the following
    # segment types: raid1, raid4, raid5*, and raid6*.
    # In the event of a failure, the following policies will determine what
    # actions are performed during the automated response to failures (when
    # dmeventd is monitoring the RAID logical volume) and when 'lvconvert' is
    # called manually with the options '--repair' and '--use-policies'.
    # "warn" - Use the system log to warn the user that a device in the RAID
    # logical volume has failed. It is left to the user to run
    # 'lvconvert --repair' manually to remove or replace the failed
    # device. As long as the number of failed devices does not
    # exceed the redundancy of the logical volume (1 device for
    # raid4/5, 2 for raid6, etc) the logical volume will remain
    # usable.
    # "allocate" - Attempt to use any extra physical volumes in the volume
    # group as spares and replace faulty devices.
    raid_fault_policy = "warn"
    # 'mirror_image_fault_policy' and 'mirror_log_fault_policy' define
    # how a device failure affecting a mirror (of "mirror" segment type) is
    # handled. A mirror is composed of mirror images (copies) and a log.
    # A disk log ensures that a mirror does not need to be re-synced
    # (all copies made the same) every time a machine reboots or crashes.
    # In the event of a failure, the specified policy will be used to determine
    # what happens. This applies to automatic repairs (when the mirror is being
    # monitored by dmeventd) and to manual lvconvert --repair when
    # --use-policies is given.
    # "remove" - Simply remove the faulty device and run without it. If
    # the log device fails, the mirror would convert to using
    # an in-memory log. This means the mirror will not
    # remember its sync status across crashes/reboots and
    # the entire mirror will be re-synced. If a
    # mirror image fails, the mirror will convert to a
    # non-mirrored device if there is only one remaining good
    # copy.
    # "allocate" - Remove the faulty device and try to allocate space on
    # a new device to be a replacement for the failed device.
    # Using this policy for the log is fast and maintains the
    # ability to remember sync state through crashes/reboots.
    # Using this policy for a mirror device is slow, as it
    # requires the mirror to resynchronize the devices, but it
    # will preserve the mirror characteristic of the device.
    # This policy acts like "remove" if no suitable device and
    # space can be allocated for the replacement.
    # "allocate_anywhere" - Not yet implemented. Useful to place the log device
    # temporarily on same physical volume as one of the mirror
    # images. This policy is not recommended for mirror devices
    # since it would break the redundant nature of the mirror. This
    # policy acts like "remove" if no suitable device and space can
    # be allocated for the replacement.
    mirror_log_fault_policy = "allocate"
    mirror_image_fault_policy = "remove"
    # 'snapshot_autoextend_threshold' and 'snapshot_autoextend_percent' define
    # how to handle automatic snapshot extension. The former defines when the
    # snapshot should be extended: when its space usage exceeds this many
    # percent. The latter defines how much extra space should be allocated for
    # the snapshot, in percent of its current size.
    # For example, if you set snapshot_autoextend_threshold to 70 and
    # snapshot_autoextend_percent to 20, whenever a snapshot exceeds 70% usage,
    # it will be extended by another 20%. For a 1G snapshot, using up 700M will
    # trigger a resize to 1.2G. When the usage exceeds 840M, the snapshot will
    # be extended to 1.44G, and so on.
    # Setting snapshot_autoextend_threshold to 100 disables automatic
    # extensions. The minimum value is 50 (A setting below 50 will be treated
    # as 50).
    snapshot_autoextend_threshold = 100
    snapshot_autoextend_percent = 20
    # 'thin_pool_autoextend_threshold' and 'thin_pool_autoextend_percent' define
    # how to handle automatic pool extension. The former defines when the
    # pool should be extended: when its space usage exceeds this many
    # percent. The latter defines how much extra space should be allocated for
    # the pool, in percent of its current size.
    # For example, if you set thin_pool_autoextend_threshold to 70 and
    # thin_pool_autoextend_percent to 20, whenever a pool exceeds 70% usage,
    # it will be extended by another 20%. For a 1G pool, using up 700M will
    # trigger a resize to 1.2G. When the usage exceeds 840M, the pool will
    # be extended to 1.44G, and so on.
    # Setting thin_pool_autoextend_threshold to 100 disables automatic
    # extensions. The minimum value is 50 (A setting below 50 will be treated
    # as 50).
    thin_pool_autoextend_threshold = 100
    thin_pool_autoextend_percent = 20
    # While activating devices, I/O to devices being (re)configured is
    # suspended, and as a precaution against deadlocks, LVM2 needs to pin
    # any memory it is using so it is not paged out. Groups of pages that
    # are known not to be accessed during activation need not be pinned
    # into memory. Each string listed in this setting is compared against
    # each line in /proc/self/maps, and the pages corresponding to any
    # lines that match are not pinned. On some systems locale-archive was
    # found to make up over 80% of the memory used by the process.
    # mlock_filter = [ "locale/locale-archive", "gconv/gconv-modules.cache" ]
    # Set to 1 to revert to the default behaviour prior to version 2.02.62
    # which used mlockall() to pin the whole process's memory while activating
    # devices.
    use_mlockall = 0
    # Monitoring is enabled by default when activating logical volumes.
    # Set to 0 to disable monitoring or use the --ignoremonitoring option.
    monitoring = 1
    # When pvmove or lvconvert must wait for the kernel to finish
    # synchronising or merging data, they check and report progress
    # at intervals of this number of seconds. The default is 15 seconds.
    # If this is set to 0 and there is only one thing to wait for, there
    # are no progress reports, but the process is awoken immediately the
    # operation is complete.
    polling_interval = 15
    # Report settings.
    # report {
    # Align columns on report output.
    # aligned=1
    # When buffered reporting is used, the report's content is appended
    # incrementally to include each object being reported until the report
    # is flushed to output which normally happens at the end of command
    # execution. Otherwise, if buffering is not used, each object is
    # reported as soon as its processing is finished.
    # buffered=1
    # Show headings for columns on report.
    # headings=1
    # A separator to use on report after each field.
    # separator=" "
    # Use a field name prefix for each field reported.
    # prefixes=0
    # Quote field values when using field name prefixes.
    # quoted=1
    # Output each column as a row. If set, this also implies report/prefixes=1.
    # colums_as_rows=0
    # Comma separated list of columns to sort by when reporting 'lvm devtypes' command.
    # See 'lvm devtypes -o help' for the list of possible fields.
    # devtypes_sort="devtype_name"
    # Comma separated list of columns to report for 'lvm devtypes' command.
    # See 'lvm devtypes -o help' for the list of possible fields.
    # devtypes_cols="devtype_name,devtype_max_partitions,devtype_description"
    # Comma separated list of columns to report for 'lvm devtypes' command in verbose mode.
    # See 'lvm devtypes -o help' for the list of possible fields.
    # devtypes_cols_verbose="devtype_name,devtype_max_partitions,devtype_description"
    # Comma separated list of columns to sort by when reporting 'lvs' command.
    # See 'lvs -o help' for the list of possible fields.
    # lvs_sort="vg_name,lv_name"
    # Comma separated list of columns to report for 'lvs' command.
    # See 'lvs -o help' for the list of possible fields.
    # lvs_cols="lv_name,vg_name,lv_attr,lv_size,pool_lv,origin,data_percent,move_pv,mirror_log,copy_percent,convert_lv"
    # Comma separated list of columns to report for 'lvs' command in verbose mode.
    # See 'lvs -o help' for the list of possible fields.
    # lvs_cols_verbose="lv_name,vg_name,seg_count,lv_attr,lv_size,lv_major,lv_minor,lv_kernel_major,lv_kernel_minor,pool_lv,origin,data_percent,metadata_percent,move_pv,copy_percent,mirror_log,convert
    # Comma separated list of columns to sort by when reporting 'vgs' command.
    # See 'vgs -o help' for the list of possible fields.
    # vgs_sort="vg_name"
    # Comma separated list of columns to report for 'vgs' command.
    # See 'vgs -o help' for the list of possible fields.
    # vgs_cols="vg_name,pv_count,lv_count,snap_count,vg_attr,vg_size,vg_free"
    # Comma separated list of columns to report for 'vgs' command in verbose mode.
    # See 'vgs -o help' for the list of possible fields.
    # vgs_cols_verbose="vg_name,vg_attr,vg_extent_size,pv_count,lv_count,snap_count,vg_size,vg_free,vg_uuid,vg_profile"
    # Comma separated list of columns to sort by when reporting 'pvs' command.
    # See 'pvs -o help' for the list of possible fields.
    # pvs_sort="pv_name"
    # Comma separated list of columns to report for 'pvs' command.
    # See 'pvs -o help' for the list of possible fields.
    # pvs_cols="pv_name,vg_name,pv_fmt,pv_attr,pv_size,pv_free"
    # Comma separated list of columns to report for 'pvs' command in verbose mode.
    # See 'pvs -o help' for the list of possible fields.
    # pvs_cols_verbose="pv_name,vg_name,pv_fmt,pv_attr,pv_size,pv_free,dev_size,pv_uuid"
    # Comma separated list of columns to sort by when reporting 'lvs --segments' command.
    # See 'lvs --segments -o help' for the list of possible fields.
    # segs_sort="vg_name,lv_name,seg_start"
    # Comma separated list of columns to report for 'lvs --segments' command.
    # See 'lvs --segments -o help' for the list of possible fields.
    # segs_cols="lv_name,vg_name,lv_attr,stripes,segtype,seg_size"
    # Comma separated list of columns to report for 'lvs --segments' command in verbose mode.
    # See 'lvs --segments -o help' for the list of possible fields.
    # segs_cols_verbose="lv_name,vg_name,lv_attr,seg_start,seg_size,stripes,segtype,stripesize,chunksize"
    # Comma separated list of columns to sort by when reporting 'pvs --segments' command.
    # See 'pvs --segments -o help' for the list of possible fields.
    # pvsegs_sort="pv_name,pvseg_start"
    # Comma separated list of columns to sort by when reporting 'pvs --segments' command.
    # See 'pvs --segments -o help' for the list of possible fields.
    # pvsegs_cols="pv_name,vg_name,pv_fmt,pv_attr,pv_size,pv_free,pvseg_start,pvseg_size"
    # Comma separated list of columns to sort by when reporting 'pvs --segments' command in verbose mode.
    # See 'pvs --segments -o help' for the list of possible fields.
    # pvsegs_cols_verbose="pv_name,vg_name,pv_fmt,pv_attr,pv_size,pv_free,pvseg_start,pvseg_size,lv_name,seg_start_pe,segtype,seg_pe_ranges"
    # Advanced section #
    # Metadata settings
    # metadata {
    # Default number of copies of metadata to hold on each PV. 0, 1 or 2.
    # You might want to override it from the command line with 0
    # when running pvcreate on new PVs which are to be added to large VGs.
    # pvmetadatacopies = 1
    # Default number of copies of metadata to maintain for each VG.
    # If set to a non-zero value, LVM automatically chooses which of
    # the available metadata areas to use to achieve the requested
    # number of copies of the VG metadata. If you set a value larger
    # than the the total number of metadata areas available then
    # metadata is stored in them all.
    # The default value of 0 ("unmanaged") disables this automatic
    # management and allows you to control which metadata areas
    # are used at the individual PV level using 'pvchange
    # --metadataignore y/n'.
    # vgmetadatacopies = 0
    # Approximate default size of on-disk metadata areas in sectors.
    # You should increase this if you have large volume groups or
    # you want to retain a large on-disk history of your metadata changes.
    # pvmetadatasize = 255
    # List of directories holding live copies of text format metadata.
    # These directories must not be on logical volumes!
    # It's possible to use LVM2 with a couple of directories here,
    # preferably on different (non-LV) filesystems, and with no other
    # on-disk metadata (pvmetadatacopies = 0). Or this can be in
    # addition to on-disk metadata areas.
    # The feature was originally added to simplify testing and is not
    # supported under low memory situations - the machine could lock up.
    # Never edit any files in these directories by hand unless you
    # you are absolutely sure you know what you are doing! Use
    # the supplied toolset to make changes (e.g. vgcfgrestore).
    # dirs = [ "/etc/lvm/metadata", "/mnt/disk2/lvm/metadata2" ]
    # Event daemon
    dmeventd {
    # mirror_library is the library used when monitoring a mirror device.
    # "libdevmapper-event-lvm2mirror.so" attempts to recover from
    # failures. It removes failed devices from a volume group and
    # reconfigures a mirror as necessary. If no mirror library is
    # provided, mirrors are not monitored through dmeventd.
    mirror_library = "libdevmapper-event-lvm2mirror.so"
    # snapshot_library is the library used when monitoring a snapshot device.
    # "libdevmapper-event-lvm2snapshot.so" monitors the filling of
    # snapshots and emits a warning through syslog when the use of
    # the snapshot exceeds 80%. The warning is repeated when 85%, 90% and
    # 95% of the snapshot is filled.
    snapshot_library = "libdevmapper-event-lvm2snapshot.so"
    # thin_library is the library used when monitoring a thin device.
    # "libdevmapper-event-lvm2thin.so" monitors the filling of
    # pool and emits a warning through syslog when the use of
    # the pool exceeds 80%. The warning is repeated when 85%, 90% and
    # 95% of the pool is filled.
    thin_library = "libdevmapper-event-lvm2thin.so"
    # Full path of the dmeventd binary.
    # executable = "/usr/sbin/dmeventd"

  • LVM drives not showing on sucessfull boot (/root is on lvm)

    I really don't know how this is possible but I have my system booting. It works fine till it makes it to checking filesystem then it complains that it can't find device /dev/mapper/array-root. That is the location of the root filesystem the one it's booting. It then drops me to a maintenance shell. I went into /etc/fstab and changed the pass on my 2 lvm mounts to 0 and I now boot into a working system.
    When I check mount it shows that root is /dev/mapper/array-root
    but when i go to /dev/mapper/ it does not exist.
    I have uselvm="yes" in /etc/rc.conf
    I really don't know whats going on here it seems like it's not detecting the lvm volumes even though it's booting off one.
    Any help would be really great.

    Have you added 'lvm2' to your hooks array in mkinitcpio.conf?

  • Default Installation on LVM is not working

    New Thread (Default installation of LVM not working):
    First I did everything exactly as the LVM guide tells me. Then I continued with the Installation Guide.  No extra options, no change from defaults. When I try to boot it, it fails.
    I tried downgrading to 2.02.103-1. No result.
    Old Thread (when I did not know that LVM was at fault):
    Current state:
    EDIT5: OK I have boiled it down to LVM. When I copied all files from etcvol to /etc in root it continued booting until it hit mounting /home... I have no idea what the §$%&/ is wrong.
    Everything that is not root and should get mounted by fstab from LVM dies o.O
    Even stranger: After I removed homevol and swapvol from fstab i get stuck at
    A start job is running for dev-sdb4.device
    I did a "proof of concept" installation just now and it worked out of the box.
    Parameters of the "proof of concept" installation:
    /dev/sda5 ext4 /
    /dev/sda4 EFI /boot
    Rest is default installation guide
    Gummiboot:
    title Arch Linux
    linux /vmlinuz-linux
    initrd /initramfs-linux.img
    options root=/dev/sda5 rw
    Changed Parameters of the broken installation:
    /dev/sda3 LVM on LUKS
    Partitions
    /dev/mapper/MyStorage-rootvol ext4 /
    /dev/sda4 EFI /boot
    /dev/mapper/MyStorage-swapvol swap
    /dev/mapper/MyStorage-homevol ext4 /home
    title Arch Linux (Encrypted)
    linux /initramfs-linux.img
    options initrd=/initramfs-linux.img cryptdevice=/dev/sda3:MyStorage root=UUID=$UUIDofRoot rw
    I added the HOOKS
    HOOKS="... encrypt lvm2 keymap ... filesystems ... shutdown ..."
    Fstab
    # /etc/fstab: static file system information
    # <file system> <dir> <type> <options> <dump> <pass>
    # UUID=767d41f7-afb5-4a9f-a39b-604c830654e7
    /dev/mapper/MyStorage-rootvol / ext4 rw,relatime,space_cache 0 0
    # UUID=6e00082a-4fc4-47a5-9e80-e2330428d2fe
    /dev/mapper/MyStorage-etcvol /etc ext4 rw,relatime,space_cache 0 0
    # UUID=ce066cf1-230a-48b4-b7a8-2635ef39a881
    /dev/mapper/MyStorage-homevol /home ext4 rw,relatime,space_cache 0 0
    # UUID=0C88-FA2A
    /dev/sdb4 /boot vfat rw,relatime,fmask=0022,dmask=0022,codepage=437,iocharset=iso8859-1,shortname=mixed,errors=remount-ro 0 2
    # UUID=9369aa55-d055-4fab-98ef-ccc97b3f3e05
    /dev/mapper/MyStorage-swapvol none swap defaults 0 0
    So what goes wrong? I can't provide the exact logs as the journal persistency module dies in this case. If you know a way to get the actual files please tell me.
    Everything works fine down to "Reached target Local File Systems"
    He runs some isEmpty checks on tmpfiles.d directories which some succeed and /etc/tmpfiles.d and /usr/local/lib/tmpfiles.d fail
    systemd-tmpfiles --create --remove --exclude-prefix=/dev gets forked
    journald gets killed by systemctl
    systemd-tmpfile SIGCHILDs and dies
    systemd-udevd gets notified and changes from start to running
    udevd-control.socket gets changes to running
    system-tmpfiles-setup exits with failure
    Failed to start Recreate Volatile Files and Directories
    systemd-update-utmp gets forked
    accepts connection on private bus
    10 times dbus request: org.freedesktop.dbus.local.disconnect() on /org/freedesktop/dbus/local
    SIGCHILD from systemctl belonged to systemd-journal-flush.service
    Failed to start Trigger Flushing of Journal to Persistent Storage
    more dbus request: org.freedesktop.dbus.local.disconnect() on /org/freedesktop/dbus/local
    systemd-update- from systemd-update-utmp SIGCHILDs and dies
    Failed to start Update UTMP about System Reboot/Shutdown
    starts system initialization sysinit.target -> active
    starts dbus.socket to listeing
    starts sockets.target to active
    systemd-tmpfiles-clean.timer gets started by monotonic timer to cleanup
    dbus-daemon gets forked
    dbus.socket from listening to running
    systemd-logind gets forked
    Failed to start D-Bus System Message Bus
    Failed to start Login Service
    I had a look at higher debug levels, but they all just showed systemd starting something which either failed or in the case of D-Bus died soon after.
    When I broke into init, I could see the volumes which contained what they were supposed to contain.
    Except that nothing really interesting. I noticed a complaint that the AHCI driver should be replaced against a specialized one. But I don't think thats the problem.
    Thank you for your help!
    Last edited by GNA (2014-01-15 21:44:01)

    Removed the fsck filehook.
    posted the fstab.
    Boot logs will follow shortly. I just have to figure out how to get them in file form when the journal does not contain anything.
    EDIT: OK I have no idea how to get my own boot logs sad but true. ill just try to add as much context as I can see but I can't possibly write all of this down by hand.
    EDIT2: Ah no wonder when flushing of journal to persistent storage fails
    EDIT3: Ill just start decreasing the complexity one by one. And as it is so easy ill start with btrfs.
    EDIT4: Changing btrfs to ext4 did not change anything.
    EDIT5: OK I have boiled it down to LVM. When I copied all files from etcvol to /etc in root it continued booting until it hit mounting /home... I have no idea what the §$%&/ is wrong.
    Everything that is not root and should get mounted by fstab from LVM dies o.O
    Even stranger: After I removed homevol and swapvol from fstab i get stuck at
    A start job is running for dev-sdb4.device
    EDIT6: Today I tried to make a "failsafe" system with lvm and luks. Turns out it is not failsafe.
    First I did everything exactly as the LVM on LUKS guide tells me. Then I continued with the Installation Guide.  No extra options, no change from defaults. The result is the very same I currently have.
    EDIT7: Now that I think about it; There was something strange during pvcreate /dev/mapper/lvm. It complained: "/dev/sdc: open failed: No medium found" which is strange as there was never a /dev/sdc to begin with.
    I'll try to install on a pure LVM partition without crypto.
    EDIT8: OK pure install only LVM is also broken.
    Last edited by GNA (2014-01-15 20:46:49)

  • Lvm pvcreate not working (need /dev/sda9 entry)

    Hi.
    Here's my partition table layout:
    [root@arch01 gregb]# uname -a
    Linux arch01 2.6.20-ARCH #1 SMP PREEMPT Sat Mar 24 10:51:35 CET 2007 i686 Intel(R) Celeron(R) CPU 2.40GHz GenuineIntel GNU/Linux
    [root@arch01 gregb]# fdisk /dev/sda
    Command (m for help): p
    Disk /dev/sda: 251.0 GB, 251000193024 bytes
    255 heads, 63 sectors/track, 30515 cylinders
    Units = cylinders of 16065 * 512 = 8225280 bytes
    Device Boot Start End Blocks Id System
    /dev/sda1 * 1 12 96358+ 83 Linux
    /dev/sda2 13 255 1951897+ 82 Linux swap / Solaris
    /dev/sda3 256 30515 243063450 5 Extended
    /dev/sda5 256 1471 9767488+ 83 Linux
    /dev/sda6 1472 2687 9767488+ 83 Linux
    /dev/sda7 2688 3903 9767488+ 83 Linux
    /dev/sda8 3904 5119 9767488+ 83 Linux
    /dev/sda9 5120 30515 203993338+ 8e Linux LVM
    [root@arch01 gregb]# lvm pvcreate /dev/sda9
    Device /dev/sda9 not found (or ignored by filtering).
    As you can see, pvcreate cannot find /dev/sda9.
    If I look in /dev, there is no entry for /dev/sda9 (but sda1-sda8
    are there).
    What do I need to do to create the /dev/sda9 entry?
    Thanks.

    It turns out that rebooting "fixed" the problem. After rebooting pvcreate on /dev/sda9 worked.
    Anyone know the name of the command/process which adds device special files to the /dev directory
    during boot?

  • [SOLVED] LVM volume not found while booting

    I have an error while trying to boot with both Syslinux and GRUB (GRUB was my 1st choice but I got this error so switched to Syslinux). It's BIOS system, my boot is on /dev/sda1 and root is inside LVM on /dev/sda2. It states that device /dev/mapper/TheVG-RootLV can not be found so the problem is with LVM. There is lvm2 hook in mkinitcpio.conf, dolvm in my kernel line... what could I do wrong? I had no problem with this exact setup on my UEFI system and GRUB. Boot partition is loaded even if it's also inside LVM (that was the case when I was trying to run it via GRUB). GPT partition table.
    Last edited by smsware (2013-05-25 16:55:48)

    This has been an ongoing issue with the Arch Linux LVM2 setup since it was changed to using lvmetad.  You did nothing wrong.  My home server still sits on an LVM, so what I did was downgraded lvm2 and device-mapper on that machine and have been ignoring the updates until I can find a better solution... though I haven't been looking very hard.
    Also, with mkinitcpio, there is no "dolvm" command line parameter.  It runs the LVM2 stuff depending on whether or not you have the lvm hook in your mkinicpio.conf (or add it otherwise with the command line).  So yu can safely remove that.
    The reason that Grub is able to read the /boot partition, but then it can still fail to find the rootfs is because these are two separate things.  Grub has its own system to give it LVM compatibility.  So it uses its stuff to read that, gets the information it needs, then loads the kernel and initramfs. It then passes the system off, at which point, this process of detecting the LVM and finding the rootfs happens all over again, albiet with the kernel and native linux userspace tools.
    My laptop was on LVM2 when this change occurred.  It worked for a while, and then just totally borked... well not toally, but it was way inconsistent and super frusturating.  So I changed to btrfs.  There is a thread about luks and lvm2 not booting, but it is actually just about lvm2, as the luks part of it was just cioncidence.  You may want to ahve a look there because there is a slight change you can make in your lvm.conf that might help.  It helped in my case, but it only made it slightly less borked, which is why I eventually changed setups.
    Edit: I remember the edit having something to do with a filter... or global filter... something about having it not scan floppies and cdroms and whatnot.
    Last edited by WonderWoofy (2013-05-23 00:05:31)

  • Read LVM VI not returning segments of specified size.

    All,
    I'm having trouble with the Read/Write LVM VIs. I'm writing to a series of multiple files, which works fine and has worked fine for quite a while now. Now the general behavior of the Read LVM VI is that it retrieves the signals from the file in small "chunks," either at the sampling frequency or a user specified number of samples. The behavior that I'm having trouble with is that even when I force the sample size to X, the VI returns an accumulated signal that grows in length with each iteration of the main for loop. The first iteration the VI will retrieve X samples, the next 2X samples, the next 3X samples, so on and so forth. As I said before the general behavior is X samples per iteration, which is how most of my VIs work, and I can't quite figure out why this VI is operating in such a manner.
    Thanks for the help!
    Regards,
    Kenneth Miller

    Hi Kenneth,
    can you upload your code or a picture of it?
    Mike

  • Can't remove partition from LVM

    ┌─[ 09/17/12 01:21:37 PM ][ ~ ]
    └─> df
    Filesystem Size Used Avail Use% Mounted on
    rootfs 29G 6.3G 22G 23% /
    dev 1.7G 0 1.7G 0% /dev
    run 1.7G 404K 1.7G 1% /run
    /dev/sdb2 29G 6.3G 22G 23% /
    shm 1.7G 0 1.7G 0% /dev/shm
    tmpfs 1.7G 72K 1.7G 1% /tmp
    /dev/sdb3 67G 57G 7.3G 89% /home
    /dev/mapper/dabox--storage-stuff 2.9T 2.1T 667G 77% /media/stuff
    /dev/sde1 299G 113G 186G 38% /media/mydrive
    ┌─[ 09/17/12 01:21:37 PM ][ ~ ]
    └─> sudo pvscan
    PV /dev/sdc1 VG dabox-storage lvm2 [1.36 TiB / 0 free]
    PV /dev/sda2 VG dabox-storage lvm2 [244.38 GiB / 0 free]
    PV /dev/sdb1 VG dabox-storage lvm2 [366.19 GiB / 0 free]
    PV /dev/sdd1 VG dabox-storage lvm2 [931.51 GiB / 0 free]
    PV /dev/sda1 VG dabox-storage lvm2 [53.70 GiB / 0 free]
    Total: 5 [2.92 TiB] / in use: 5 [2.92 TiB] / in no VG: 0 [0 ]
    ┌─[ 09/17/12 01:21:37 PM ][ ~ ]
    └─> sudo vgscan
    Reading all physical volumes. This may take a while...
    Found volume group "dabox-storage" using metadata type lvm2
    ┌─[ 09/17/12 01:21:37 PM ][ ~ ]
    └─> sudo lvscan
    ACTIVE '/dev/dabox-storage/stuff' [2.92 TiB] inherit
    ┌─[ 09/17/12 01:21:37 PM ][ ~ ]
    └─>
    So I have 667GB free on this LV as you can see from df, yet this is happening when I try to remove a volume:
    └─> sudo pvmove -v /dev/sdb1
    Finding volume group "dabox-storage"
    No extents available for allocation
    ┌─[ 09/17/12 01:21:37 PM ][ ~ ]
    └─> sudo pvmove -v /dev/sda1
    Finding volume group "dabox-storage"
    No extents available for allocation
    I keep getting this message no matter which PV I try.
    Here's pvdisplay -m
    sudo pvdisplay -m
    [sudo] password for bobby:
    --- Physical volume ---
    PV Name /dev/sdc1
    VG Name dabox-storage
    PV Size 1.36 TiB / not usable 897.00 KiB
    Allocatable yes (but full)
    PE Size 4.00 MiB
    Total PE 357699
    Free PE 0
    Allocated PE 357699
    PV UUID Khz636-PmiK-pGwn-YIAA-V0pV-iQsp-viRTLt
    --- Physical Segments ---
    Physical extent 0 to 357698:
    Logical volume /dev/dabox-storage/stuff
    Logical extents 0 to 357698
    --- Physical volume ---
    PV Name /dev/sda2
    VG Name dabox-storage
    PV Size 244.38 GiB / not usable 2.89 MiB
    Allocatable yes (but full)
    PE Size 4.00 MiB
    Total PE 62561
    Free PE 0
    Allocated PE 62561
    PV UUID fuNir3-tgpT-wq8v-ztLv-0E08-JVS2-YA7ZCc
    --- Physical Segments ---
    Physical extent 0 to 62560:
    Logical volume /dev/dabox-storage/stuff
    Logical extents 357699 to 420259
    --- Physical volume ---
    PV Name /dev/sdb1
    VG Name dabox-storage
    PV Size 366.19 GiB / not usable 2.12 MiB
    Allocatable yes (but full)
    PE Size 4.00 MiB
    Total PE 93744
    Free PE 0
    Allocated PE 93744
    PV UUID lHtf0O-uqSh-rq9N-RHHo-RF0r-4051-zl3dNX
    --- Physical Segments ---
    Physical extent 0 to 93743:
    Logical volume /dev/dabox-storage/stuff
    Logical extents 420260 to 514003
    --- Physical volume ---
    PV Name /dev/sdd1
    VG Name dabox-storage
    PV Size 931.51 GiB / not usable 3.19 MiB
    Allocatable yes (but full)
    PE Size 4.00 MiB
    Total PE 238466
    Free PE 0
    Allocated PE 238466
    PV UUID g091ul-udvl-XmJ0-KG19-0rb4-Dag7-5GFeUH
    --- Physical Segments ---
    Physical extent 0 to 238465:
    Logical volume /dev/dabox-storage/stuff
    Logical extents 514004 to 752469
    --- Physical volume ---
    PV Name /dev/sda1
    VG Name dabox-storage
    PV Size 53.71 GiB / not usable 3.92 MiB
    Allocatable yes (but full)
    PE Size 4.00 MiB
    Total PE 13748
    Free PE 0
    Allocated PE 13748
    PV UUID YNqPYg-Zbn3-z1dj-2HlR-mwoX-aG4U-qPjV10
    --- Physical Segments ---
    Physical extent 0 to 13747:
    Logical volume /dev/dabox-storage/stuff
    Logical extents 752470 to 766217
    What I am trying to do is remove /dev/sdb1 at the least, and/or sda1+2 so I can take out that drive. I'd really appreciate any help with this folks, thanks!! (Any other output required please ask.)

    slytux wrote:
    pvmove and the partition by itselfy may not be enough, you could specify the target to move to as the second argument to pvmove.
    Ex. take out sdb1:
    pvmove /dev/sdb1 /dev/sdc1
    Yup, have tried that with all sorts of combinations:
    └─> sudo pvmove /dev/sdb1 /dev/sdc1
    No free extents on physical volume "/dev/sdc1"
    No specified PVs have space available
    ┌─[ 09/17/12 11:27:53 PM ][ ~ ]
    └─> sudo pvmove /dev/sdb1 /dev/sdd1
    No free extents on physical volume "/dev/sdd1"
    No specified PVs have space available
    ┌─[ 09/17/12 11:27:53 PM ][ ~ ]
    └─> sudo pvmove /dev/sda1 /dev/sdb1
    No free extents on physical volume "/dev/sdb1"
    No specified PVs have space available
    ┌─[ 09/17/12 11:27:53 PM ][ ~ ]
    └─> sudo pvmove /dev/sda1 /dev/sdc1
    No free extents on physical volume "/dev/sdc1"
    No specified PVs have space available
    ┌─[ 09/17/12 11:27:53 PM ][ ~ ]
    └─> sudo pvmove /dev/sda1 /dev/sdd1
    No free extents on physical volume "/dev/sdd1"
    No specified PVs have space available
    I've pretty much given up, I bought my SSD tonight, and now I'm going to buy a 2TB drive tomorrow and move my important stuff over, then dismantle the LVM and not use it again. Thanks for your help mate I appreciate it

  • Not able to create a Volume Group in the oracle VM server installation

    Hi,
    I am installing Oracle VM server. In the installtion ( Partition time ), i click on create custom layout and i want to create a / partition as a logical volumn under volume group. But there is no option to create a Volume group actually. So can anybody help how to create a volume group actually and how to create a logical volume under volume group.

    user10274248 wrote:
    So can anybody help how to create a volume group actually and how to create a logical volume under volume group.LVM is not recommended in Dom0, which is why support for LVM has been removed from the installer. LVM is not cluster aware and does not support write barriers, both of which are important for a filesystem that is used for storing file-backed disk images.

  • Partition layout greyed out – not possible to create a third partition

    Hi all,
    Being in need of a new laptop, I bought myself a Macbook Pro retina, 13inch, 512gb SSD last weekend. I’ve been a windows user for almost 20 years now, with occasional exploits into Unix/Linux, but I’ve had it in mind to try out OS X for quite a while. I like the hardware and the design of the Macbook Pro and the fact that Boot camp promised to make dual booting into win7 a walk in the park, lead me to just try it out. However…
    When I tried running boot camp (based on creating bootable USB’s):
    ISO number 1 first didn’t get recognized by boot camp
    Omitting the bootable USB creation by boot camp, downloading the drivers on a separate USB stick and using my own bootable USB lead boot camp to state that it didn’t recognize a windows install disk (even though the USB is bootable and does work)
    Creating a full new ISO and using that one to create a bootable USB by boot camp worked, but gave an error during the copying of the files (can’t create USB, error during copying windows installation files)
    Making a new image of that ISO within OS X to avoid compatibility issues gave the same error
    Using another USB stick gave the same error
    Unmounting the image still gave the same error
    At that point I gave up on wanting to use boot camp and I decided to do the install myself. Going to disk utility to create new partitions triggered a new issue:
    Via the “+”-sign, I can create a second partition on my drive. I can name that partition, format it any way I want and make it as big as I want.
    After doing that, the “-“-sign is available to remove that partition again and to merge all back into one big HFS partition.
    The tool doesn’t allow me to create 3 partitions though. The partition layout box is greyed out and fixed on “current”, and after the second partition is created, I can create a third one
    Rebooting and doing disk utility via CMD+R gives the same issue.
    I looked on the internet, found people with similar issues, but until now no real solution (unless a “boot via USB and partition from there”-solution, which I didn’t try yet).
    Why doesn’t the system allow me to create 3 partitions?
    What can I do to still create 3 partitions?
    Is working via a bootable OS X USB really the only way to do this?

    user10274248 wrote:
    So can anybody help how to create a volume group actually and how to create a logical volume under volume group.LVM is not recommended in Dom0, which is why support for LVM has been removed from the installer. LVM is not cluster aware and does not support write barriers, both of which are important for a filesystem that is used for storing file-backed disk images.

  • BIOS-GPT grub cannot find the root device in LVM setup

    Greetings Arch Linux Community,
    Recently I started my journey of installing Arch Linux on my VMware Workstation.
    I successfully installed the non-LVM instance not long before I noticed the advantages of LVM, then I just started a new Arch Linux setup with LVM.
    However, after following the Beginners' Guide and LVM Guide carefully, I ran into a problem on boot that I cannot solve by myself due to my poor knowledge on boot/grub.
    Here is the error message:
    ERROR: device '4b28d500-8ed3-4fe5-ad26-97fbb01f979b' not found. Skipping fsck
    ERROR: Unable to find root device '4b28d500-8ed3-4fe5-ad26-97fbb01f979b'
    The error message itself is pretty clear; saying that the root device specified by my grub.cfg cannot be found and thus cannot be loaded.
    However, I'm 100% sure that the device is actually there and just for some reason grub cannot see it.
    Below please see the info from lsblk, pvdisplay, vgdisplay, lvdisplay, and some related info from fstab, grub.cfg and mkinitcpio.conf:
    root@archiso / # lsblk -f
    NAME FSTYPE LABEL UUID MOUNTPOINT
    fd0
    sda
    ├─sda1
    ├─sda2 swap a576bed1-929a-4ea5-b771-6b2e3b2e3808
    ├─sda3 ext4 659e5bcf-b34b-4083-b8e5-e336fdd20c3d /mnt/boot
    └─sda4 LVM2_member GNznA5-fwUh-Gw6D-JvY4-rZKJ-d6fP-TIyoRZ
    ├─VolGroup00-lvroot ext4 4b28d500-8ed3-4fe5-ad26-97fbb01f979b /mnt
    └─VolGroup00-lvhome ext4 dc1f1322-8d8e-4456-b61d-95966ae4e25d /mnt/home
    sr0 iso9660 ARCH_201405 2014-05-01-05-20-54-00 /run/archiso/bootmnt
    loop0 squashfs /run/archiso/sfs/root-image
    loop1 ext4 f07bfb26-3377-487a-a022-efd840614a0d
    └─arch_root-image ext4 f07bfb26-3377-487a-a022-efd840614a0d /
    loop2 ext4 f07bfb26-3377-487a-a022-efd840614a0d
    └─arch_root-image ext4 f07bfb26-3377-487a-a022-efd840614a0d /
    root@archiso / # pvdisplay
    --- Physical volume ---
    PV Name /dev/sda4
    VG Name VolGroup00
    PV Size 8.50 GiB / not usable 2.98 MiB
    Allocatable yes (but full)
    PE Size 4.00 MiB
    Total PE 2175
    Free PE 0
    Allocated PE 2175
    PV UUID GNznA5-fwUh-Gw6D-JvY4-rZKJ-d6fP-TIyoRZ
    root@archiso / # vgdisplay
    --- Volume group ---
    VG Name VolGroup00
    System ID
    Format lvm2
    Metadata Areas 1
    Metadata Sequence No 3
    VG Access read/write
    VG Status resizable
    MAX LV 0
    Cur LV 2
    Open LV 2
    Max PV 0
    Cur PV 1
    Act PV 1
    VG Size 8.50 GiB
    PE Size 4.00 MiB
    Total PE 2175
    Alloc PE / Size 2175 / 8.50 GiB
    Free PE / Size 0 / 0
    VG UUID x8VegQ-OKr6-kNNk-HwHe-xxi1-Ecdt-W9Ydj2
    root@archiso / # lvdisplay
    --- Logical volume ---
    LV Path /dev/VolGroup00/lvroot
    LV Name lvroot
    VG Name VolGroup00
    LV UUID q8LxrZ-PXmf-K2Sl-7SNl-DlsF-IZ3s-2m2eHw
    LV Write Access read/write
    LV Creation host, time archiso, 2014-05-17 16:24:23 +0000
    LV Status available
    # open 1
    LV Size 5.00 GiB
    Current LE 1280
    Segments 1
    Allocation inherit
    Read ahead sectors auto
    - currently set to 256
    Block device 254:1
    --- Logical volume ---
    LV Path /dev/VolGroup00/lvhome
    LV Name lvhome
    VG Name VolGroup00
    LV UUID 3q32Nf-CRZz-iYpB-5317-075y-ulj3-uurdwE
    LV Write Access read/write
    LV Creation host, time archiso, 2014-05-17 16:24:49 +0000
    LV Status available
    # open 1
    LV Size 3.50 GiB
    Current LE 895
    Segments 1
    Allocation inherit
    Read ahead sectors auto
    - currently set to 256
    Block device 254:2
    root@archiso / # cat /mnt/etc/fstab
    # /etc/fstab: static file system information
    # <file system> <dir> <type> <options> <dump> <pass>
    # /dev/mapper/VolGroup00-lvroot
    UUID=4b28d500-8ed3-4fe5-ad26-97fbb01f979b / ext4 rw,relatime,data=ordered 0 1
    # /dev/mapper/VolGroup00-lvhome
    UUID=dc1f1322-8d8e-4456-b61d-95966ae4e25d /home ext4 rw,relatime,data=ordered 0 2
    # /dev/sda3
    UUID=659e5bcf-b34b-4083-b8e5-e336fdd20c3d /boot ext4 rw,relatime,data=ordered 0 2
    # /dev/sda2
    UUID=a576bed1-929a-4ea5-b771-6b2e3b2e3808 none swap defaults 0 0
    root@archiso / # cat /mnt/boot/grub/grub.cfg | grep "root"
    set root='lvmid/x8VegQ-OKr6-kNNk-HwHe-xxi1-Ecdt-W9Ydj2/q8LxrZ-PXmf-K2Sl-7SNl-DlsF-IZ3s-2m2eHw'
    search --no-floppy --fs-uuid --set=root --hint='lvmid/x8VegQ-OKr6-kNNk-HwHe-xxi1-Ecdt-W9Ydj2/q8LxrZ-PXmf-K2Sl-7SNl-DlsF-IZ3s-2m2eHw' 4b28d500-8ed3-4fe5-ad26-97fbb01f979b
    search --no-floppy --fs-uuid --set=root 4b28d500-8ed3-4fe5-ad26-97fbb01f979b
    set root='hd0,gpt3'
    search --no-floppy --fs-uuid --set=root --hint-bios=hd0,gpt3 --hint-efi=hd0,gpt3 --hint-baremetal=ahci0,gpt3 659e5bcf-b34b-4083-b8e5-e336fdd20c3d
    search --no-floppy --fs-uuid --set=root 659e5bcf-b34b-4083-b8e5-e336fdd20c3d
    linux /vmlinuz-linux root=UUID=4b28d500-8ed3-4fe5-ad26-97fbb01f979b rw quiet
    set root='hd0,gpt3'
    search --no-floppy --fs-uuid --set=root --hint-bios=hd0,gpt3 --hint-efi=hd0,gpt3 --hint-baremetal=ahci0,gpt3 659e5bcf-b34b-4083-b8e5-e336fdd20c3d
    search --no-floppy --fs-uuid --set=root 659e5bcf-b34b-4083-b8e5-e336fdd20c3d
    linux /vmlinuz-linux root=UUID=4b28d500-8ed3-4fe5-ad26-97fbb01f979b rw quiet
    set root='hd0,gpt3'
    search --no-floppy --fs-uuid --set=root --hint-bios=hd0,gpt3 --hint-efi=hd0,gpt3 --hint-baremetal=ahci0,gpt3 659e5bcf-b34b-4083-b8e5-e336fdd20c3d
    search --no-floppy --fs-uuid --set=root 659e5bcf-b34b-4083-b8e5-e336fdd20c3d
    linux /vmlinuz-linux root=UUID=4b28d500-8ed3-4fe5-ad26-97fbb01f979b rw quiet
    root@archiso / # cat /mnt/etc/mkinitcpio.conf| grep "HOOKS"
    # HOOKS
    # This is the most important setting in this file. The HOOKS control the
    # order in which HOOKS are added. Run 'mkinitcpio -H <hook name>' for
    # HOOKS="base"
    # HOOKS="base udev autodetect block filesystems"
    # HOOKS="base udev block filesystems"
    # HOOKS="base udev block mdadm encrypt filesystems"
    # HOOKS="base udev block lvm2 filesystems"
    HOOKS="base udev autodetect modconf block lvm2 filesystems keyboard fsck"
    Last edited by garrio (2014-05-18 00:42:36)

    Efka wrote:when i was installing Arch, i installed grub on /dev/sda and my sda (gpt )was - sda1 (ext4)for / and sda2(ext4) for /home so when grub is installed it is installed into SDA ( or im wrong? ) but not into sda1 or sda2,  but when i was doing my install i didnt made /boot ext4 partition (sda3   ext4 659e5bcf-b34b-4083-b8e5-e336fdd20c3d   /mnt/boot)  . i think u trying to mount grub into ext4 partition. is this right ? if yes... is it how it suposed to be ?
    Sorry, im newbie ( dont take my advice seriously ), but im very curious about his, because once i couldnt install grub as well so i went for syslinux instead , and i know i will have one more fresh install of arch to do, so im curious
    Hi Efka, I used grub-install and specified "/dev/sda" for argument. What I understand is that the sda is the identifier for the hard disk where all the sdaX reside. And if you look into the output of my lsblk, you will see that there is one partition sda1 that has no filetype, label and UUID on display; that is actually the bios boot partition where, as I understand, the grub code embeds itself.
    We can have different partition scheme to deploy and having a separate partition for /boot is just my personal preference ans it should work as well.

  • Multiple acquisition in the same .lvm: how to read

    Hello,
    I've generated a lvm file wich contains several set of data :
        in a while loop, I do an acquisition of 360 samples, then write it to a lvm file
                           and do an other acquisition and APPEND it to the lvm file.
    The result seems to be good. I can read this file with excel. (2 headers before the first data, the data, one header, the data, one header, and so on...)
    I'd like to read all the data with the express vi "lire un fichier de mesures"
    No problem for the first data, I obtain 360 samples.
    Then I got an error (error 100) telling that ther is some unattended data in the lvm file. (note that the eof flag is false...)
    Is it a feature of this tool, a mistake in my mind...
    Thanks

    ci-après 2 fichiers :
       le premier est séparé par des headers,
       le second non.
    Dans le second, j'ai une succession d'acquisition de 500 données. Preuve en est que si l'on ouvre avec excel, on peut supperposer les courbes. Je voudrais faire la même chose avec LV.
    Je ne comprends pas pourquoi je n'ai pas été informé par mail de cette réponse. D'où mon retard.
    Merci de jeter un oeil.
    christophe
    Attachments:
    Einstat-1à6tranche1.zip ‏558 KB
    testfrorondot5.zip ‏7 KB

  • LVM; Advantages vs. Disadvantages

    I was just browsing the wiki for some features i would like and i came across LVM. It sounds great!, i have currently three disks in my system. One 640GB, one 750GB and one brand new 1000GB drive and i would like to make them into one large volume. I have heard of Raid 0 but it's too dangerous. Then there is raid 5 but then i need all the same size disks, and i don't have that. Also i lose one disk in storage, but that's not too big of a problem. But with LVM you can use drives of different sizes. I'm just wondering, if one drive fails, do i lose all my data? Can i add and remove drives without reformatting the entire group? I'm planning on putting it on /home. So are there LVM users out there? What do you recommend?
    Last edited by Vintendo (2008-06-12 23:42:53)

    Lvm does not by default provide any redundancy (I think it has some options for mirroring volumes (raid 1), but those aren't commonly used).
    If a disk fails, you loose the data on that disk and by default lvm does not stripe data across disks (but it may be able to), so there will still be some recoverable data (it may take some effort though).
    If you have enough unallocated space, you can remove a disk, move the logical volumes while running; its quite flexible. You can combine your storage into one pool, yet still have some control over where it is stored.
    Software raid (whatever level), can work across differently sized disks, if the raid is combining similarly sized partitions (on different disks of course), and lvm can take the leftover space too.
    One of the drawbacks is that it will take a bit more effort to read the disks in another system, or from another operating system (ie. IFS does not work)
    The setup is a bit daunting at first, especially if you put root on lvm/raid, but it is pretty logical, and the documentation is decent.

  • Create ASMDISK on LVM

    the first, i created logical volume: lvcreate -L 2000M -n lvdata1 vg01 ( volume group is: vg01)
    lvcreate -L 2000M -n lvdata2 vg01
    lvcreate -L 2000M -n lvdata3 vg01
    i created the ASMDISKS: ASM1 , ASM2, ASM3 ( use asmlib): /etc/init.d/oracleasm createdisk ASM1 /dev/vg01/lvdata1,...
    when i executed: /etc/init.d/oracleasm listdisks , result as:
    DATA1
    DATA2
    ASM1
    ASM2
    ASM3
    connected to asm instance:
    then , i created a diskgroup: dgdata01 by execute cmd: create diskgroup dg03 normal redundancy disk 'ORCL:ASM1','ORCL:ASM2';
    execute commandline: show parameter asm_diskgroup , appear: dg01,dg02,dg03, successfully.
    When i restart server, then i execute: /etc/init.d/oracleasm listdisks, the result : DATA1 ,DATA2
    ls -l /dev/oracleasm/disks appear: DATA1, DATA2
    -->the asm disks: ASM1,ASM2,ASM3 not found.
    thanks,

    Hi,
    a.) Why do you want an logical volume manager below a volume manager? ASM is a volume manager, so it does not make sense to put it on top of LVM.
    b.) If this is for a cluster, this won't work anyway. LVM is not clusteraware.
    Even though: Look here for a solution:
    RHEL/OL5 oracleasm listdisks fails not list LVM-managed ASM devices after reboot (Doc ID 558596.1)
    Sebastian

  • LVM 2.0 Problem with System Refresh

    Hello,
    we're using LVM 2.0 (latest SP)  for refreshing a QA System. For testing we use the task list SAP_BASIS_COPY_REFRESH via the LVM.
    The SAP table export is working and the data is written to /tmp/VCM_pca/<SID>/vcm_exp_<SID>_2014_08_28_15_23.
    After that the system is build via LVM with NetApp volume clonening.
    The sytem is coming up and task list SAP_BASIS_COPY_REFRESH should delete table data from the source system and import the saved from the previously step.
    But we get the following error:
    Task list run SAP_BASIS_COPY_REFRESH_20140828132612 does not exist in the file system
    Executing RFC 'STC_TM_SESSION_IMPORT' with the following arguments: RFC Destination Name: 'LVM_SystemID.ER5.SystemHost.saper5.<domain>.Client.000.Import'; Input Parameters
    '{PARAM_SESSION_ID=SAP_BASIS_COPY_REFRESH_20140828132612}'.
    So LVM is not able to find the export data. When I execute the task list manually via STC02 in the new system the task list is executed and the data is imported. 
    Is it maybe a time problem?
    The export folder is named _2014_08_28_15_23  but LVM is searching for SAP_BASIS_COPY_REFRESH_20140828132612, so there is a time difference of two hours.
    Our systems have the same time.... !
    Any ideas?
    Thanks and Best Regards,
    Andreas

    Hi Andreas,
    I checked with our DEV team and below is the feedback:
    The way it works is - the refresh task-list will export the data and come to a special confirmation task where LVM confirms the task and saves the task list to the file system. Once the copy is done, LVM imports the task-list and continues with the import.
    LVM creates a folder for storing the task-list and the export data and in your case, LVM is looking in the wrong folder for the import
    task-list and the import data.
    The name of the folder & the name of the task list run are two different things.
    The time difference of two hours probably means; that LVM is in a different time zone (CEST) then the target host / system (UTC). The folder is created by LVM whereas; the task-list is run by the system. In general, this should not cause any problems however, it still makes sense to check the time-zone maintained within LVM with the time-zone of the system to rule this out completely.
    LVM adds the path to the export directory to the DEFAULT profile of the target system (ctc/export_data = …). You should check if this is the correct folder, containing the task list run. This info. can also be found in the LVM logs.
    Hope this helps!
    Best Regards,
    Adarsh

Maybe you are looking for

  • How to use my iPhone 4 to select Internet radio stations on MacBook Pro

    Okay, once again I return here to seek guidance and wisdom.  Fellow members here told me about Air Video, a terrific set of programs for iOS and OS X to stream videos from my MacBook Pro to my iPhone 4.  It is awesome! Now, my other need (?).  I want

  • Problem During Installation of ECC 6 on Windows 2003

    <b><b>Here I am facing problem during Import ABAP Phase. It is throwing Error like that:: CJS-30022 Program 'Migration Moitor' exists with return code 103. After completeing 9 stages, it's getting problem in Import packages and installation get stuck

  • Starting a Web Server

    Ok, here it goes... What I would like to do is turn a busted up Imac into a self contained home web server. My main computer is a PowerMac G3 B&W running OS9 and my frankenstien is an old G3 Imac with OS9 also that I'm going to strip down to just a m

  • Problem with folders

    i noticed that in the "info" of a video file you can set the video type to either movie, music video, or tv show. but when i look under videos on my ipod it only shows music videos and movies, but no tv shows menu, and so i cant access my tv show fil

  • Understanding Output Error Messages

    I don't understand the "heirarchy" of Flash Output Error Messages. For example: TypeError: Error #1009: Cannot access a property or method of a null object reference.     at bBrand_Gallery() ArgumentError: Error #2025: The supplied DisplayObject must