BAPI_ALM_ORDER_MAINTAIN - operation-decription INITIAL

Hi all,
i'm trying to use BAPI_ALM_ORDER_MAINTAIN to add new operations.
If i pass values to STANDARD_TEXT_KEY and to DESCRIPTION i'll find on operation only the standard key.
The standard KEY is defined without any text in transaction CA10.
If i try to set description using IW32 i have no problem, both elements are filled.
Best regards,
SAPDUMMY

Hi madhurao123, thanks for your answer.
My issue is not about text.
I'm talking about two operations fields (BAPI_ALM_ORDER_OPERATION-STANDARD_TEXT_KEY and DESCRIPTION).
If i use a STANDARD_TEXT_KEY that has not a text (transaction CA10) and i fill the field DESCRIPTION, i'll have an operation with the standard_text_key and not with the description (blank).
This is my code:
  CONCATENATE ord noper INTO t_methods-objectkey.
  t_methods-refnumber  = nmethop.
  t_methods-objecttype = 'OPERATION'.
  t_methods-method     = 'CREATE'.
  APPEND t_methods.
  CLEAR t_methods.
  t_operation-activity     = noper.
  t_operation-control_key  = 'Z002'.
  t_operation-standard_text_key  = 'PREV'.
  t_operation-work_cntr    = tg_header-mn_wk_ctr.
  t_operation-plant        = zpmtb0001-iwerk.
  CONCATENATE 'Preventivo:' preventivo '/' 'Versione:' simulazione
  INTO t_operation-description SEPARATED BY space.
  t_operation-acttype      = 'BO0003'.
  t_operation-mrp_relevant       = '1'.
  t_operation-quantity           = '1'.
  t_operation-base_uom           = 'NUM'.
  t_operation-langu              = 'I'.
  APPEND t_operation.
  CLEAR  t_operation.
  t_operation_up-activity     = 'X'.
  t_operation_up-standard_text_key  = 'X'.
  t_operation_up-control_key  = 'X'.
  t_operation_up-work_cntr    = 'X'.
  t_operation_up-plant        = 'X'.
  t_operation_up-description  = 'X'.
  t_operation_up-acttype      = 'X'.
  t_operation_up-mrp_relevant = 'X'.
  t_operation_up-quantity           = 'X'.
  t_operation_up-base_uom           = 'X'.
  t_operation-langu              = 'X'.
  APPEND t_operation_up.
  CLEAR  t_operation_up.
Thanks.

Similar Messages

  • Post operation processing initiated

    Hi,
    We are implementing Oim 11.1.1.15.4
    We have created a request template for providing access to resource ABC
    I have attached default Beneficiary Manager Approval process with this template at operation level.
    There is no other approval associated with this.
    Request level and template level approval are Auto Approval.
    I haven't applied any validation yet.
    After the beneficiary manager has approved the request the request state is changed to "Post Operation Processing Initiated"
    The task is pending with "Internal User"
    Kindly suggest where it is getting stuck?
    How to resolve it?
    Thanks in advance.

    check if you have any post operation job dependent on it which is in pending state.
    Post Operation Processing Initiated(Below definition given in Oracle doc)
    After the actual requested operation is completed, if there exists any additional operation that needs to be executed as post-processing, the request engine moves the request to the Post Operation Processing Initiated stage, before initiating those operations. This happens after Operation Completed status.

  • Post Operation Processing Initiated status after completion of approvals

    Hi All,
    I have implemented an approval work flow that has two levels of approvals. The two levels of approvals are completed successfully, but the status of the request is not changing to "Request Completed". The status of the request is showing as "Post Operation Processing Initiated".
    Please let me know what needs to be done for the status of the request to be changed to "Request Completed".
    Thanks in advance for your help.

    Check for any manual provisioning task assigned to system admin or any other user according to the configuration you have provided and complete it. "Post Operation Processing Initiated" status comes after operation level approval gets completed and some provisioning task is pending. If everything is completed check for Task to object status mapping and change it to Provisioned as suggested by Rajiv.
    If you still could not find what is the reason then connect to OIM db and query OSI and ORC tables to check the task which is in pending status.
    Edited by: Durgaprasad on Jan 28, 2013 8:58 PM

  • BAPI_ALM_ORDER_MAINTAIN operation create

    Hi, Dear All
    When I create new service order using BAPI_ALM_ORDER_MAINTAIN, How come there is only the first operation input created?
    Thanks.
    Here is my coding:
    DATA: IT_METHODS TYPE BAPI_ALM_ORDER_METHOD OCCURS 0 WITH HEADER LINE,
          IT_HEADER  TYPE BAPI_ALM_ORDER_HEADERS_I OCCURS 0 WITH HEADER LINE,
          IT_HEADER_UP TYPE BAPI_ALM_ORDER_HEADERS_UP OCCURS 0 WITH HEADER LINE,
          IT_PARTNER TYPE BAPI_ALM_ORDER_PARTN_MUL OCCURS 0 WITH HEADER LINE,
          IT_PARTNER_UP TYPE BAPI_ALM_ORDER_PARTN_MUL_UP OCCURS 0 WITH HEADER LINE,
          IT_OPERATION TYPE BAPI_ALM_ORDER_OPERATION OCCURS 0 WITH HEADER LINE,
          IT_OPERATION_UP TYPE BAPI_ALM_ORDER_OPERATION_UP OCCURS 0 WITH HEADER LINE,
          IT_SRULE TYPE BAPI_ALM_ORDER_SRULE OCCURS 0 WITH HEADER LINE,
          IT_SRULE_UP TYPE BAPI_ALM_ORDER_SRULE_UP OCCURS 0 WITH HEADER LINE,
          RETURN     TYPE BAPIRET2 OCCURS 0 WITH HEADER LINE,
          ET_NUMBERS TYPE BAPI_ALM_NUMBERS OCCURS 0 WITH HEADER LINE.
    IT_METHODS-REFNUMBER  = 1.
    IT_METHODS-OBJECTTYPE = 'OPERATION'.
    IT_METHODS-METHOD     = 'CREATE'.
    IT_METHODS-OBJECTKEY  = '000000020664'.
    APPEND IT_METHODS.
    IT_METHODS-REFNUMBER  = 1.
    IT_METHODS-OBJECTTYPE = ''.
    IT_METHODS-METHOD     = 'SAVE'.
    IT_METHODS-OBJECTKEY  = '000000020664'.
    APPEND IT_METHODS.
    IT_OPERATION-ACTIVITY = '0030'.
    IT_OPERATION-CONTROL_KEY = 'SM01'.
    IT_OPERATION-WORK_CNTR = 'M1A1'.
    IT_OPERATION-PLANT = '110A'.
    IT_OPERATION-STANDARD_TEXT_KEY = 'C2-'.
    IT_OPERATION-DESCRIPTION = 'Description 1'.
    IT_OPERATION-LANGU = 'E'.
    APPEND IT_OPERATION.
    IT_OPERATION-ACTIVITY = '0040'.
    IT_OPERATION-CONTROL_KEY = 'SM01'.
    IT_OPERATION-WORK_CNTR = 'M1A1'.
    IT_OPERATION-PLANT = '110A'.
    IT_OPERATION-STANDARD_TEXT_KEY = 'ZSTAND'.
    IT_OPERATION-DESCRIPTION = 'Description 2'.
    IT_OPERATION-LANGU = 'E'.
    APPEND IT_OPERATION.
    IT_OPERATION_UP-ACTIVITY           = 'X'.
    IT_OPERATION_UP-CONTROL_KEY        = 'X'.
    IT_OPERATION_UP-WORK_CNTR          = 'X'.
    IT_OPERATION_UP-PLANT              = 'X'.
    IT_OPERATION_UP-STANDARD_TEXT_KEY  = 'X'.
    IT_OPERATION_UP-DESCRIPTION        = 'X'.
    IT_OPERATION_UP-LANGU              = 'X'.
    APPEND IT_OPERATION_UP.
    APPEND IT_OPERATION_UP.
    CALL FUNCTION 'BAPI_ALM_ORDER_MAINTAIN'
      TABLES
        IT_METHODS              = IT_METHODS
        IT_HEADER               = IT_HEADER
        IT_HEADER_UP            = IT_HEADER_UP
      IT_HEADER_SRV           =
      IT_HEADER_SRV_UP        =
      IT_USERSTATUS           =
       IT_PARTNER              = IT_PARTNER
       IT_PARTNER_UP           = IT_PARTNER_UP
        IT_OPERATION            = IT_OPERATION
        IT_OPERATION_UP         = IT_OPERATION_UP
      IT_RELATION             =
      IT_RELATION_UP          =
      IT_COMPONENT            =
      IT_COMPONENT_UP         =
      IT_OBJECTLIST           =
      IT_OBJECTLIST_UP        =
      IT_OLIST_RELATION       =
      IT_TEXT                 =
      IT_TEXT_LINES           =
       IT_SRULE                = IT_SRULE
       IT_SRULE_UP             = IT_SRULE_UP
      IT_TASKLISTS            =
      EXTENSION_IN            =
        RETURN                  = RETURN
        ET_NUMBERS              = ET_NUMBERS
    CALL FUNCTION 'BAPI_TRANSACTION_COMMIT'
    EXPORTING
       WAIT          = 'X'

    hi,
    If you want to create two operations then we have to append one more operation line in IT_METHODS with
    REFNUMBER = 2.
    ADD the code which is in bold to your code to get required output.
    IT_METHODS-REFNUMBER = 1.
    IT_METHODS-OBJECTTYPE = 'OPERATION'.
    IT_METHODS-METHOD = 'CREATE'.
    IT_METHODS-OBJECTKEY = '000000020664'.
    APPEND IT_METHODS.
    *IT_METHODS-REFNUMBER = 2.*
    *IT_METHODS-OBJECTTYPE = 'OPERATION'.*
    *IT_METHODS-METHOD = 'CREATE'.*
    *IT_METHODS-OBJECTKEY = '000000020664'.*
    *APPEND IT_METHODS*
    IT_METHODS-REFNUMBER = 1.
    IT_METHODS-OBJECTTYPE = ''.
    IT_METHODS-METHOD = 'SAVE'.
    IT_METHODS-OBJECTKEY = '000000020664'.
    APPEND IT_METHODS.
    IT_OPERATION-ACTIVITY = '0030'.
    IT_OPERATION-CONTROL_KEY = 'SM01'.
    IT_OPERATION-WORK_CNTR = 'M1A1'.
    IT_OPERATION-PLANT = '110A'.
    IT_OPERATION-STANDARD_TEXT_KEY = 'C2-'.
    IT_OPERATION-DESCRIPTION = 'Description 1'.
    IT_OPERATION-LANGU = 'E'.
    APPEND IT_OPERATION.
    IT_OPERATION-ACTIVITY = '0040'.
    IT_OPERATION-CONTROL_KEY = 'SM01'.
    IT_OPERATION-WORK_CNTR = 'M1A1'.
    IT_OPERATION-PLANT = '110A'.
    IT_OPERATION-STANDARD_TEXT_KEY = 'ZSTAND'.
    IT_OPERATION-DESCRIPTION = 'Description 2'.
    IT_OPERATION-LANGU = 'E'.
    APPEND IT_OPERATION.
    IT_OPERATION_UP-ACTIVITY = 'X'.
    IT_OPERATION_UP-CONTROL_KEY = 'X'.
    IT_OPERATION_UP-WORK_CNTR = 'X'.
    IT_OPERATION_UP-PLANT = 'X'.
    IT_OPERATION_UP-STANDARD_TEXT_KEY = 'X'.
    IT_OPERATION_UP-DESCRIPTION = 'X'.
    IT_OPERATION_UP-LANGU = 'X'.
    APPEND IT_OPERATION_UP.
    APPEND IT_OPERATION_UP.
    CALL FUNCTION 'BAPI_ALM_ORDER_MAINTAIN'
    TABLES
    IT_METHODS = IT_METHODS
    IT_HEADER = IT_HEADER
    IT_HEADER_UP = IT_HEADER_UP
    IT_HEADER_SRV =
    IT_HEADER_SRV_UP =
    IT_USERSTATUS =
    IT_PARTNER = IT_PARTNER
    IT_PARTNER_UP = IT_PARTNER_UP
    IT_OPERATION = IT_OPERATION
    IT_OPERATION_UP = IT_OPERATION_UP
    IT_RELATION =
    IT_RELATION_UP =
    IT_COMPONENT =
    IT_COMPONENT_UP =
    IT_OBJECTLIST =
    IT_OBJECTLIST_UP =
    IT_OLIST_RELATION =
    IT_TEXT =
    IT_TEXT_LINES =
    IT_SRULE = IT_SRULE
    IT_SRULE_UP = IT_SRULE_UP
    IT_TASKLISTS =
    EXTENSION_IN =
    RETURN = RETURN
    ET_NUMBERS = ET_NUMBERS
    CALL FUNCTION 'BAPI_TRANSACTION_COMMIT'
    EXPORTING
    WAIT = 'X'
    Regards,
    Sravanthi

  • BAPI_ALM_ORDER_MAINTAIN Operation User Status Update

    Hi All,
    I am trying to use this BAPI_ALM_ORDER Maintain to update the user status at the Operation level, but am facing the following issue:
    In the IT_METHODS table, I am passing the USERSTATUS CHANGE (with the right object key - aufnr + vornr), and SAVE Method.
    And am filling the User Status table (IT_USERSTATUS) with the correct user status to be updated at the Operation Level.
    But when I execute the BAPI - it throws an error saying that the User status XXXX doesnt exist in the XXXXXX (This is the wrong profile. This profile is at the Header level and not at the operation level) Profile.
    So somehow its trying to update at the Header level and not at the operation level.
    Any idea what can be the reason?
    Regards
    Ankur

    Hi,
    BAPI_ALM_ORDER_MAINTAIN user status Change
    check out this link..
    Thanks & regards.

  • OIM 11gR2- Approved Requests remain in Operation Initiated Status

    Hi,
    We are using OIM 11gR2 and we are seeing the following behaviour whenver we request for Application Instance
    1) End User logs in and requests for Application Instance.
    2) The request is created and assigned to the manager for approval.
    3) Manager logs in and approves the request.
    4) The Requested resource is assigned to the user and the status of the resource is Provisioned. All the tasks in the Resource History are in Completed status.
    5) If we see the status of the request, it remains in Operation Initiated status.
    We expect the status of the Request to be Request Completed and not remain in Operation Initiated. We are sending a mail notification to the manager once resource is provisioned.
    If we remove the notification part, the status of the Request is coming as Request Completed as expected. However with notification getting triggered we are getting problem.
    Please suggest a solution.
    Thanks and Regards,
    Mayuri
    Edited by: 943112 on Mar 11, 2013 7:09 AM

    Usually after the request is approved and if there is any pending task in provisioning it goes to 'Post Operation Processing Initiated' status. If 'Task to Object Status' mapping is not done properly the request stays in the same status even though all the tasks are completed.
    check this link for various request status
    http://docs.oracle.com/cd/E27559_01/user.1112/e27151/req_mangmnt_user.htm#BGBGIIDH
    In your case it is going to 'Operation Initiated' status when notification is attached. Can you tell where have you triggered notification? In SOA approval task or in Provisioning process? If it is in SOA check whether proper status is returned to the callback webservice after that. Else if it is in provisioning check for task object status mapping.

  • Regarding Service Order(IW34) creation using BAPI_ALM_ORDER_MAINTAIN

    Hi Gurus,
    While Creating Service order through BAPI:BAPI_ALM_ORDER_MAINTAIN, Operations tab is only updating for One activity eventhough I Passed 4 activities. also the components tab with the same problem.
    Could any plz suggest....
    Thanks in Advance..
    Regards,
    Venkat.

    Venkat,
      Check whether you have maintained the proper methods for the key in IT_METHODS table parameter.You should be able to find the relevant values to be passed on the long text of the parameter within the BAPI.
    Regards
    Narasimhan

  • The Secure Store Service application Secure Store Service Proxy is not accessible

    I am working on setting up a new SharePoint 2013 Farm for our external web site which is currently running on 2010 SharePoint. Because this is an External Website for our organization we have an internally available Publishing Site and there is an Web App
    Extension to provide Anon Access to users using the www address. Currently I am doing everything on a single server for testing with HOSTS entries to loop back so I can test with real host headers and not impact the 2010 prod environment. We will be setting
    up 2 more WFEs with a NLB before moving this into prod but it isn't there yet. We brought the Site Collection Databases over but are still running in 2010 mode for now.
    I migrated a copy of the Secure Store Database over from 2010 and had it put on the 2013 Farm's Database Server. Set up Secure Store and the Secure Store Proxy. I went into the Secure Store Proxy in CA and refreshed the Key with the proper Pass Phrase. After
    doing that I could see the Target Application IDs listed. Everything was looking good so I went off to test. If I am on the publishing site I can go to a page that has the solution on it and it operates as expected. It does a query to a non SharePoint database
    and returns information filtered based on the search parameters. The only odd thing I see is if I click on the "Site Actions" I get a message "An error has occurred with the data fetch. Please refresh page and retry." Need to look into
    that more but it only happens on pages that have these solutions. Maybe a clue. If I go to the Anonymous Access page (www......) and try and use the solution, it immediately (way too quick to appear like it is doing anything) comes back with "The security
    token request cannot be completed."
    If I dig through the usage logs I am finding the following.
    The Secure Store Service application Secure Store Service Proxy is not accessible. The full exception text is: The security token request cannot be completed.
    Unexpected exception from endpoint address : https://[Servername]:32844/4e87fd3aabb640fb8cc3ed52188cf5c0/SecureStoreService.svc/https
    Logging unknown/unexpected client side exception: InvalidOperationException. This will cause this application server to be removed from the load balancer queue. Exception: System.InvalidOperationException: The security token request cannot be completed.
        at Microsoft.SharePoint.SPSecurityContext.SecurityTokenForServiceContext(Uri contextUri)     at Microsoft.SharePoint.SPChannelFactoryOperations.InternalCreateChannelActingAsLoggedOnUser[TChannel](ChannelFactory`1 factory, EndpointAddress
    address, Uri via)     at Microsoft.Office.SecureStoreService.Server.SecureStoreServiceApplicationProxy.GetChannel(Uri address)     at Microsoft.Office.SecureStoreService.Server.SecureStoreServiceApplicationProxy.Execute[T](String operationName,
    Boolean validateCanary, ExecuteDelegate`1 operation).
    Initially I was fighting a firewall issue because it wasn't working (different errors) on both sides. Had Network Engineers open firewall ports needed and now it works on the publishing side. Still the same error on the Anon side. I suspect I have missed
    something when it comes to configuring the Secure Store to allow access to anonymous connections.
    Let me know if you need any more information but that should paint a pretty good picture as to how things are set up.
    Thanks for any help on this one. Searches haven't found much that has helped so far.

    Hi Sennister,
    I recommend to verify the things below:
    Did this issue occur with all the pages in the anonymous side?
    Check if the Claims to Windows Token Service is started.
    Change the <identity impersonate="true" /> in the web.config to see if the issue still occurs.
    Thanks,
    Victoria
    Forum Support
    Please remember to mark the replies as answers if they help and unmark them if they provide no help. If you have feedback for TechNet Subscriber Support, contact
    [email protected]
    Victoria Xia
    TechNet Community Support

  • LVM Volumes not available after update

    Hi All!
    I haven't updated my system for about two months and today I updated it. Now I have the problem that I cannot boot properly. I have my root partition in an LVM volume and on boot I get the message
    ERROR: device 'UUID=xxx' not found. Skipping fs
    ERROR: Unable to find root device 'UUID=xxx'
    After that I land in the recovery shell. After some research I found, that "lvm lvdisplay" showed that my volumes where not available and I had to reenable them with "lvm vgchange -a y".
    Issuing any lvm command also produced the following warning:
    WARNING: lvmetad is running but disabled. Restart lvmetad before enabling it!
    Anyway, after issuing the commands and exiting the recovery shell, the system booted again. However, I would prefer being able to boot without manual actions.
    Thanks in advance!
    Further information:
    vgdisplay
    --- Volume group ---
    VG Name ArchLVM
    System ID
    Format lvm2
    Metadata Areas 1
    Metadata Sequence No 3
    VG Access read/write
    VG Status resizable
    MAX LV 0
    Cur LV 2
    Open LV 1
    Max PV 0
    Cur PV 1
    Act PV 1
    VG Size 232.69 GiB
    PE Size 4.00 MiB
    Total PE 59568
    Alloc PE / Size 59568 / 232.69 GiB
    Free PE / Size 0 / 0
    VG UUID SoB3M1-v1fD-1abI-PNJ3-6IOn-FfdI-0RoLK5
    lvdisplay (LV Status was 'not available' right after booting)
    --- Logical volume ---
    LV Path /dev/ArchLVM/Swap
    LV Name Swap
    VG Name ArchLVM
    LV UUID XRYBrz-LojR-k6SD-XIxV-wHnY-f3VG-giKL6V
    LV Write Access read/write
    LV Creation host, time archiso, 2014-05-16 14:43:06 +0200
    LV Status available
    # open 0
    LV Size 8.00 GiB
    Current LE 2048
    Segments 1
    Allocation inherit
    Read ahead sectors auto
    - currently set to 256
    Block device 254:0
    --- Logical volume ---
    LV Path /dev/ArchLVM/Root
    LV Name Root
    VG Name ArchLVM
    LV UUID lpjDl4-Jqzu-ZWkq-Uphc-IaOo-6Rzd-cIh5yv
    LV Write Access read/write
    LV Creation host, time archiso, 2014-05-16 14:43:27 +0200
    LV Status available
    # open 1
    LV Size 224.69 GiB
    Current LE 57520
    Segments 1
    Allocation inherit
    Read ahead sectors auto
    - currently set to 256
    Block device 254:1
    /etc/fstab
    # /etc/fstab: static file system information
    # <file system> <dir> <type> <options> <dump> <pass>
    # /dev/mapper/ArchLVM-Root
    UUID=2db82d1a-47a4-4e30-a819-143e8fb75199 / ext4 rw,relatime,data=ordered 0 1
    #/dev/mapper/ArchLVM-Root / ext4 rw,relatime,data=ordered 0 1
    # /dev/sda1
    UUID=72691888-a781-4cdd-a98e-2613d87925d0 /boot ext2 rw,relatime 0 2
    /etc/mkinitcpio.conf
    # vim:set ft=sh
    # MODULES
    # The following modules are loaded before any boot hooks are
    # run. Advanced users may wish to specify all system modules
    # in this array. For instance:
    # MODULES="piix ide_disk reiserfs"
    MODULES=""
    # BINARIES
    # This setting includes any additional binaries a given user may
    # wish into the CPIO image. This is run last, so it may be used to
    # override the actual binaries included by a given hook
    # BINARIES are dependency parsed, so you may safely ignore libraries
    BINARIES=""
    # FILES
    # This setting is similar to BINARIES above, however, files are added
    # as-is and are not parsed in any way. This is useful for config files.
    FILES=""
    # HOOKS
    # This is the most important setting in this file. The HOOKS control the
    # modules and scripts added to the image, and what happens at boot time.
    # Order is important, and it is recommended that you do not change the
    # order in which HOOKS are added. Run 'mkinitcpio -H <hook name>' for
    # help on a given hook.
    # 'base' is _required_ unless you know precisely what you are doing.
    # 'udev' is _required_ in order to automatically load modules
    # 'filesystems' is _required_ unless you specify your fs modules in MODULES
    # Examples:
    ## This setup specifies all modules in the MODULES setting above.
    ## No raid, lvm2, or encrypted root is needed.
    # HOOKS="base"
    ## This setup will autodetect all modules for your system and should
    ## work as a sane default
    # HOOKS="base udev autodetect block filesystems"
    ## This setup will generate a 'full' image which supports most systems.
    ## No autodetection is done.
    # HOOKS="base udev block filesystems"
    ## This setup assembles a pata mdadm array with an encrypted root FS.
    ## Note: See 'mkinitcpio -H mdadm' for more information on raid devices.
    # HOOKS="base udev block mdadm encrypt filesystems"
    ## This setup loads an lvm2 volume group on a usb device.
    # HOOKS="base udev block lvm2 filesystems"
    ## NOTE: If you have /usr on a separate partition, you MUST include the
    # usr, fsck and shutdown hooks.
    HOOKS="base udev autodetect modconf block lvm2 filesystems keyboard fsck"
    # COMPRESSION
    # Use this to compress the initramfs image. By default, gzip compression
    # is used. Use 'cat' to create an uncompressed image.
    #COMPRESSION="gzip"
    #COMPRESSION="bzip2"
    #COMPRESSION="lzma"
    #COMPRESSION="xz"
    #COMPRESSION="lzop"
    #COMPRESSION="lz4"
    # COMPRESSION_OPTIONS
    # Additional options for the compressor
    #COMPRESSION_OPTIONS=""
    /boot/grub/grub.cfg
    # DO NOT EDIT THIS FILE
    # It is automatically generated by grub-mkconfig using templates
    # from /etc/grub.d and settings from /etc/default/grub
    ### BEGIN /etc/grub.d/00_header ###
    insmod part_gpt
    insmod part_msdos
    if [ -s $prefix/grubenv ]; then
    load_env
    fi
    if [ "${next_entry}" ] ; then
    set default="${next_entry}"
    set next_entry=
    save_env next_entry
    set boot_once=true
    else
    set default="0"
    fi
    if [ x"${feature_menuentry_id}" = xy ]; then
    menuentry_id_option="--id"
    else
    menuentry_id_option=""
    fi
    export menuentry_id_option
    if [ "${prev_saved_entry}" ]; then
    set saved_entry="${prev_saved_entry}"
    save_env saved_entry
    set prev_saved_entry=
    save_env prev_saved_entry
    set boot_once=true
    fi
    function savedefault {
    if [ -z "${boot_once}" ]; then
    saved_entry="${chosen}"
    save_env saved_entry
    fi
    function load_video {
    if [ x$feature_all_video_module = xy ]; then
    insmod all_video
    else
    insmod efi_gop
    insmod efi_uga
    insmod ieee1275_fb
    insmod vbe
    insmod vga
    insmod video_bochs
    insmod video_cirrus
    fi
    if [ x$feature_default_font_path = xy ] ; then
    font=unicode
    else
    insmod part_msdos
    insmod lvm
    insmod ext2
    set root='lvmid/SoB3M1-v1fD-1abI-PNJ3-6IOn-FfdI-0RoLK5/lpjDl4-Jqzu-ZWkq-Uphc-IaOo-6Rzd-cIh5yv'
    if [ x$feature_platform_search_hint = xy ]; then
    search --no-floppy --fs-uuid --set=root --hint='lvmid/SoB3M1-v1fD-1abI-PNJ3-6IOn-FfdI-0RoLK5/lpjDl4-Jqzu-ZWkq-Uphc-IaOo-6Rzd-cIh5yv' 2db82d1a-47a4-4e30-a819-143e8fb75199
    else
    search --no-floppy --fs-uuid --set=root 2db82d1a-47a4-4e30-a819-143e8fb75199
    fi
    font="/usr/share/grub/unicode.pf2"
    fi
    if loadfont $font ; then
    set gfxmode=auto
    load_video
    insmod gfxterm
    fi
    terminal_input console
    terminal_output gfxterm
    if [ x$feature_timeout_style = xy ] ; then
    set timeout_style=menu
    set timeout=5
    # Fallback normal timeout code in case the timeout_style feature is
    # unavailable.
    else
    set timeout=5
    fi
    ### END /etc/grub.d/00_header ###
    ### BEGIN /etc/grub.d/10_linux ###
    menuentry 'Arch Linux' --class arch --class gnu-linux --class gnu --class os $menuentry_id_option 'gnulinux-simple-2db82d1a-47a4-4e30-a819-143e8fb75199' {
    load_video
    set gfxpayload=keep
    insmod gzio
    insmod part_msdos
    insmod ext2
    set root='hd0,msdos1'
    if [ x$feature_platform_search_hint = xy ]; then
    search --no-floppy --fs-uuid --set=root --hint-bios=hd0,msdos1 --hint-efi=hd0,msdos1 --hint-baremetal=ahci0,msdos1 72691888-a781-4cdd-a98e-2613d87925d0
    else
    search --no-floppy --fs-uuid --set=root 72691888-a781-4cdd-a98e-2613d87925d0
    fi
    echo 'Loading Linux linux ...'
    linux /vmlinuz-linux root=UUID=2db82d1a-47a4-4e30-a819-143e8fb75199 rw quiet
    echo 'Loading initial ramdisk ...'
    initrd /initramfs-linux.img
    submenu 'Advanced options for Arch Linux' $menuentry_id_option 'gnulinux-advanced-2db82d1a-47a4-4e30-a819-143e8fb75199' {
    menuentry 'Arch Linux, with Linux linux' --class arch --class gnu-linux --class gnu --class os $menuentry_id_option 'gnulinux-linux-advanced-2db82d1a-47a4-4e30-a819-143e8fb75199' {
    load_video
    set gfxpayload=keep
    insmod gzio
    insmod part_msdos
    insmod ext2
    set root='hd0,msdos1'
    if [ x$feature_platform_search_hint = xy ]; then
    search --no-floppy --fs-uuid --set=root --hint-bios=hd0,msdos1 --hint-efi=hd0,msdos1 --hint-baremetal=ahci0,msdos1 72691888-a781-4cdd-a98e-2613d87925d0
    else
    search --no-floppy --fs-uuid --set=root 72691888-a781-4cdd-a98e-2613d87925d0
    fi
    echo 'Loading Linux linux ...'
    linux /vmlinuz-linux root=UUID=2db82d1a-47a4-4e30-a819-143e8fb75199 rw quiet
    echo 'Loading initial ramdisk ...'
    initrd /initramfs-linux.img
    menuentry 'Arch Linux, with Linux linux (fallback initramfs)' --class arch --class gnu-linux --class gnu --class os $menuentry_id_option 'gnulinux-linux-fallback-2db82d1a-47a4-4e30-a819-143e8fb75199' {
    load_video
    set gfxpayload=keep
    insmod gzio
    insmod part_msdos
    insmod ext2
    set root='hd0,msdos1'
    if [ x$feature_platform_search_hint = xy ]; then
    search --no-floppy --fs-uuid --set=root --hint-bios=hd0,msdos1 --hint-efi=hd0,msdos1 --hint-baremetal=ahci0,msdos1 72691888-a781-4cdd-a98e-2613d87925d0
    else
    search --no-floppy --fs-uuid --set=root 72691888-a781-4cdd-a98e-2613d87925d0
    fi
    echo 'Loading Linux linux ...'
    linux /vmlinuz-linux root=UUID=2db82d1a-47a4-4e30-a819-143e8fb75199 rw quiet
    echo 'Loading initial ramdisk ...'
    initrd /initramfs-linux-fallback.img
    ### END /etc/grub.d/10_linux ###
    ### BEGIN /etc/grub.d/20_linux_xen ###
    ### END /etc/grub.d/20_linux_xen ###
    ### BEGIN /etc/grub.d/30_os-prober ###
    ### END /etc/grub.d/30_os-prober ###
    ### BEGIN /etc/grub.d/40_custom ###
    # This file provides an easy way to add custom menu entries. Simply type the
    # menu entries you want to add after this comment. Be careful not to change
    # the 'exec tail' line above.
    ### END /etc/grub.d/40_custom ###
    ### BEGIN /etc/grub.d/41_custom ###
    if [ -f ${config_directory}/custom.cfg ]; then
    source ${config_directory}/custom.cfg
    elif [ -z "${config_directory}" -a -f $prefix/custom.cfg ]; then
    source $prefix/custom.cfg;
    fi
    ### END /etc/grub.d/41_custom ###
    ### BEGIN /etc/grub.d/60_memtest86+ ###
    ### END /etc/grub.d/60_memtest86+ ###
    Last edited by Kirodema (2014-07-16 07:31:34)

    use_lvmetad = 0
    lvm2-lvmetad is not enabled or running on my system. Shall I activate it?
    # This is an example configuration file for the LVM2 system.
    # It contains the default settings that would be used if there was no
    # /etc/lvm/lvm.conf file.
    # Refer to 'man lvm.conf' for further information including the file layout.
    # To put this file in a different directory and override /etc/lvm set
    # the environment variable LVM_SYSTEM_DIR before running the tools.
    # N.B. Take care that each setting only appears once if uncommenting
    # example settings in this file.
    # This section allows you to set the way the configuration settings are handled.
    config {
    # If enabled, any LVM2 configuration mismatch is reported.
    # This implies checking that the configuration key is understood
    # by LVM2 and that the value of the key is of a proper type.
    # If disabled, any configuration mismatch is ignored and default
    # value is used instead without any warning (a message about the
    # configuration key not being found is issued in verbose mode only).
    checks = 1
    # If enabled, any configuration mismatch aborts the LVM2 process.
    abort_on_errors = 0
    # Directory where LVM looks for configuration profiles.
    profile_dir = "/etc/lvm/profile"
    # This section allows you to configure which block devices should
    # be used by the LVM system.
    devices {
    # Where do you want your volume groups to appear ?
    dir = "/dev"
    # An array of directories that contain the device nodes you wish
    # to use with LVM2.
    scan = [ "/dev" ]
    # If set, the cache of block device nodes with all associated symlinks
    # will be constructed out of the existing udev database content.
    # This avoids using and opening any inapplicable non-block devices or
    # subdirectories found in the device directory. This setting is applied
    # to udev-managed device directory only, other directories will be scanned
    # fully. LVM2 needs to be compiled with udev support for this setting to
    # take effect. N.B. Any device node or symlink not managed by udev in
    # udev directory will be ignored with this setting on.
    obtain_device_list_from_udev = 1
    # If several entries in the scanned directories correspond to the
    # same block device and the tools need to display a name for device,
    # all the pathnames are matched against each item in the following
    # list of regular expressions in turn and the first match is used.
    preferred_names = [ ]
    # Try to avoid using undescriptive /dev/dm-N names, if present.
    # preferred_names = [ "^/dev/mpath/", "^/dev/mapper/mpath", "^/dev/[hs]d" ]
    # A filter that tells LVM2 to only use a restricted set of devices.
    # The filter consists of an array of regular expressions. These
    # expressions can be delimited by a character of your choice, and
    # prefixed with either an 'a' (for accept) or 'r' (for reject).
    # The first expression found to match a device name determines if
    # the device will be accepted or rejected (ignored). Devices that
    # don't match any patterns are accepted.
    # Be careful if there there are symbolic links or multiple filesystem
    # entries for the same device as each name is checked separately against
    # the list of patterns. The effect is that if the first pattern in the
    # list to match a name is an 'a' pattern for any of the names, the device
    # is accepted; otherwise if the first pattern in the list to match a name
    # is an 'r' pattern for any of the names it is rejected; otherwise it is
    # accepted.
    # Don't have more than one filter line active at once: only one gets used.
    # Run vgscan after you change this parameter to ensure that
    # the cache file gets regenerated (see below).
    # If it doesn't do what you expect, check the output of 'vgscan -vvvv'.
    # If lvmetad is used, then see "A note about device filtering while
    # lvmetad is used" comment that is attached to global/use_lvmetad setting.
    # By default we accept every block device:
    filter = [ "a/.*/" ]
    # Exclude the cdrom drive
    # filter = [ "r|/dev/cdrom|" ]
    # When testing I like to work with just loopback devices:
    # filter = [ "a/loop/", "r/.*/" ]
    # Or maybe all loops and ide drives except hdc:
    # filter =[ "a|loop|", "r|/dev/hdc|", "a|/dev/ide|", "r|.*|" ]
    # Use anchors if you want to be really specific
    # filter = [ "a|^/dev/hda8$|", "r/.*/" ]
    # Since "filter" is often overridden from command line, it is not suitable
    # for system-wide device filtering (udev rules, lvmetad). To hide devices
    # from LVM-specific udev processing and/or from lvmetad, you need to set
    # global_filter. The syntax is the same as for normal "filter"
    # above. Devices that fail the global_filter are not even opened by LVM.
    # global_filter = []
    # The results of the filtering are cached on disk to avoid
    # rescanning dud devices (which can take a very long time).
    # By default this cache is stored in the /etc/lvm/cache directory
    # in a file called '.cache'.
    # It is safe to delete the contents: the tools regenerate it.
    # (The old setting 'cache' is still respected if neither of
    # these new ones is present.)
    # N.B. If obtain_device_list_from_udev is set to 1 the list of
    # devices is instead obtained from udev and any existing .cache
    # file is removed.
    cache_dir = "/etc/lvm/cache"
    cache_file_prefix = ""
    # You can turn off writing this cache file by setting this to 0.
    write_cache_state = 1
    # Advanced settings.
    # List of pairs of additional acceptable block device types found
    # in /proc/devices with maximum (non-zero) number of partitions.
    # types = [ "fd", 16 ]
    # If sysfs is mounted (2.6 kernels) restrict device scanning to
    # the block devices it believes are valid.
    # 1 enables; 0 disables.
    sysfs_scan = 1
    # By default, LVM2 will ignore devices used as component paths
    # of device-mapper multipath devices.
    # 1 enables; 0 disables.
    multipath_component_detection = 1
    # By default, LVM2 will ignore devices used as components of
    # software RAID (md) devices by looking for md superblocks.
    # 1 enables; 0 disables.
    md_component_detection = 1
    # By default, if a PV is placed directly upon an md device, LVM2
    # will align its data blocks with the md device's stripe-width.
    # 1 enables; 0 disables.
    md_chunk_alignment = 1
    # Default alignment of the start of a data area in MB. If set to 0,
    # a value of 64KB will be used. Set to 1 for 1MiB, 2 for 2MiB, etc.
    # default_data_alignment = 1
    # By default, the start of a PV's data area will be a multiple of
    # the 'minimum_io_size' or 'optimal_io_size' exposed in sysfs.
    # - minimum_io_size - the smallest request the device can perform
    # w/o incurring a read-modify-write penalty (e.g. MD's chunk size)
    # - optimal_io_size - the device's preferred unit of receiving I/O
    # (e.g. MD's stripe width)
    # minimum_io_size is used if optimal_io_size is undefined (0).
    # If md_chunk_alignment is enabled, that detects the optimal_io_size.
    # This setting takes precedence over md_chunk_alignment.
    # 1 enables; 0 disables.
    data_alignment_detection = 1
    # Alignment (in KB) of start of data area when creating a new PV.
    # md_chunk_alignment and data_alignment_detection are disabled if set.
    # Set to 0 for the default alignment (see: data_alignment_default)
    # or page size, if larger.
    data_alignment = 0
    # By default, the start of the PV's aligned data area will be shifted by
    # the 'alignment_offset' exposed in sysfs. This offset is often 0 but
    # may be non-zero; e.g.: certain 4KB sector drives that compensate for
    # windows partitioning will have an alignment_offset of 3584 bytes
    # (sector 7 is the lowest aligned logical block, the 4KB sectors start
    # at LBA -1, and consequently sector 63 is aligned on a 4KB boundary).
    # But note that pvcreate --dataalignmentoffset will skip this detection.
    # 1 enables; 0 disables.
    data_alignment_offset_detection = 1
    # If, while scanning the system for PVs, LVM2 encounters a device-mapper
    # device that has its I/O suspended, it waits for it to become accessible.
    # Set this to 1 to skip such devices. This should only be needed
    # in recovery situations.
    ignore_suspended_devices = 0
    # ignore_lvm_mirrors: Introduced in version 2.02.104
    # This setting determines whether logical volumes of "mirror" segment
    # type are scanned for LVM labels. This affects the ability of
    # mirrors to be used as physical volumes. If 'ignore_lvm_mirrors'
    # is set to '1', it becomes impossible to create volume groups on top
    # of mirror logical volumes - i.e. to stack volume groups on mirrors.
    # Allowing mirror logical volumes to be scanned (setting the value to '0')
    # can potentially cause LVM processes and I/O to the mirror to become
    # blocked. This is due to the way that the "mirror" segment type handles
    # failures. In order for the hang to manifest itself, an LVM command must
    # be run just after a failure and before the automatic LVM repair process
    # takes place OR there must be failures in multiple mirrors in the same
    # volume group at the same time with write failures occurring moments
    # before a scan of the mirror's labels.
    # Note that these scanning limitations do not apply to the LVM RAID
    # types, like "raid1". The RAID segment types handle failures in a
    # different way and are not subject to possible process or I/O blocking.
    # It is encouraged that users set 'ignore_lvm_mirrors' to 1 if they
    # are using the "mirror" segment type. Users that require volume group
    # stacking on mirrored logical volumes should consider using the "raid1"
    # segment type. The "raid1" segment type is not available for
    # active/active clustered volume groups.
    # Set to 1 to disallow stacking and thereby avoid a possible deadlock.
    ignore_lvm_mirrors = 1
    # During each LVM operation errors received from each device are counted.
    # If the counter of a particular device exceeds the limit set here, no
    # further I/O is sent to that device for the remainder of the respective
    # operation. Setting the parameter to 0 disables the counters altogether.
    disable_after_error_count = 0
    # Allow use of pvcreate --uuid without requiring --restorefile.
    require_restorefile_with_uuid = 1
    # Minimum size (in KB) of block devices which can be used as PVs.
    # In a clustered environment all nodes must use the same value.
    # Any value smaller than 512KB is ignored.
    # Ignore devices smaller than 2MB such as floppy drives.
    pv_min_size = 2048
    # The original built-in setting was 512 up to and including version 2.02.84.
    # pv_min_size = 512
    # Issue discards to a logical volumes's underlying physical volume(s) when
    # the logical volume is no longer using the physical volumes' space (e.g.
    # lvremove, lvreduce, etc). Discards inform the storage that a region is
    # no longer in use. Storage that supports discards advertise the protocol
    # specific way discards should be issued by the kernel (TRIM, UNMAP, or
    # WRITE SAME with UNMAP bit set). Not all storage will support or benefit
    # from discards but SSDs and thinly provisioned LUNs generally do. If set
    # to 1, discards will only be issued if both the storage and kernel provide
    # support.
    # 1 enables; 0 disables.
    issue_discards = 0
    # This section allows you to configure the way in which LVM selects
    # free space for its Logical Volumes.
    allocation {
    # When searching for free space to extend an LV, the "cling"
    # allocation policy will choose space on the same PVs as the last
    # segment of the existing LV. If there is insufficient space and a
    # list of tags is defined here, it will check whether any of them are
    # attached to the PVs concerned and then seek to match those PV tags
    # between existing extents and new extents.
    # Use the special tag "@*" as a wildcard to match any PV tag.
    # Example: LVs are mirrored between two sites within a single VG.
    # PVs are tagged with either @site1 or @site2 to indicate where
    # they are situated.
    # cling_tag_list = [ "@site1", "@site2" ]
    # cling_tag_list = [ "@*" ]
    # Changes made in version 2.02.85 extended the reach of the 'cling'
    # policies to detect more situations where data can be grouped
    # onto the same disks. Set this to 0 to revert to the previous
    # algorithm.
    maximise_cling = 1
    # Whether to use blkid library instead of native LVM2 code to detect
    # any existing signatures while creating new Physical Volumes and
    # Logical Volumes. LVM2 needs to be compiled with blkid wiping support
    # for this setting to take effect.
    # LVM2 native detection code is currently able to recognize these signatures:
    # - MD device signature
    # - swap signature
    # - LUKS signature
    # To see the list of signatures recognized by blkid, check the output
    # of 'blkid -k' command. The blkid can recognize more signatures than
    # LVM2 native detection code, but due to this higher number of signatures
    # to be recognized, it can take more time to complete the signature scan.
    use_blkid_wiping = 1
    # Set to 1 to wipe any signatures found on newly-created Logical Volumes
    # automatically in addition to zeroing of the first KB on the LV
    # (controlled by the -Z/--zero y option).
    # The command line option -W/--wipesignatures takes precedence over this
    # setting.
    # The default is to wipe signatures when zeroing.
    wipe_signatures_when_zeroing_new_lvs = 1
    # Set to 1 to guarantee that mirror logs will always be placed on
    # different PVs from the mirror images. This was the default
    # until version 2.02.85.
    mirror_logs_require_separate_pvs = 0
    # Set to 1 to guarantee that cache_pool metadata will always be
    # placed on different PVs from the cache_pool data.
    cache_pool_metadata_require_separate_pvs = 0
    # Specify the minimal chunk size (in kiB) for cache pool volumes.
    # Using a chunk_size that is too large can result in wasteful use of
    # the cache, where small reads and writes can cause large sections of
    # an LV to be mapped into the cache. However, choosing a chunk_size
    # that is too small can result in more overhead trying to manage the
    # numerous chunks that become mapped into the cache. The former is
    # more of a problem than the latter in most cases, so we default to
    # a value that is on the smaller end of the spectrum. Supported values
    # range from 32(kiB) to 1048576 in multiples of 32.
    # cache_pool_chunk_size = 64
    # Set to 1 to guarantee that thin pool metadata will always
    # be placed on different PVs from the pool data.
    thin_pool_metadata_require_separate_pvs = 0
    # Specify chunk size calculation policy for thin pool volumes.
    # Possible options are:
    # "generic" - if thin_pool_chunk_size is defined, use it.
    # Otherwise, calculate the chunk size based on
    # estimation and device hints exposed in sysfs:
    # the minimum_io_size. The chunk size is always
    # at least 64KiB.
    # "performance" - if thin_pool_chunk_size is defined, use it.
    # Otherwise, calculate the chunk size for
    # performance based on device hints exposed in
    # sysfs: the optimal_io_size. The chunk size is
    # always at least 512KiB.
    # thin_pool_chunk_size_policy = "generic"
    # Specify the minimal chunk size (in KB) for thin pool volumes.
    # Use of the larger chunk size may improve performance for plain
    # thin volumes, however using them for snapshot volumes is less efficient,
    # as it consumes more space and takes extra time for copying.
    # When unset, lvm tries to estimate chunk size starting from 64KB
    # Supported values are in range from 64 to 1048576.
    # thin_pool_chunk_size = 64
    # Specify discards behaviour of the thin pool volume.
    # Select one of "ignore", "nopassdown", "passdown"
    # thin_pool_discards = "passdown"
    # Set to 0, to disable zeroing of thin pool data chunks before their
    # first use.
    # N.B. zeroing larger thin pool chunk size degrades performance.
    # thin_pool_zero = 1
    # This section that allows you to configure the nature of the
    # information that LVM2 reports.
    log {
    # Controls the messages sent to stdout or stderr.
    # There are three levels of verbosity, 3 being the most verbose.
    verbose = 0
    # Set to 1 to suppress all non-essential messages from stdout.
    # This has the same effect as -qq.
    # When this is set, the following commands still produce output:
    # dumpconfig, lvdisplay, lvmdiskscan, lvs, pvck, pvdisplay,
    # pvs, version, vgcfgrestore -l, vgdisplay, vgs.
    # Non-essential messages are shifted from log level 4 to log level 5
    # for syslog and lvm2_log_fn purposes.
    # Any 'yes' or 'no' questions not overridden by other arguments
    # are suppressed and default to 'no'.
    silent = 0
    # Should we send log messages through syslog?
    # 1 is yes; 0 is no.
    syslog = 1
    # Should we log error and debug messages to a file?
    # By default there is no log file.
    #file = "/var/log/lvm2.log"
    # Should we overwrite the log file each time the program is run?
    # By default we append.
    overwrite = 0
    # What level of log messages should we send to the log file and/or syslog?
    # There are 6 syslog-like log levels currently in use - 2 to 7 inclusive.
    # 7 is the most verbose (LOG_DEBUG).
    level = 0
    # Format of output messages
    # Whether or not (1 or 0) to indent messages according to their severity
    indent = 1
    # Whether or not (1 or 0) to display the command name on each line output
    command_names = 0
    # A prefix to use before the message text (but after the command name,
    # if selected). Default is two spaces, so you can see/grep the severity
    # of each message.
    prefix = " "
    # To make the messages look similar to the original LVM tools use:
    # indent = 0
    # command_names = 1
    # prefix = " -- "
    # Set this if you want log messages during activation.
    # Don't use this in low memory situations (can deadlock).
    # activation = 0
    # Some debugging messages are assigned to a class and only appear
    # in debug output if the class is listed here.
    # Classes currently available:
    # memory, devices, activation, allocation, lvmetad, metadata, cache,
    # locking
    # Use "all" to see everything.
    debug_classes = [ "memory", "devices", "activation", "allocation",
    "lvmetad", "metadata", "cache", "locking" ]
    # Configuration of metadata backups and archiving. In LVM2 when we
    # talk about a 'backup' we mean making a copy of the metadata for the
    # *current* system. The 'archive' contains old metadata configurations.
    # Backups are stored in a human readable text format.
    backup {
    # Should we maintain a backup of the current metadata configuration ?
    # Use 1 for Yes; 0 for No.
    # Think very hard before turning this off!
    backup = 1
    # Where shall we keep it ?
    # Remember to back up this directory regularly!
    backup_dir = "/etc/lvm/backup"
    # Should we maintain an archive of old metadata configurations.
    # Use 1 for Yes; 0 for No.
    # On by default. Think very hard before turning this off.
    archive = 1
    # Where should archived files go ?
    # Remember to back up this directory regularly!
    archive_dir = "/etc/lvm/archive"
    # What is the minimum number of archive files you wish to keep ?
    retain_min = 10
    # What is the minimum time you wish to keep an archive file for ?
    retain_days = 30
    # Settings for the running LVM2 in shell (readline) mode.
    shell {
    # Number of lines of history to store in ~/.lvm_history
    history_size = 100
    # Miscellaneous global LVM2 settings
    global {
    # The file creation mask for any files and directories created.
    # Interpreted as octal if the first digit is zero.
    umask = 077
    # Allow other users to read the files
    #umask = 022
    # Enabling test mode means that no changes to the on disk metadata
    # will be made. Equivalent to having the -t option on every
    # command. Defaults to off.
    test = 0
    # Default value for --units argument
    units = "h"
    # Since version 2.02.54, the tools distinguish between powers of
    # 1024 bytes (e.g. KiB, MiB, GiB) and powers of 1000 bytes (e.g.
    # KB, MB, GB).
    # If you have scripts that depend on the old behaviour, set this to 0
    # temporarily until you update them.
    si_unit_consistency = 1
    # Whether or not to display unit suffix for sizes. This setting has
    # no effect if the units are in human-readable form (global/units="h")
    # in which case the suffix is always displayed.
    suffix = 1
    # Whether or not to communicate with the kernel device-mapper.
    # Set to 0 if you want to use the tools to manipulate LVM metadata
    # without activating any logical volumes.
    # If the device-mapper kernel driver is not present in your kernel
    # setting this to 0 should suppress the error messages.
    activation = 1
    # If we can't communicate with device-mapper, should we try running
    # the LVM1 tools?
    # This option only applies to 2.4 kernels and is provided to help you
    # switch between device-mapper kernels and LVM1 kernels.
    # The LVM1 tools need to be installed with .lvm1 suffices
    # e.g. vgscan.lvm1 and they will stop working after you start using
    # the new lvm2 on-disk metadata format.
    # The default value is set when the tools are built.
    # fallback_to_lvm1 = 0
    # The default metadata format that commands should use - "lvm1" or "lvm2".
    # The command line override is -M1 or -M2.
    # Defaults to "lvm2".
    # format = "lvm2"
    # Location of proc filesystem
    proc = "/proc"
    # Type of locking to use. Defaults to local file-based locking (1).
    # Turn locking off by setting to 0 (dangerous: risks metadata corruption
    # if LVM2 commands get run concurrently).
    # Type 2 uses the external shared library locking_library.
    # Type 3 uses built-in clustered locking.
    # Type 4 uses read-only locking which forbids any operations that might
    # change metadata.
    # N.B. Don't use lvmetad with locking type 3 as lvmetad is not yet
    # supported in clustered environment. If use_lvmetad=1 and locking_type=3
    # is set at the same time, LVM always issues a warning message about this
    # and then it automatically disables lvmetad use.
    locking_type = 1
    # Set to 0 to fail when a lock request cannot be satisfied immediately.
    wait_for_locks = 1
    # If using external locking (type 2) and initialisation fails,
    # with this set to 1 an attempt will be made to use the built-in
    # clustered locking.
    # If you are using a customised locking_library you should set this to 0.
    fallback_to_clustered_locking = 1
    # If an attempt to initialise type 2 or type 3 locking failed, perhaps
    # because cluster components such as clvmd are not running, with this set
    # to 1 an attempt will be made to use local file-based locking (type 1).
    # If this succeeds, only commands against local volume groups will proceed.
    # Volume Groups marked as clustered will be ignored.
    fallback_to_local_locking = 1
    # Local non-LV directory that holds file-based locks while commands are
    # in progress. A directory like /tmp that may get wiped on reboot is OK.
    locking_dir = "/run/lock/lvm"
    # Whenever there are competing read-only and read-write access requests for
    # a volume group's metadata, instead of always granting the read-only
    # requests immediately, delay them to allow the read-write requests to be
    # serviced. Without this setting, write access may be stalled by a high
    # volume of read-only requests.
    # NB. This option only affects locking_type = 1 viz. local file-based
    # locking.
    prioritise_write_locks = 1
    # Other entries can go here to allow you to load shared libraries
    # e.g. if support for LVM1 metadata was compiled as a shared library use
    # format_libraries = "liblvm2format1.so"
    # Full pathnames can be given.
    # Search this directory first for shared libraries.
    # library_dir = "/lib"
    # The external locking library to load if locking_type is set to 2.
    # locking_library = "liblvm2clusterlock.so"
    # Treat any internal errors as fatal errors, aborting the process that
    # encountered the internal error. Please only enable for debugging.
    abort_on_internal_errors = 0
    # Check whether CRC is matching when parsed VG is used multiple times.
    # This is useful to catch unexpected internal cached volume group
    # structure modification. Please only enable for debugging.
    detect_internal_vg_cache_corruption = 0
    # If set to 1, no operations that change on-disk metadata will be permitted.
    # Additionally, read-only commands that encounter metadata in need of repair
    # will still be allowed to proceed exactly as if the repair had been
    # performed (except for the unchanged vg_seqno).
    # Inappropriate use could mess up your system, so seek advice first!
    metadata_read_only = 0
    # 'mirror_segtype_default' defines which segtype will be used when the
    # shorthand '-m' option is used for mirroring. The possible options are:
    # "mirror" - The original RAID1 implementation provided by LVM2/DM. It is
    # characterized by a flexible log solution (core, disk, mirrored)
    # and by the necessity to block I/O while reconfiguring in the
    # event of a failure.
    # There is an inherent race in the dmeventd failure handling
    # logic with snapshots of devices using this type of RAID1 that
    # in the worst case could cause a deadlock.
    # Ref: https://bugzilla.redhat.com/show_bug.cgi?id=817130#c10
    # "raid1" - This implementation leverages MD's RAID1 personality through
    # device-mapper. It is characterized by a lack of log options.
    # (A log is always allocated for every device and they are placed
    # on the same device as the image - no separate devices are
    # required.) This mirror implementation does not require I/O
    # to be blocked in the kernel in the event of a failure.
    # This mirror implementation is not cluster-aware and cannot be
    # used in a shared (active/active) fashion in a cluster.
    # Specify the '--type <mirror|raid1>' option to override this default
    # setting.
    mirror_segtype_default = "raid1"
    # 'raid10_segtype_default' determines the segment types used by default
    # when the '--stripes/-i' and '--mirrors/-m' arguments are both specified
    # during the creation of a logical volume.
    # Possible settings include:
    # "raid10" - This implementation leverages MD's RAID10 personality through
    # device-mapper.
    # "mirror" - LVM will layer the 'mirror' and 'stripe' segment types. It
    # will do this by creating a mirror on top of striped sub-LVs;
    # effectively creating a RAID 0+1 array. This is suboptimal
    # in terms of providing redundancy and performance. Changing to
    # this setting is not advised.
    # Specify the '--type <raid10|mirror>' option to override this default
    # setting.
    raid10_segtype_default = "raid10"
    # The default format for displaying LV names in lvdisplay was changed
    # in version 2.02.89 to show the LV name and path separately.
    # Previously this was always shown as /dev/vgname/lvname even when that
    # was never a valid path in the /dev filesystem.
    # Set to 1 to reinstate the previous format.
    # lvdisplay_shows_full_device_path = 0
    # Whether to use (trust) a running instance of lvmetad. If this is set to
    # 0, all commands fall back to the usual scanning mechanisms. When set to 1
    # *and* when lvmetad is running (automatically instantiated by making use of
    # systemd's socket-based service activation or run as an initscripts service
    # or run manually), the volume group metadata and PV state flags are obtained
    # from the lvmetad instance and no scanning is done by the individual
    # commands. In a setup with lvmetad, lvmetad udev rules *must* be set up for
    # LVM to work correctly. Without proper udev rules, all changes in block
    # device configuration will be *ignored* until a manual 'pvscan --cache'
    # is performed. These rules are installed by default.
    # If lvmetad has been running while use_lvmetad was 0, it MUST be stopped
    # before changing use_lvmetad to 1 and started again afterwards.
    # If using lvmetad, the volume activation is also switched to automatic
    # event-based mode. In this mode, the volumes are activated based on
    # incoming udev events that automatically inform lvmetad about new PVs
    # that appear in the system. Once the VG is complete (all the PVs are
    # present), it is auto-activated. The activation/auto_activation_volume_list
    # setting controls which volumes are auto-activated (all by default).
    # A note about device filtering while lvmetad is used:
    # When lvmetad is updated (either automatically based on udev events
    # or directly by pvscan --cache <device> call), the devices/filter
    # is ignored and all devices are scanned by default. The lvmetad always
    # keeps unfiltered information which is then provided to LVM commands
    # and then each LVM command does the filtering based on devices/filter
    # setting itself.
    # To prevent scanning devices completely, even when using lvmetad,
    # the devices/global_filter must be used.
    # N.B. Don't use lvmetad with locking type 3 as lvmetad is not yet
    # supported in clustered environment. If use_lvmetad=1 and locking_type=3
    # is set at the same time, LVM always issues a warning message about this
    # and then it automatically disables lvmetad use.
    use_lvmetad = 0
    # Full path of the utility called to check that a thin metadata device
    # is in a state that allows it to be used.
    # Each time a thin pool needs to be activated or after it is deactivated
    # this utility is executed. The activation will only proceed if the utility
    # has an exit status of 0.
    # Set to "" to skip this check. (Not recommended.)
    # The thin tools are available as part of the device-mapper-persistent-data
    # package from https://github.com/jthornber/thin-provisioning-tools.
    # thin_check_executable = "/usr/bin/thin_check"
    # Array of string options passed with thin_check command. By default,
    # option "-q" is for quiet output.
    # With thin_check version 2.1 or newer you can add "--ignore-non-fatal-errors"
    # to let it pass through ignorable errors and fix them later.
    # thin_check_options = [ "-q" ]
    # Full path of the utility called to repair a thin metadata device
    # is in a state that allows it to be used.
    # Each time a thin pool needs repair this utility is executed.
    # See thin_check_executable how to obtain binaries.
    # thin_repair_executable = "/usr/bin/thin_repair"
    # Array of extra string options passed with thin_repair command.
    # thin_repair_options = [ "" ]
    # Full path of the utility called to dump thin metadata content.
    # See thin_check_executable how to obtain binaries.
    # thin_dump_executable = "/usr/bin/thin_dump"
    # If set, given features are not used by thin driver.
    # This can be helpful not just for testing, but i.e. allows to avoid
    # using problematic implementation of some thin feature.
    # Features:
    # block_size
    # discards
    # discards_non_power_2
    # external_origin
    # metadata_resize
    # external_origin_extend
    # thin_disabled_features = [ "discards", "block_size" ]
    activation {
    # Set to 1 to perform internal checks on the operations issued to
    # libdevmapper. Useful for debugging problems with activation.
    # Some of the checks may be expensive, so it's best to use this
    # only when there seems to be a problem.
    checks = 0
    # Set to 0 to disable udev synchronisation (if compiled into the binaries).
    # Processes will not wait for notification from udev.
    # They will continue irrespective of any possible udev processing
    # in the background. You should only use this if udev is not running
    # or has rules that ignore the devices LVM2 creates.
    # The command line argument --nodevsync takes precedence over this setting.
    # If set to 1 when udev is not running, and there are LVM2 processes
    # waiting for udev, run 'dmsetup udevcomplete_all' manually to wake them up.
    udev_sync = 1
    # Set to 0 to disable the udev rules installed by LVM2 (if built with
    # --enable-udev_rules). LVM2 will then manage the /dev nodes and symlinks
    # for active logical volumes directly itself.
    # N.B. Manual intervention may be required if this setting is changed
    # while any logical volumes are active.
    udev_rules = 1
    # Set to 1 for LVM2 to verify operations performed by udev. This turns on
    # additional checks (and if necessary, repairs) on entries in the device
    # directory after udev has completed processing its events.
    # Useful for diagnosing problems with LVM2/udev interactions.
    verify_udev_operations = 0
    # If set to 1 and if deactivation of an LV fails, perhaps because
    # a process run from a quick udev rule temporarily opened the device,
    # retry the operation for a few seconds before failing.
    retry_deactivation = 1
    # How to fill in missing stripes if activating an incomplete volume.
    # Using "error" will make inaccessible parts of the device return
    # I/O errors on access. You can instead use a device path, in which
    # case, that device will be used to in place of missing stripes.
    # But note that using anything other than "error" with mirrored
    # or snapshotted volumes is likely to result in data corruption.
    missing_stripe_filler = "error"
    # The linear target is an optimised version of the striped target
    # that only handles a single stripe. Set this to 0 to disable this
    # optimisation and always use the striped target.
    use_linear_target = 1
    # How much stack (in KB) to reserve for use while devices suspended
    # Prior to version 2.02.89 this used to be set to 256KB
    reserved_stack = 64
    # How much memory (in KB) to reserve for use while devices suspended
    reserved_memory = 8192
    # Nice value used while devices suspended
    process_priority = -18
    # If volume_list is defined, each LV is only activated if there is a
    # match against the list.
    # "vgname" and "vgname/lvname" are matched exactly.
    # "@tag" matches any tag set in the LV or VG.
    # "@*" matches if any tag defined on the host is also set in the LV or VG
    # If any host tags exist but volume_list is not defined, a default
    # single-entry list containing "@*" is assumed.
    # volume_list = [ "vg1", "vg2/lvol1", "@tag1", "@*" ]
    # If auto_activation_volume_list is defined, each LV that is to be
    # activated with the autoactivation option (--activate ay/-a ay) is
    # first checked against the list. There are two scenarios in which
    # the autoactivation option is used:
    # - automatic activation of volumes based on incoming PVs. If all the
    # PVs making up a VG are present in the system, the autoactivation
    # is triggered. This requires lvmetad (global/use_lvmetad=1) and udev
    # to be running. In this case, "pvscan --cache -aay" is called
    # automatically without any user intervention while processing
    # udev events. Please, make sure you define auto_activation_volume_list
    # properly so only the volumes you want and expect are autoactivated.
    # - direct activation on command line with the autoactivation option.
    # In this case, the user calls "vgchange --activate ay/-a ay" or
    # "lvchange --activate ay/-a ay" directly.
    # By default, the auto_activation_volume_list is not defined and all
    # volumes will be activated either automatically or by using --activate ay/-a ay.
    # N.B. The "activation/volume_list" is still honoured in all cases so even
    # if the VG/LV passes the auto_activation_volume_list, it still needs to
    # pass the volume_list for it to be activated in the end.
    # If auto_activation_volume_list is defined but empty, no volumes will be
    # activated automatically and --activate ay/-a ay will do nothing.
    # auto_activation_volume_list = []
    # If auto_activation_volume_list is defined and it's not empty, only matching
    # volumes will be activated either automatically or by using --activate ay/-a ay.
    # "vgname" and "vgname/lvname" are matched exactly.
    # "@tag" matches any tag set in the LV or VG.
    # "@*" matches if any tag defined on the host is also set in the LV or VG
    # auto_activation_volume_list = [ "vg1", "vg2/lvol1", "@tag1", "@*" ]
    # If read_only_volume_list is defined, each LV that is to be activated
    # is checked against the list, and if it matches, it as activated
    # in read-only mode. (This overrides '--permission rw' stored in the
    # metadata.)
    # "vgname" and "vgname/lvname" are matched exactly.
    # "@tag" matches any tag set in the LV or VG.
    # "@*" matches if any tag defined on the host is also set in the LV or VG
    # read_only_volume_list = [ "vg1", "vg2/lvol1", "@tag1", "@*" ]
    # Each LV can have an 'activation skip' flag stored persistently against it.
    # During activation, this flag is used to decide whether such an LV is skipped.
    # The 'activation skip' flag can be set during LV creation and by default it
    # is automatically set for thin snapshot LVs. The 'auto_set_activation_skip'
    # enables or disables this automatic setting of the flag while LVs are created.
    # auto_set_activation_skip = 1
    # For RAID or 'mirror' segment types, 'raid_region_size' is the
    # size (in KiB) of each:
    # - synchronization operation when initializing
    # - each copy operation when performing a 'pvmove' (using 'mirror' segtype)
    # This setting has replaced 'mirror_region_size' since version 2.02.99
    raid_region_size = 512
    # Setting to use when there is no readahead value stored in the metadata.
    # "none" - Disable readahead.
    # "auto" - Use default value chosen by kernel.
    readahead = "auto"
    # 'raid_fault_policy' defines how a device failure in a RAID logical
    # volume is handled. This includes logical volumes that have the following
    # segment types: raid1, raid4, raid5*, and raid6*.
    # In the event of a failure, the following policies will determine what
    # actions are performed during the automated response to failures (when
    # dmeventd is monitoring the RAID logical volume) and when 'lvconvert' is
    # called manually with the options '--repair' and '--use-policies'.
    # "warn" - Use the system log to warn the user that a device in the RAID
    # logical volume has failed. It is left to the user to run
    # 'lvconvert --repair' manually to remove or replace the failed
    # device. As long as the number of failed devices does not
    # exceed the redundancy of the logical volume (1 device for
    # raid4/5, 2 for raid6, etc) the logical volume will remain
    # usable.
    # "allocate" - Attempt to use any extra physical volumes in the volume
    # group as spares and replace faulty devices.
    raid_fault_policy = "warn"
    # 'mirror_image_fault_policy' and 'mirror_log_fault_policy' define
    # how a device failure affecting a mirror (of "mirror" segment type) is
    # handled. A mirror is composed of mirror images (copies) and a log.
    # A disk log ensures that a mirror does not need to be re-synced
    # (all copies made the same) every time a machine reboots or crashes.
    # In the event of a failure, the specified policy will be used to determine
    # what happens. This applies to automatic repairs (when the mirror is being
    # monitored by dmeventd) and to manual lvconvert --repair when
    # --use-policies is given.
    # "remove" - Simply remove the faulty device and run without it. If
    # the log device fails, the mirror would convert to using
    # an in-memory log. This means the mirror will not
    # remember its sync status across crashes/reboots and
    # the entire mirror will be re-synced. If a
    # mirror image fails, the mirror will convert to a
    # non-mirrored device if there is only one remaining good
    # copy.
    # "allocate" - Remove the faulty device and try to allocate space on
    # a new device to be a replacement for the failed device.
    # Using this policy for the log is fast and maintains the
    # ability to remember sync state through crashes/reboots.
    # Using this policy for a mirror device is slow, as it
    # requires the mirror to resynchronize the devices, but it
    # will preserve the mirror characteristic of the device.
    # This policy acts like "remove" if no suitable device and
    # space can be allocated for the replacement.
    # "allocate_anywhere" - Not yet implemented. Useful to place the log device
    # temporarily on same physical volume as one of the mirror
    # images. This policy is not recommended for mirror devices
    # since it would break the redundant nature of the mirror. This
    # policy acts like "remove" if no suitable device and space can
    # be allocated for the replacement.
    mirror_log_fault_policy = "allocate"
    mirror_image_fault_policy = "remove"
    # 'snapshot_autoextend_threshold' and 'snapshot_autoextend_percent' define
    # how to handle automatic snapshot extension. The former defines when the
    # snapshot should be extended: when its space usage exceeds this many
    # percent. The latter defines how much extra space should be allocated for
    # the snapshot, in percent of its current size.
    # For example, if you set snapshot_autoextend_threshold to 70 and
    # snapshot_autoextend_percent to 20, whenever a snapshot exceeds 70% usage,
    # it will be extended by another 20%. For a 1G snapshot, using up 700M will
    # trigger a resize to 1.2G. When the usage exceeds 840M, the snapshot will
    # be extended to 1.44G, and so on.
    # Setting snapshot_autoextend_threshold to 100 disables automatic
    # extensions. The minimum value is 50 (A setting below 50 will be treated
    # as 50).
    snapshot_autoextend_threshold = 100
    snapshot_autoextend_percent = 20
    # 'thin_pool_autoextend_threshold' and 'thin_pool_autoextend_percent' define
    # how to handle automatic pool extension. The former defines when the
    # pool should be extended: when its space usage exceeds this many
    # percent. The latter defines how much extra space should be allocated for
    # the pool, in percent of its current size.
    # For example, if you set thin_pool_autoextend_threshold to 70 and
    # thin_pool_autoextend_percent to 20, whenever a pool exceeds 70% usage,
    # it will be extended by another 20%. For a 1G pool, using up 700M will
    # trigger a resize to 1.2G. When the usage exceeds 840M, the pool will
    # be extended to 1.44G, and so on.
    # Setting thin_pool_autoextend_threshold to 100 disables automatic
    # extensions. The minimum value is 50 (A setting below 50 will be treated
    # as 50).
    thin_pool_autoextend_threshold = 100
    thin_pool_autoextend_percent = 20
    # While activating devices, I/O to devices being (re)configured is
    # suspended, and as a precaution against deadlocks, LVM2 needs to pin
    # any memory it is using so it is not paged out. Groups of pages that
    # are known not to be accessed during activation need not be pinned
    # into memory. Each string listed in this setting is compared against
    # each line in /proc/self/maps, and the pages corresponding to any
    # lines that match are not pinned. On some systems locale-archive was
    # found to make up over 80% of the memory used by the process.
    # mlock_filter = [ "locale/locale-archive", "gconv/gconv-modules.cache" ]
    # Set to 1 to revert to the default behaviour prior to version 2.02.62
    # which used mlockall() to pin the whole process's memory while activating
    # devices.
    use_mlockall = 0
    # Monitoring is enabled by default when activating logical volumes.
    # Set to 0 to disable monitoring or use the --ignoremonitoring option.
    monitoring = 1
    # When pvmove or lvconvert must wait for the kernel to finish
    # synchronising or merging data, they check and report progress
    # at intervals of this number of seconds. The default is 15 seconds.
    # If this is set to 0 and there is only one thing to wait for, there
    # are no progress reports, but the process is awoken immediately the
    # operation is complete.
    polling_interval = 15
    # Report settings.
    # report {
    # Align columns on report output.
    # aligned=1
    # When buffered reporting is used, the report's content is appended
    # incrementally to include each object being reported until the report
    # is flushed to output which normally happens at the end of command
    # execution. Otherwise, if buffering is not used, each object is
    # reported as soon as its processing is finished.
    # buffered=1
    # Show headings for columns on report.
    # headings=1
    # A separator to use on report after each field.
    # separator=" "
    # Use a field name prefix for each field reported.
    # prefixes=0
    # Quote field values when using field name prefixes.
    # quoted=1
    # Output each column as a row. If set, this also implies report/prefixes=1.
    # colums_as_rows=0
    # Comma separated list of columns to sort by when reporting 'lvm devtypes' command.
    # See 'lvm devtypes -o help' for the list of possible fields.
    # devtypes_sort="devtype_name"
    # Comma separated list of columns to report for 'lvm devtypes' command.
    # See 'lvm devtypes -o help' for the list of possible fields.
    # devtypes_cols="devtype_name,devtype_max_partitions,devtype_description"
    # Comma separated list of columns to report for 'lvm devtypes' command in verbose mode.
    # See 'lvm devtypes -o help' for the list of possible fields.
    # devtypes_cols_verbose="devtype_name,devtype_max_partitions,devtype_description"
    # Comma separated list of columns to sort by when reporting 'lvs' command.
    # See 'lvs -o help' for the list of possible fields.
    # lvs_sort="vg_name,lv_name"
    # Comma separated list of columns to report for 'lvs' command.
    # See 'lvs -o help' for the list of possible fields.
    # lvs_cols="lv_name,vg_name,lv_attr,lv_size,pool_lv,origin,data_percent,move_pv,mirror_log,copy_percent,convert_lv"
    # Comma separated list of columns to report for 'lvs' command in verbose mode.
    # See 'lvs -o help' for the list of possible fields.
    # lvs_cols_verbose="lv_name,vg_name,seg_count,lv_attr,lv_size,lv_major,lv_minor,lv_kernel_major,lv_kernel_minor,pool_lv,origin,data_percent,metadata_percent,move_pv,copy_percent,mirror_log,convert
    # Comma separated list of columns to sort by when reporting 'vgs' command.
    # See 'vgs -o help' for the list of possible fields.
    # vgs_sort="vg_name"
    # Comma separated list of columns to report for 'vgs' command.
    # See 'vgs -o help' for the list of possible fields.
    # vgs_cols="vg_name,pv_count,lv_count,snap_count,vg_attr,vg_size,vg_free"
    # Comma separated list of columns to report for 'vgs' command in verbose mode.
    # See 'vgs -o help' for the list of possible fields.
    # vgs_cols_verbose="vg_name,vg_attr,vg_extent_size,pv_count,lv_count,snap_count,vg_size,vg_free,vg_uuid,vg_profile"
    # Comma separated list of columns to sort by when reporting 'pvs' command.
    # See 'pvs -o help' for the list of possible fields.
    # pvs_sort="pv_name"
    # Comma separated list of columns to report for 'pvs' command.
    # See 'pvs -o help' for the list of possible fields.
    # pvs_cols="pv_name,vg_name,pv_fmt,pv_attr,pv_size,pv_free"
    # Comma separated list of columns to report for 'pvs' command in verbose mode.
    # See 'pvs -o help' for the list of possible fields.
    # pvs_cols_verbose="pv_name,vg_name,pv_fmt,pv_attr,pv_size,pv_free,dev_size,pv_uuid"
    # Comma separated list of columns to sort by when reporting 'lvs --segments' command.
    # See 'lvs --segments -o help' for the list of possible fields.
    # segs_sort="vg_name,lv_name,seg_start"
    # Comma separated list of columns to report for 'lvs --segments' command.
    # See 'lvs --segments -o help' for the list of possible fields.
    # segs_cols="lv_name,vg_name,lv_attr,stripes,segtype,seg_size"
    # Comma separated list of columns to report for 'lvs --segments' command in verbose mode.
    # See 'lvs --segments -o help' for the list of possible fields.
    # segs_cols_verbose="lv_name,vg_name,lv_attr,seg_start,seg_size,stripes,segtype,stripesize,chunksize"
    # Comma separated list of columns to sort by when reporting 'pvs --segments' command.
    # See 'pvs --segments -o help' for the list of possible fields.
    # pvsegs_sort="pv_name,pvseg_start"
    # Comma separated list of columns to sort by when reporting 'pvs --segments' command.
    # See 'pvs --segments -o help' for the list of possible fields.
    # pvsegs_cols="pv_name,vg_name,pv_fmt,pv_attr,pv_size,pv_free,pvseg_start,pvseg_size"
    # Comma separated list of columns to sort by when reporting 'pvs --segments' command in verbose mode.
    # See 'pvs --segments -o help' for the list of possible fields.
    # pvsegs_cols_verbose="pv_name,vg_name,pv_fmt,pv_attr,pv_size,pv_free,pvseg_start,pvseg_size,lv_name,seg_start_pe,segtype,seg_pe_ranges"
    # Advanced section #
    # Metadata settings
    # metadata {
    # Default number of copies of metadata to hold on each PV. 0, 1 or 2.
    # You might want to override it from the command line with 0
    # when running pvcreate on new PVs which are to be added to large VGs.
    # pvmetadatacopies = 1
    # Default number of copies of metadata to maintain for each VG.
    # If set to a non-zero value, LVM automatically chooses which of
    # the available metadata areas to use to achieve the requested
    # number of copies of the VG metadata. If you set a value larger
    # than the the total number of metadata areas available then
    # metadata is stored in them all.
    # The default value of 0 ("unmanaged") disables this automatic
    # management and allows you to control which metadata areas
    # are used at the individual PV level using 'pvchange
    # --metadataignore y/n'.
    # vgmetadatacopies = 0
    # Approximate default size of on-disk metadata areas in sectors.
    # You should increase this if you have large volume groups or
    # you want to retain a large on-disk history of your metadata changes.
    # pvmetadatasize = 255
    # List of directories holding live copies of text format metadata.
    # These directories must not be on logical volumes!
    # It's possible to use LVM2 with a couple of directories here,
    # preferably on different (non-LV) filesystems, and with no other
    # on-disk metadata (pvmetadatacopies = 0). Or this can be in
    # addition to on-disk metadata areas.
    # The feature was originally added to simplify testing and is not
    # supported under low memory situations - the machine could lock up.
    # Never edit any files in these directories by hand unless you
    # you are absolutely sure you know what you are doing! Use
    # the supplied toolset to make changes (e.g. vgcfgrestore).
    # dirs = [ "/etc/lvm/metadata", "/mnt/disk2/lvm/metadata2" ]
    # Event daemon
    dmeventd {
    # mirror_library is the library used when monitoring a mirror device.
    # "libdevmapper-event-lvm2mirror.so" attempts to recover from
    # failures. It removes failed devices from a volume group and
    # reconfigures a mirror as necessary. If no mirror library is
    # provided, mirrors are not monitored through dmeventd.
    mirror_library = "libdevmapper-event-lvm2mirror.so"
    # snapshot_library is the library used when monitoring a snapshot device.
    # "libdevmapper-event-lvm2snapshot.so" monitors the filling of
    # snapshots and emits a warning through syslog when the use of
    # the snapshot exceeds 80%. The warning is repeated when 85%, 90% and
    # 95% of the snapshot is filled.
    snapshot_library = "libdevmapper-event-lvm2snapshot.so"
    # thin_library is the library used when monitoring a thin device.
    # "libdevmapper-event-lvm2thin.so" monitors the filling of
    # pool and emits a warning through syslog when the use of
    # the pool exceeds 80%. The warning is repeated when 85%, 90% and
    # 95% of the pool is filled.
    thin_library = "libdevmapper-event-lvm2thin.so"
    # Full path of the dmeventd binary.
    # executable = "/usr/sbin/dmeventd"

  • Excel 2013 crash - related to digital signature?

    This is a cross-posting of a thread originally started in Microsoft Answers (Office- Excel) forum, advised to post in this forum by MS support engineer Mohan Suryanarayan (link to other forum: here).
    I have a VB macro in file A (which otherwise contains only several Excel Tables with named ranges). My digital signature (for signing code) is attached to file A.
    As the macro executes, the Open file dialog is called up so that the User can select a secondary file to open (file B or C).
    Files B and C contain several sheets with formulas and Excel Tables, and also their own extensive VB macros. My digital signature is also attached to files B & C.
    The remainder of the macro in file A simply executes an update of the Excel Tables in the secondary file, before saving & closing it.
    I have been finding that during execution of the file A macro, Excel crashes and closes all open Excel files.
    If I remove the digital signature from the secondary files then the macro in file A executes with no problem.
    The macro was written in Excel 2007 and on a different laptop, with different digital signature, and used to have no problem at all. When a new laptop was purchased a new digital signature was obtained and placed into all of the files.
    Execution of file A is a critical process - I need help with stopping the crash, while maintaining a digital signature in the secondary files (so that their own macros can execute).
    Event Viewer error shows the following:
    Faulting application name: EXCEL.EXE, version: 15.0.4667.1000, time stamp: 0x543d366c
    Faulting module name: mso.dll, version: 0.0.0.0, time stamp: 0x5447696f
    Exception code: 0xc0000602
    Fault offset: 0x011aafdd
    Faulting process id: 0x2d58
    Faulting application start time: 0x01d01e17eafcd0bb
    Faulting application path: C:\Program Files\Microsoft Office 15\root\office15\EXCEL.EXE
    Faulting module path: C:\Program Files (x86)\Common Files\Microsoft Shared\Office15\mso.dll
    JL Latham replied:
    Tom, all I can give you is relatively non-information.  I haven't experienced that type of behavior even in Excel 2013 with digitally signed macros in multiple workbooks.  But my digital signature is a commercial one from DigiCert, not
    self-cert.  I'm not even sure what piece of Excel or Office processes the digital signature (that is, whether it is mso.dll or not).  But your whole process has more moving parts than most of mine involving multiple workbooks do - I quite often turn
    off .EnableEvents when doing things like that just so event driven macros in the other workbook(s) do not execute.
    Do the macros in the 'B files' have to execute during all of this processing in file A?  If not, have you tried turning interrupts off:
    Application.EnableEvents = False
    before starting to work with the B-files?  Maybe something in them, such as a Worksheet_Change() or other event triggered process is messing things up?
    If you do use  Application.EnableEvents=False, remember that you MUST re-enable that later in the code with Application.EnableEvents=True to get responses to things like button clicks, changes in a worksheet or other event driven processes
    to run 'normally' again.
    Have you looked at this discussion:
    http://answers.microsoft.com/en-us/office/forum/office_2013_release-excel/excel-2013-crashing-immediately-when-attempting-to/ea2ab100-5525-4a0d-a3cf-6977319f954f
    Sounds very similar to yours - original signature created in 2007, and crashing in 2013.  Some suggestions were made, but whether they fixed the issue or not is unknown: OP never came back and said "fixed it" or "didn't fix it".
    I replied:
    In answer, the secondary B files don't execute any part of their macros during the A file process, and I hadn't included an EnableEvents = False within the file A code. However, I have just tried it to check if it made any difference and unfortunately it didn't.
    I had read the thread you gave a link to before posting, and also the sub threads given within that one, but didn't see anything I thought of immediate parallel other than an Excel 2013 crash.
    The way I read the linked thread was that a User added a signature in Excel 2007 and when they tried to open the file Excel (2007?) crashes, although other people using Excel 2013 can open it (not clear whether it means others can open the one with
    the 2007 signature or without it).
    Anyway, the differences in my situation are that:
    (1) my B-files were originally created in Excel 2007 on a different laptop to current and had their own signature attached. I maintain a 'stock' of B-file versions, some of which have the old signature and some the new. If I run the A-file macro
    on one of the B-files with the old signature then it executes ok;
    (2) I am able to open the B-files ok, no matter whether the signature in them derives from the old laptop with Excel 2007 or the new laptop with Excel 2013;
    (3) the macros within the B-files execute ok (this includes any versions of the B-files which still have an old signature attached, by removing the old signature and replacing with the new and executing on the new laptop);
    (4) the offending operation which triggers the crash (when the B-files have the new signature attached) is a Resize of the Excel Tables within the B-file. I understand that I may eventually have to find a workaround for this operation but initially
    I want to sort out the signature issue (I'm reasonably sure the signature issue is not a red herring where the Resize is the 'real' problem).
    Regards, Tom

    Hi,
    According to your description, this issue occurred with some particular files that stored in SharePoint site. Did the user  download the file to local disk to print?
    Based on the event log >>Faulting module name: EXCEL.EXE<<, it does not show the root Faulting module. Thus, please try to follow this KB to do general troubleshooting:
    https://support.microsoft.com/en-us/kb/2758592
    Then, if you have used "printer status application", please go to Device settings -> Installable options -> Printer Status notification and select disable.
    If this issue still exists, we may need to collect the App crash dump file:
    https://technet.microsoft.com/en-us/sysinternals/dd996900.aspx
    You can try to analyze dump by yourself if you would like to:
    How to analyze app crash dump file:
    http://blogs.technet.com/b/askperf/archive/2007/05/29/basic-debugging-of-an-application-crash.aspx
    Regards,
    George Zhao
    TechNet Community Support
    It's recommended to download and install
    Configuration Analyzer Tool (OffCAT), which is developed by Microsoft Support teams. Once the tool is installed, you can run it at any time to scan for hundreds of known issues in Office
    programs.
    Please remember to mark the replies as answers if they help, and unmark the answers if they provide no help. If you have feedback for TechNet Support, contact [email protected]

  • Is there any version of Indesign that allows to open multiple instance at the same time

    I am using Indesign CS3 to generate the bulk reports automatically(programmatically) using Adobe indesign library . One of the key limitation to this tool is it allows to open only one instance at a time unlike word where multiple instances can be opened at the same time. Is there any adobe indesign too version desktop/server which allows to open multiple instances at the same time. Any help on is appreciated. Thanks.

    @PeterG
    You said "You might want to examine the End User Licensing Agreement (EULA) for  any prohibition against using these two instances at the same time with  the same license; the license does permit two installs for the same user  that are not used simultaneously." I thought that meant you had read the EULA. It's pretty unambiguous on that point.
    For everyone's edification about multiple installations and server use with a standard license here's the relevant text:
    2. Software License.
    If you obtained the Software and any required serial number(s) from Adobe or one of its authorized
    licensees and as long as you comply with the terms of this agreement, Adobe grants you a non-exclusive
    license to install and use the Software in a manner consistent with its design and Documentation and as
    further set forth below. See Section 16 for specific provisions related to the use of certain products and
    components, for example font software, Acrobat, After Effects, Adobe Presenter, Contribute, Adobe
    Device Central, Flash Player, Flash Builder, and Adobe Runtimes.
    2.1 Limited Use. The Software, or portions of the Software, may allow installation and use without a serial
    number. If so, you may install but not use such non-serialized Software on any number of Computers as
    part of an organizational deployment plan. Further, you may install and use such non-serialized Software
    on any number of Computers for demonstration, evaluation and training purposes only and only if any
    Output Files or other materials produced through such use are used only for internal, non-commercial and
    non-production purposes. You may not use such non-serialized Software after any applicable time-out
    period has ended, unless you input a valid serial number under Section 2.2. ACCESS TO AND USE OF
    ANY OUTPUT FILES CREATED WITH SUCH NON-SERIALIZED SOFTWARE IS ENTIRELY AT
    YOUR OWN RISK.
    2.2 General Use. You may install and use one copy of the Software only on the Permitted Number of your
    compatible Computers into which you enter a valid serial number.
    2.3 Distribution from Server. You may copy an image of the Software onto Computer file server(s) within
    your Internal Network for the purpose of downloading and installing the Software onto Computers within
    the same Internal Network for use as permitted by Section 2.1 and 2.2.
    2.4 Server Use. You may install the Software on Computer file server(s) within your Internal Network only
    for use of the Software initiated by an individual from a Computer within the same Internal Network only
    as permitted by Section 2.2. The total number of users (not the concurrent number of users) able to use the
    Software on such Computer file server(s) may not exceed the Permitted Number.
    By way of example, the foregoing does not permit you to install or access (either directly or through
    commands, data or instructions) the Software: (a) from or to a Computer not part of your Internal Network,
    (b) for enabling Web hosted workgroups or services available to the public, (c) by any individual or entity
    to use, download, copy or otherwise benefit from the functionality of the Software unless licensed to do so
    by Adobe, (d) as a component of a system, workflow or service accessible by more than the Permitted
    Number of users, or (e) for operations not initiated by an individual user (e.g., automated server
    processing).
    2.5 Portable or Home Computer Use. Subject to the important restrictions set forth in Section 2.6, the
    primary user of the Computer on which the Software is installed under Section 2.2 (“Primary User”) may
    install a second copy of the Software for his or her exclusive use on either a portable Computer or a
    Computer located at his or her home, provided that the Software on the portable or home Computer is not
    used at the same time as the Software on the primary Computer.
    2.6 Restrictions on Secondary Use by Volume Licensees. If the Software was obtained under an Adobe
    volume license program (currently known as Adobe Open Options) by any licensee other than an
    educational volume licensee, the second copy of the Software made under Section 2.5 must be used solely
    for the benefit and business of that volume licensee. For more information about secondary use by volume
    licensees, please visit our Web Site at http://www.adobe.com/go/open_options.
    There is also language stating that if you have a dual boot or other system capable of running more than one OS you will need a separate license for each platform if you want to run the software on more than one.

  • The cluster resource 'SQL Server' could not be brought online error while in completion stage of sql server 2012 failover installation

    While installing sql server 2012 with failover clustering I got the following error:
    The cluster resource 'SQL Server' could not be brought online. Error: There was a failure to call cluster code from a provider. Exception message: Generic failure. Status code: 5942. Description: The resource failed to come online due to the failure of one
    or more provider resources.
    The two nodes of the sql server failover have been successfully prepared. The error came while running the advance cluster completion option.
    In the clustered events in failover cluster manager the following message is displayed:
    Cluster resource 'SQL Network Name (CPCTestSQLClus)' of type 'Network name' in clustered role 'SQL Server (MSSQLSERVER)' failed.
    I have done various checking and the issue remained. I tried to start the sql

    Please find below part of the content of the cluster log. I hope this is helpful in resolving the issue:
    00000998.000010e8::2013/05/08-20:45:00.233 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Online called for resource SQL Network Name (CPCTestSQLClus)
    000005a4.0000050c::2013/05/08-20:45:00.233 INFO  [RCM] HandleMonitorReply: ONLINERESOURCE for 'SQL Network Name (CPCTestSQLClus)', gen(6) result 997/0.
    000005a4.0000050c::2013/05/08-20:45:00.233 INFO  [RCM] Res SQL Network Name (CPCTestSQLClus): OnlineCallIssued -> OnlinePending( StateUnknown )
    000005a4.0000050c::2013/05/08-20:45:00.233 INFO  [RCM] TransitionToState(SQL Network Name (CPCTestSQLClus)) OnlineCallIssued-->OnlinePending.
    00000998.000012f0::2013/05/08-20:45:00.233 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Entering Online thread
    000005a4.0000050c::2013/05/08-20:45:00.233 INFO  [GEM] Sending 1 messages as a batched GEM message
    00000998.000012f0::2013/05/08-20:45:00.233 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Initializing config info
    00000998.000012f0::2013/05/08-20:45:00.248 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Skipping initialization of 'Identity' module because netname is not yet created
    00000998.000012f0::2013/05/08-20:45:00.248 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Performing actual online of resource.
    00000998.000012f0::2013/05/08-20:45:00.248 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Obtaining IP info for resource SQL IP Address 1 (CPCTestSQLClus)
    00000998.000012f0::2013/05/08-20:45:00.248 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Using provider SQL IP Address 1 (CPCTestSQLClus), ip address 10.144.166.160, mask 255.255.255.0, prefix length 24
    00000998.000012f0::2013/05/08-20:45:00.248 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Resource has 1 IPs
    00000998.000012f0::2013/05/08-20:45:00.248 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: IP: Type Ipv4, Address 10.144.166.160:~0~, Prefix 10.144.166.160/24, Online true, Transport \Device\NetBt_If3
    00000998.000012f0::2013/05/08-20:45:00.248 INFO  [RES] Network Name: Agent: Initializing (082d8ef5-62a8-4fee-acbc-e89b769e9181,Configuration)
    00000998.000012f0::2013/05/08-20:45:00.248 INFO  [RES] Network Name: Agent: Sending request Netname/InitializeIndirect to NN:Agent
    00000998.000010e8::2013/05/08-20:45:00.248 INFO  [RES] Network Name: Agent: OnInitialize (082d8ef5-62a8-4fee-acbc-e89b769e9181,Configuration)
    00000998.000010e8::2013/05/08-20:45:00.248 INFO  [RES] Network Name: Agent: CanModuleBeInitializedImp - Module can be initialized, current state Offline
    00000998.000010e8::2013/05/08-20:45:00.248 INFO  [RES] Network Name: Agent: Sending Initialize to module NN:082d8ef5-62a8-4fee-acbc-e89b769e9181:Configuration
    00000998.000010e8::2013/05/08-20:45:00.248 INFO  [RES] Network Name: Agent: Sending request Netname/Initialize to NN:082d8ef5-62a8-4fee-acbc-e89b769e9181:Configuration
    00000998.000010e8::2013/05/08-20:45:00.248 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Configuration: Waiting for ongoing slow strand to complete (if any)
    00000998.000010e8::2013/05/08-20:45:00.248 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Configuration: Signaling main strand to initialize module
    00000998.000010e8::2013/05/08-20:45:00.248 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Configuration: Calling initialize of the configuration implementation
    000005a4.00001350::2013/05/08-20:45:00.248 INFO  [GEM] Sending 1 messages as a batched GEM message
    00000998.000010e8::2013/05/08-20:45:00.248 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Configuration: Setting 'StatusKerberos' in clusdb returned status 0
    00000998.000010e8::2013/05/08-20:45:00.248 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Configuration: Creating a New AD NetName (type Singleton)...
    00000998.000010e8::2013/05/08-20:45:00.248 INFO  [RES] Network Name: Agent: Initializing (082d8ef5-62a8-4fee-acbc-e89b769e9181,AccountAD)
    00000998.000010e8::2013/05/08-20:45:00.248 INFO  [RES] Network Name: Agent: Sending request Netname/InitializeIndirect to NN:Agent
    00000998.000010e8::2013/05/08-20:45:00.248 INFO  [RES] Network Name: Agent: OnInitialize (082d8ef5-62a8-4fee-acbc-e89b769e9181,AccountAD)
    00000998.000010e8::2013/05/08-20:45:00.248 INFO  [RES] Network Name: Agent: CanModuleBeInitializedImp - Module can be initialized, current state Closing
    00000998.000010e8::2013/05/08-20:45:00.248 INFO  [RES] Network Name: Agent: Sending Initialize to module NN:082d8ef5-62a8-4fee-acbc-e89b769e9181:AccountAD
    00000998.000010e8::2013/05/08-20:45:00.248 INFO  [RES] Network Name: Agent: Sending request Netname/Initialize to NN:082d8ef5-62a8-4fee-acbc-e89b769e9181:AccountAD
    00000998.000010e8::2013/05/08-20:45:00.248 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: AccountAD: Waiting for ongoing slow strand to complete (if any)
    00000998.000010e8::2013/05/08-20:45:00.248 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: AccountAD: Signaling main strand to initialize module
    00000998.000010e8::2013/05/08-20:45:00.248 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: AccountAD: Initializing Name: CPCTestSQLClus, NetbiosName: CPCTESTSQLCLUS, Type: Singleton, Created: false
    00000998.000010e8::2013/05/08-20:45:00.248 INFO  [RES] Network Name: [NNLIB] GetCreatingDC - name CPCTestSQLClus, lookup flags = 1
    00000998.000010e8::2013/05/08-20:45:00.452 INFO  [RES] Network Name: [NNLIB] GetDCWithFlags - using DC
    \\MY01DC03.Domain.net, domain name Domain.net, IsReadOnly 0, Object exists 0
    00000998.000010e8::2013/05/08-20:45:00.452 INFO  [RES] Network Name:  [NN] Setting crypto access members for encrypt. New container = false.
    000005a4.00001350::2013/05/08-20:45:00.452 INFO  [GEM] Sending 1 messages as a batched GEM message
    00000998.000010e8::2013/05/08-20:45:00.452 INFO  [RES] Network Name: Agent: Sending request Netname/LastDcChange to NN:082d8ef5-62a8-4fee-acbc-e89b769e9181:Configuration
    00000998.0000075c::2013/05/08-20:45:00.452 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Configuration: LastDc Changed, configOnly: 1
    000005a4.00001350::2013/05/08-20:45:00.452 INFO  [GEM] Sending 1 messages as a batched GEM message
    00000998.0000075c::2013/05/08-20:45:00.468 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Configuration: Setting last DC in clusdb returned status 0
    00000998.0000075c::2013/05/08-20:45:00.468 INFO  [RES] Network Name:  [NN] got sync reply: 0
    00000998.000010e8::2013/05/08-20:45:00.468 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: AccountAD: LastDc Changed (configOnly true), result: 0
    00000998.000010e8::2013/05/08-20:45:00.468 INFO  [RES] Network Name: Agent: Sending request Netname/PasswordChange to NN:082d8ef5-62a8-4fee-acbc-e89b769e9181:Configuration
    00000998.0000075c::2013/05/08-20:45:00.468 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Configuration: Password Changed, configOnly: 1
    000005a4.00001350::2013/05/08-20:45:00.468 INFO  [GEM] Sending 1 messages as a batched GEM message
    00000998.0000075c::2013/05/08-20:45:00.468 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Configuration: Setting password in clusdb returned status 0
    00000998.0000075c::2013/05/08-20:45:00.468 INFO  [RES] Network Name:  [NN] got sync reply: 0
    00000998.000010e8::2013/05/08-20:45:00.468 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: AccountAD: PasswordChanged (configOnly true), result: 0
    00000998.000010e8::2013/05/08-20:45:00.468 INFO  [RES] Network Name:  [NN] Setting crypto access members for decrypt. New container = false.
    00000998.00000ab8::2013/05/08-20:45:00.636 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Getting Read/Write private properties
    00000998.00000ab8::2013/05/08-20:45:00.647 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Getting Read only private properties
    00000998.0000075c::2013/05/08-20:45:00.784 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Getting Read/Write private properties
    00000998.00000ab8::2013/05/08-20:45:00.815 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Getting Read only private properties
    00000998.00000ab8::2013/05/08-20:45:00.815 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Getting Read/Write private properties
    00000998.00000ab8::2013/05/08-20:45:00.830 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Getting Read only private properties
    00000998.00000ab8::2013/05/08-20:45:00.830 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Getting Read/Write private properties
    00000998.00000ab8::2013/05/08-20:45:00.830 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Getting Read only private properties
    00000998.000010e8::2013/05/08-20:45:00.924 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: AccountAD: OU name for VCO is OU=NewComputers,OU=WORKSTATIONS,DC=Domain,DC=net
    00000998.000010e8::2013/05/08-20:45:00.924 INFO  [RES] Network Name: Agent: Sending request Netname/GetToken to NN:c03fb105-9a50-4d00-8bee-afaf693f7efc:Identity
    00000998.00000ab8::2013/05/08-20:45:00.924 INFO  [RES] Network Name <Cluster Name>: Identity: Module is not yet initialized. Trying to obtain a new token.
    00000998.0000075c::2013/05/08-20:45:00.924 INFO  [RES] Network Name <Cluster Name>: Identity: Obtaining new token
    00000998.0000075c::2013/05/08-20:45:00.924 INFO  [RES] Network Name:  [NN] Setting crypto access members for decrypt. New container = false.
    00000998.0000075c::2013/05/08-20:45:00.924 INFO  [RES] Network Name: [NNLIB] Priming local KDC cache to
    \\MY01DC02.Domain.net for domain Domain.net
    00000998.0000075c::2013/05/08-20:45:00.924 INFO  [RES] Network Name: [NNLIB] PopulateKerbKDCLookupCache - DC flags 0
    00000998.0000075c::2013/05/08-20:45:00.940 INFO  [RES] Network Name: [NNLIB] LsaCallAuthenticationPackage success with a request of size 94, result size 0 (status: 0, subStatus: 0)
    00000998.0000075c::2013/05/08-20:45:00.940 INFO  [RES] Network Name: [NNLIB] Priming local KDC cache to
    \\MY01DC02.Domain.net for domain label Domain
    00000998.0000075c::2013/05/08-20:45:00.940 INFO  [RES] Network Name: [NNLIB] LsaCallAuthenticationPackage success with a request of size 86, result size 0 (status: 0, subStatus: 0)
    00000998.0000075c::2013/05/08-20:45:01.018 WARN  [RES] Network Name: [NNLIB] LogonUserEx fails for user CPCTestWinClus$: 1326 (useSecondaryPassword: 0)
    00000998.0000075c::2013/05/08-20:45:01.096 WARN  [RES] Network Name: [NNLIB] LogonUserEx fails for user CPCTestWinClus$: 1326 (useSecondaryPassword: 1)
    00000998.0000075c::2013/05/08-20:45:01.096 INFO  [RES] Network Name: [NNLIB] Logon failed for user CPCTestWinClus$ (Error 1326), DC
    \\MY01DC02.Domain.net, domain Domain.net
    00000998.0000075c::2013/05/08-20:45:01.096 INFO  [RES] Network Name <Cluster Name>: Identity: Obtaining Windows Token for Name: CPCTestWinClus, SamName: CPCTestWinClus$, Type: Singleton, Result: 1326, LastDC:
    \\MY01DC02.Domain.net
    00000998.0000075c::2013/05/08-20:45:01.096 INFO  [RES] Network Name <Cluster Name>: Identity: Slow Operation, FinishWithReply: 1326
    00000998.0000075c::2013/05/08-20:45:01.096 INFO  [RES] Network Name <Cluster Name>: Identity: InternalReplyHandler with event: 1326
    00000998.0000075c::2013/05/08-20:45:01.096 INFO  [RES] Network Name <Cluster Name>: Identity: End of Slow Operation, state: Error/Idle, prevWorkState: Idle
    00000998.00000ab8::2013/05/08-20:45:01.096 WARN  [RES] Network Name <Cluster Name>: Identity: Get Token Request, currently doesnt have a token!
    00000998.00000ab8::2013/05/08-20:45:01.096 INFO  [RES] Network Name:  [NN] got sync reply: 0
    00000998.000010e8::2013/05/08-20:45:01.096 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: AccountAD: End of Slow Operation, state: Initializing/Writing, prevWorkState: Writing
    00000998.000010e8::2013/05/08-20:45:01.096 WARN  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: AccountAD: Slow operation has exception ERROR_INVALID_HANDLE(6)' because of '::ImpersonateLoggedOnUser( GetToken() )'
    00000998.000010e8::2013/05/08-20:45:01.096 INFO  [RES] Network Name: Agent: OnInitializeReply, Failure on (082d8ef5-62a8-4fee-acbc-e89b769e9181,AccountAD): 6
    00000998.000010e8::2013/05/08-20:45:01.096 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Configuration: InitializeReplyCreation of NetName (type Singleton), result: 6, IsCanceled: false
    000005a4.00000cd8::2013/05/08-20:45:01.096 INFO  [GEM] Sending 1 messages as a batched GEM message
    00000998.00000ab8::2013/05/08-20:45:01.096 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Getting Read/Write private properties
    00000998.000010e8::2013/05/08-20:45:01.096 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Configuration: Setting 'StatusKerberos' in clusdb returned status 0
    00000998.000010e8::2013/05/08-20:45:01.096 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Configuration: Deleting ResourceData, CreatingDC, ObjectGUID for a newly created netname from cluster database
    000005a4.00000cd8::2013/05/08-20:45:01.096 INFO  [GEM] Sending 1 messages as a batched GEM message
    000005a4.00000cd8::2013/05/08-20:45:01.112 INFO  [GEM] Sending 1 messages as a batched GEM message
    00000998.00000ab8::2013/05/08-20:45:01.112 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Getting Read only private properties
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: OnInitializeReply, Failure on (082d8ef5-62a8-4fee-acbc-e89b769e9181,Configuration): 6
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: SyncReplyHandler Configuration, result: 6
    00000998.000012f0::2013/05/08-20:45:01.112 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: PerformOnline - Initialization of Configuration module finished with result: 6
    00000998.000012f0::2013/05/08-20:45:01.112 ERR   [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Online thread Failed: ERROR_SUCCESS(0)' because of 'Initializing netname configuration for SQL Network Name (CPCTestSQLClus) failed with
    error 6.'
    00000998.000012f0::2013/05/08-20:45:01.112 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: All resources offline. Cleaning up.
    00000998.000012f0::2013/05/08-20:45:01.112 ERR   [RHS] Online for resource SQL Network Name (CPCTestSQLClus) failed.
    000005a4.0000050c::2013/05/08-20:45:01.112 WARN  [RCM] HandleMonitorReply: ONLINERESOURCE for 'SQL Network Name (CPCTestSQLClus)', gen(6) result 5018/0.
    000005a4.0000050c::2013/05/08-20:45:01.112 INFO  [RCM] Res SQL Network Name (CPCTestSQLClus): OnlinePending -> ProcessingFailure( StateUnknown )
    000005a4.0000050c::2013/05/08-20:45:01.112 INFO  [RCM] TransitionToState(SQL Network Name (CPCTestSQLClus)) OnlinePending-->ProcessingFailure.
    000005a4.0000050c::2013/05/08-20:45:01.112 ERR   [RCM] rcm::RcmResource::HandleFailure: (SQL Network Name (CPCTestSQLClus))
    000005a4.0000050c::2013/05/08-20:45:01.112 INFO  [RCM] resource SQL Network Name (CPCTestSQLClus): failure count: 6, restartAction: 2 persistentState: 1.
    000005a4.00000cd8::2013/05/08-20:45:01.112 INFO  [GEM] Sending 1 messages as a batched GEM message
    000005a4.0000050c::2013/05/08-20:45:01.112 INFO  [RCM] Greater than restartPeriod time has elapsed since first failure of SQL Network Name (CPCTestSQLClus), resetting failureTime and failureCount.
    000005a4.0000050c::2013/05/08-20:45:01.112 INFO  [RCM] Will queue immediate restart (500 milliseconds) of SQL Network Name (CPCTestSQLClus) after terminate is complete.
    000005a4.0000050c::2013/05/08-20:45:01.112 INFO  [RCM] Res SQL Network Name (CPCTestSQLClus): ProcessingFailure -> WaitingToTerminate( DelayRestartingResource )
    000005a4.0000050c::2013/05/08-20:45:01.112 INFO  [RCM] TransitionToState(SQL Network Name (CPCTestSQLClus)) ProcessingFailure-->[WaitingToTerminate to DelayRestartingResource].
    000005a4.0000050c::2013/05/08-20:45:01.112 INFO  [RCM] Res SQL Network Name (CPCTestSQLClus): [WaitingToTerminate to DelayRestartingResource] -> Terminating( DelayRestartingResource )
    000005a4.0000050c::2013/05/08-20:45:01.112 INFO  [RCM] TransitionToState(SQL Network Name (CPCTestSQLClus)) [WaitingToTerminate to DelayRestartingResource]-->[Terminating to DelayRestartingResource].
    00000998.00000ab8::2013/05/08-20:45:01.112 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Terminate called for resource SQL Network Name (CPCTestSQLClus)
    00000998.00000ab8::2013/05/08-20:45:01.112 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Entering Offline thread
    00000998.00000ab8::2013/05/08-20:45:01.112 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Performing actual offline of resource.
    00000998.00000ab8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: Closing (082d8ef5-62a8-4fee-acbc-e89b769e9181,AdminShare)
    00000998.00000ab8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: Sending request Netname/CloseIndirect to NN:Agent
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: OnClose (082d8ef5-62a8-4fee-acbc-e89b769e9181,AdminShare)
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: Sending request Netname/Close to NN:082d8ef5-62a8-4fee-acbc-e89b769e9181:AdminShare
    00000998.000010e8::2013/05/08-20:45:01.112 ERR   [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: AdminShare: OnCloseBase, Error Already Closing, previous state: Closing/Ending
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: OnCloseReply (082d8ef5-62a8-4fee-acbc-e89b769e9181,AdminShare) result: 1904
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: SyncReplyHandler Configuration, result: 1904
    00000998.00000ab8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: Closing (082d8ef5-62a8-4fee-acbc-e89b769e9181,Configuration)
    00000998.00000ab8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: Sending request Netname/CloseIndirect to NN:Agent
    000005a4.00000cd8::2013/05/08-20:45:01.112 INFO  [RCM-rbtr] giving default token to group SQL Server (MSSQLSERVER)
    000005a4.00000cd8::2013/05/08-20:45:01.112 INFO  [RCM-rbtr] giving default token to group SQL Server (MSSQLSERVER)
    000005a4.00000cd8::2013/05/08-20:45:01.112 INFO  [RCM-rbtr] giving default token to group SQL Server (MSSQLSERVER)
    000005a4.00000cd8::2013/05/08-20:45:01.112 INFO  [RCM-rbtr] giving default token to group SQL Server (MSSQLSERVER)
    00000998.0000075c::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: OnClose (082d8ef5-62a8-4fee-acbc-e89b769e9181,Configuration)
    00000998.0000075c::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: Sending request Netname/Close to NN:082d8ef5-62a8-4fee-acbc-e89b769e9181:Configuration
    00000998.0000075c::2013/05/08-20:45:01.112 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Configuration: Canceling work, state: Closing/Idle
    00000998.0000075c::2013/05/08-20:45:01.112 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Configuration: OnCloseBase, previous state: Initializing/Idle
    00000998.0000075c::2013/05/08-20:45:01.112 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Configuration: Closing... (PreviousState: Initializing, Created: false, Type: Singleton)
    00000998.0000075c::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: Closing (082d8ef5-62a8-4fee-acbc-e89b769e9181,AccountAD)
    00000998.0000075c::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: Sending request Netname/CloseIndirect to NN:Agent
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: OnClose (082d8ef5-62a8-4fee-acbc-e89b769e9181,AccountAD)
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: Sending request Netname/Close to NN:082d8ef5-62a8-4fee-acbc-e89b769e9181:AccountAD
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: AccountAD: Canceling work, state: Closing/Idle
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: AccountAD: OnCloseBase, previous state: Initializing/Idle
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: OnCloseReply (082d8ef5-62a8-4fee-acbc-e89b769e9181,AccountAD) result: 0
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Configuration: Finish Closing NetName (type Singleton) Module AccountAD, result 0, remaining... 0 (0)
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: -- listing 3 instances ---------------------------------------------
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: --- 7 modules for instance: 082d8ef5-62a8-4fee-acbc-e89b769e9181:
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: ---- Module: Configuration with states: Offline/Ending (BeingBorn)
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: ---- Module: AccountAD with states: Closing/Ending (BeingBorn)
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: ---- Module: Client with states: Offline/Idle (BeingBorn)
    00000998.0000075c::2013/05/08-20:45:01.112 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Configuration: Closed
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: ---- Module: Identity with states: Offline/Idle (BeingBorn)
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: ---- Module: Dns with states: Offline/Idle (BeingBorn)
    00000998.0000075c::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: OnCloseReply (082d8ef5-62a8-4fee-acbc-e89b769e9181,Configuration) result: 0
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: ---- Module: Netbios with states: Offline/Idle (BeingBorn)
    00000998.0000075c::2013/05/08-20:45:01.112 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: SyncReplyHandler Configuration, result: 0
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: ---- Module: AdminShare with states: Closing/Ending (BeingBorn)
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: --- 7 modules for instance: 43dec41b-661c-42a8-ae23-b635d7ab11f5:
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: ---- Module: Configuration with states: Offline/Ending (BeingBorn)
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: ---- Module: AccountAD with states: Closing/Ending (BeingBorn)
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: ---- Module: Client with states: Offline/Idle (BeingBorn)
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: ---- Module: Identity with states: Offline/Idle (BeingBorn)
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: ---- Module: Dns with states: Offline/Idle (BeingBorn)
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: ---- Module: Netbios with states: Offline/Idle (BeingBorn)
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: ---- Module: AdminShare with states: Closing/Ending (BeingBorn)
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: --- 7 modules for instance: c03fb105-9a50-4d00-8bee-afaf693f7efc:
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: ---- Module: Configuration with states: Initialized/Idle (Alive)
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: ---- Module: AccountAD with states: Initializing/Idle (BeingBorn)
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: ---- Module: Client with states: Initialized/Idle (Alive)
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: ---- Module: Identity with states: Error/Idle (BeingBorn)
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: ---- Module: Dns with states: Initialized/Idle (Alive)
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: ---- Module: Netbios with states: Initialized/Idle (Alive)
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: ---- Module: AdminShare with states: Initialized/Idle (BeingBorn)
    00000998.000010e8::2013/05/08-20:45:01.112 INFO  [RES] Network Name: Agent: -- 0 Zombie modules
    000005a4.0000050c::2013/05/08-20:45:01.112 INFO  [RCM] HandleMonitorReply: TERMINATERESOURCE for 'SQL Network Name (CPCTestSQLClus)', gen(7) result 0/0.
    000005a4.0000050c::2013/05/08-20:45:01.112 INFO  [RCM] Res SQL Network Name (CPCTestSQLClus): [Terminating to DelayRestartingResource] -> DelayRestartingResource( StateUnknown )
    000005a4.0000050c::2013/05/08-20:45:01.112 INFO  [RCM] TransitionToState(SQL Network Name (CPCTestSQLClus)) [Terminating to DelayRestartingResource]-->DelayRestartingResource.
    000005a4.0000050c::2013/05/08-20:45:01.112 WARN  [RCM] Queueing immediate delay restart of resource SQL Network Name (CPCTestSQLClus) in 500 ms.
    000005a4.00001248::2013/05/08-20:45:01.112 INFO  [RCM-rbtr] giving default token to group SQL Server (MSSQLSERVER)
    000005a4.00001248::2013/05/08-20:45:01.112 INFO  [RCM-rbtr] giving default token to group SQL Server (MSSQLSERVER)
    000005a4.00001248::2013/05/08-20:45:01.112 INFO  [RCM-rbtr] giving default token to group SQL Server (MSSQLSERVER)
    000005a4.00001248::2013/05/08-20:45:01.112 INFO  [RCM-rbtr] giving default token to group SQL Server (MSSQLSERVER)
    000005a4.00001350::2013/05/08-20:45:01.112 INFO  [GEM] Sending 1 messages as a batched GEM message
    00000998.00000b78::2013/05/08-20:45:01.112 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Getting Read/Write private properties
    000005a4.00001350::2013/05/08-20:45:01.112 INFO  [GEM] Sending 1 messages as a batched GEM message
    00000998.00000b78::2013/05/08-20:45:01.112 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Getting Read only private properties
    00000998.00000b78::2013/05/08-20:45:01.127 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Getting Read/Write private properties
    000005a4.00001350::2013/05/08-20:45:01.127 INFO  [GEM] Sending 1 messages as a batched GEM message
    00000998.00000b78::2013/05/08-20:45:01.127 INFO  [RES] Network Name <SQL Network Name (CPCTestSQLClus)>: Getting Read only private properties
    000005a4.00001350::2013/05/08-20:45:01.127 INFO  [GEM] Sending 1 messages as a batched GEM message
    000005a4.00001350::2013/05/08-20:45:01.127 INFO  [GEM] Sending 1 messages as a batched GEM message
    000005a4.00001350::2013/05/08-20:45:01.143 INFO  [GEM] Sending 1 messages as a batched GEM message
    00000998.00000b78::2013/05/08-20:45:01.252 INFO  [RES] Network Name: Agent: Sending request Netname/RecheckConfig to NN:c03fb105-9a50-4d00-8bee-afaf693f7efc:Netbios
    00000998.00000ab8::2013/05/08-20:45:01.252 INFO  [RES] Network Name <Cluster Name>: Netbios: Slow Operation, FinishWithReply: 0
    00000998.00000ab8::2013/05/08-20:45:01.252 INFO  [RES] Network Name:  [NN] got sync reply: 0
    00000998.00000ab8::2013/05/08-20:45:01.252 INFO  [RES] Network Name <Cluster Name>: Netbios: End of Slow Operation, state: Initialized/Idle, prevWorkState: Idle

  • Database Adapter

    Hi All,
    My requirement is, based on the input selection from the frontend we have to query the database and then display the records based on the input. Sometimes we may have to display all the records from database. So I am using a DB Adapter in a SOA Composite and using Select operation. we have SQL query written in that Select operation. Initially the service was working fine because the number of records were very few. So it was fetching the records from the table and then we are able to display in the UI as well. Now the problem is we have more number of records in that table (more than 10000). So SOA Composite takes a lot of time to execute. Thats why we couldnt display all these records in our frontend. Can anyone help on this.
    Thanks,
    Maanik.

    Hello Maanik,
    I am not familiar with the DB Adapter but I would not assume that you want to display 10000+ records?
    First you may need to check where the majority of the time is spent, in the database or in the network getting the data to your application. Testing the query can be done through SQL*PLUS where the command 'set timing on'. But with a 10000 rows, I would not immediately expect the database to be the culprit.
    You may also want to consider to use an optimizer hint [url http://download.oracle.com/docs/cd/B10501_01/server.920/a96533/hintsref.htm#4942]FIRST_ROWS in your query. This will speed up the execution of the first few rows in your data set. Should you application display the information in subsets, this hint will be useful as the database returns the rest of the rows while the using is viewing the first rows through the application.
    Thanks,
    EJ

  • Excel 2010, file save as Question

    I save my file name based on a cell "C35". This Cell is a concatenation and I
    have recently noticed that some of the data imported from an xml database
    contains illegal characters and Windows won’t save the file. How can I  substitute
    windows offending characters with say an underscore before save as?
     Dim MyName
     MyName = Range("$C$35").Text
     If MyName = "" Then
     MsgBox ("Verify Save as Filename C35")
     Exit Sub
     End If
    Application.Dialogs(xlDialogSaveAs).Show MyName, 52<o:p></o:p>
    heads up

    Because it can be quite difficult to determine what is a legal character and what is not legal how about just restricting to alpha and numeric like the following.
    Code modified with better syntax for Like operator since initial post.
    Sub KeepAlphaNumeric()
        Dim MyName As String
        Dim i As Long
        Dim strNewName As String
        MyName = Range("$C$35").Text
        For i = 1 To Len(MyName)
            If Mid(MyName, i, 1) Like "[A-Za-z0-9]" Then
                strNewName = strNewName & Mid(MyName, i, 1)
            Else
                strNewName = strNewName & "_"
            End If
        Next i
        MsgBox strNewName
    End Sub
    Regards, OssieMac

  • Made a mess of my clean install

    OKay, I took a copy of my lion install onto dvd.  I clicked reinstall which worked fine, however I wanted a clean install of lion on my machine as I have issues with too much content too little hardrive space.
    So my thinking was clean install and then only place back on there what I need once I have cleaned it and removed anything I dont want....so I took a clone backup and also a timemanchine backup
    Here is when it went wrong.
    I clicked the disk utlity, and erased the contents of the hardrive.  I then stupidly read the how to incorrectly and moved the disk drive to the hardrive, thinking this would then  restore from the disk, but all it has done is made the drive locked and i'm no longer able to restore, erase or boot from the disk or the hardrive.
    Next major mistake, I selected restore from timemachine, so its now overwritten my clone copy with the disk copy.
    I'm in  a right mess now and I dont know how to get out of it, any ideas?

    CornwallAppleMac wrote:
     I created the boot disk after I had downloaded from the app store, I did not upgrade immediately, so placed a copy of the contents of the upgrade and copied this to a dvd
    Ok, I see what you did. You just saved the downloaded Lion Installer, you didn't create a bootable Lion DVD as outlined in this procedure.
    http://eggfreckles.net/notes/burning-a-lion-boot-disc/
    If you have another Mac #2 and good single layer DVD's, go ahead and copy that Lion Installer to the hard drive of Mac #2 and create the bootable Lion DVD as outlined in the link instructions.
    Your going to need this to boot from and install both the Lion Recovery Partition and the Lion OS X Partition again.
    I had one hardrive, with two partitions, one with the timemachine backup and the other partition with clone.
    Ok, the clone partiton got TM-ed, but you still got a TM to restore from, which is what we are going to focus on now and then make another clone later to do the "cleanup" operation you initially wanted to do.
    I booted from the dvd and went into disk utility where I erased the hardrive.
    How did you boot from this DVD? Did you hold c and boot, or option or??
    Was this Lion DVD created as outlined in the Eggfreckles link?
    once this was compelted it rebooted and I went back to the disk utility, however I moved the DVD contents across to the internal hardrive.....I know I should have gone to reinstall on the lion install from the DVD now.
    Yes, but you can boot off the DVD again and erase and install Lion from the DVD like you should have, once you do that just restore from TM.
    Clicked restore and it seems to have restored the copy of the dvd to my hardrive.
    It will restore a copy of the DVD installer, not the actual install of Lion and the Lion Recovery Partition
    But you realize your mistake now.
    I then connected my hardrive with the time machine and clone partitions, it asked me select a backup source, so I clicked on timemachine.  It asked me to click on the drive to restore to, I thought it has selected my internal drive but it has selected my partition with the clone on it and wiped it.
    That's ok, you WISELY made two backups, although on the same drive which wasn't the best choice as if that hard drive died it would take both TM and the clone down.
    So since you have a Lion/Lion Recovery install DVD and a TM to restore from, all you have to do is those in that order and your back to normal.
    Once your back to normal on your boot drive, THEN clone over the Lion Partition to the clone partiton again (actually you should use another drive)
    That should get you back exactly where you were before you started.
    So now I have a timemachine folder with backups on, which I have just copied onto another drive just in case I mess up again using another mac I have. 
    However when i go back to my Imac, it says select the boot up disk and takes me to the lion restore options, if I click restore lion it takes me to what was my internal drive but says it is locked.  this disk is locked.  which is now called MAC OS X install ESD.
    I have a cd with my lion on it, and a hardrive.  If I click boot to dvd it says cannot find files, if I boot to hardrive it takes me to the lion install screen but wont go any further.
    Ok, what your doing is moving/copying stuff around and running things without knowing what is what and what they do.
    You likely need to STOP and educate yourself first.
    You have a Lion boot DVD and you have a TM restore, that will get you back to before all this started, but the way your going your going to delete something or mess up TM.
    If you can't c boot off the Lion DVD and erase and format the drive, install Lion and reboot, connect the TM drive and restore from that, then you need to stop and take it to someone who can.
    Your working with your last backup, you mess up now and all your files will be gone.

Maybe you are looking for

  • When opening PDF files in CS6 they show inverted as default?

    Yes (Command I) is a quick fix to this issue but wondering why the default is set as invert?

  • How to hide elements to printer inside muse? / Como esconder elementos a la impresora en muse

    English Hello guys, there is any option to hide elements on muse site for no printing, I'm working on a catalog with pront option but i dont want the navigation menu buttons on my print sheets, any idea? Spanish Hola a todos, hay alguna opcion que bl

  • Automatically apply title to file name?

    I've got a number of files / radio shows (m4a format) for which I have set the title manually. However, their file names are still of the form Lilipuz Kinderhoerspiel-ep09.m4a or similar. Does anybody know of an automatic way to make the title appear

  • APs not registering with WLC

    Hi guys, i'll try to explain our trouble as best as i can: Im trying to do some basic scenario here. Picture this: One L3 Switch connecting a 4402 Controller, 3 Administrative Vlans for APs (63, 93 and 127) and 3 Vlans for clients (one for each Admin

  • Convert to MP3 for Samsung A-900 phone

    I'm trying to put some tunes on my Samsung SPH-A900 phone. It won't accept them unless they're in MP3 format. I've used the ADVANCED>CONVERT SELECTION TO MP3, but when I Get Info on the converted file, it shows that it's an MPEG Audio File. How can I