CCMSR Agent -- Executable not available

Hi Experts,
I`m trying to install ccmsr agent in visual administrator > monitoring-> ccmsagent , but the following message appears:
*Dialog Instance : cmap2_CMP_00*
*Instance Profile Information   /usr/sap/CMP/SYS/profile/CMP_D00_cmap2 is valid*
*Agent Configuration File Information*
*/usr/sap/ccms/CMP_00/sapccmsr/csmconf is valid*
*Agent Executable File Information*
*not available*
*Dialog Instance : cmap1_CMP_00*
*Instance Profile Information*
*/usr/sap/CMP/SYS/profile/CMP_D00_cmap1 is valid*
*Agent Configuration File Information*
*/usr/sap/ccms/CMP_00/sapccmsr/csmconf is valid*
*Agent Executable File Information*
*not available*
I download the ccmsr SAR file and descompress but I dont know why I can solve this.
On the oher hand I don`t know why the central instance is not showing in the message, I want to register it too.
Regards
Edited by: jorge velasquez on Jul 13, 2009 3:41 PM
Edited by: jorge velasquez on Jul 13, 2009 3:41 PM

Hi,
Please check whether the below sequence of activities were followed -
1] Generating the CSMCONF file on the Solution Manager System
2] Unpacking the CCMAGENT file
3] Installing the SAPCCMSR Agent (making sure the owner of the file is the <sid>adm account for the current
system.)
4] Register the agent using visual administrator - making sure the correct csmconf file is selected; and the local/CEN user id password are correct.
Then check the status of the registration using,
sapccmsr -status -j2ee pf=/usr/sap/**/SYS/profile/**
Can also try registering at OS level using,
sapccmsr -R [-f <config file>] [[-j2ee] pf=<complete path of profile of managed instance>]
Hope this helps!
Regards,
Aditya

Similar Messages

  • Agent updates not available after restoring DPM database

    Hi,
    I am using DPM 2012 R2 which I have recently successfully updated to version 4.2.1297.0 (A hotfix installed to resolve an issue with Rollup 5).
    I updated 3 agents successfully and planned to do the remaining 8 at a later date.  However after updating the first 3 I was trying to resolve a reporting issue which I did successfully and I carried out a DPM database restore in the process followed
    by a DPMsync -sync command.
    I am now planning to update the remaining 8 agents but under the management tab the agent status is now showing as OK instead of Update Available as it did previously before the restore.  I have checked the protected clients and they do not have the 4.2.1297.0
    version installed (I checked this in case the installation had occurred).  I then decided I would try and re-install the hotfix on the DPM server (kb3040326) but it seems I can't first uninstall the update on it's own so will have to uninstall DPM altogether.
    I have tried to reinstall the hotfix again on the DPM server without removing it first and I get a message advising:
    "A required hotfix KB3021791 is missing.  Install it from http://support.microsoft.com/kb/3021791"
    This update refers to Update Rollup 5 for DPM 2012 R2 which, again, is already installed on the DPM server and appears under Add/Remove programs.
    It seems that since the restore the DPM server is aware that the Rollup and subsequent hotfixes are installed but the DPM application is not (although when I click the About DPM button, the version is 4.2.1297.0).  I am unable to update the agents through
    the DPM GUI with the latest hotfix as it is not available to them and I am reluctant to install the latest version directly onto the clients without sorting this problem out first.
    If anyone has any advice that would be great,
    Thanks.

    This isn't a bug in Leopard or it's recent updates so it's not something you should expect to see fixed in a subsequent release. However, you're not the only one who has found there to be problems after an update and assumed the update was responsible when it's either coincidental, or simply highlighted a problem that existed on the system previously.
    First thing to try when any device or ports appear to vanish on the system is to perform an SMC reset:
    -From the Apple menu, choose Shut Down (or if the computer is not responding, hold the power button until it turns off).
    -Unplug all cables from the computer, including the power cord and any display cables.
    -Wait at least 15 seconds.
    -Plug the power cord back in, making sure the power button is not being pressed at the time.
    -Then reconnect your keyboard and mouse to the computer.
    -Press the power button on the back to start up your computer.
    Since the SMC (System Management Controller) looks after all the internal devices, when it goes wrong it can cause very bizarre apparent failures! Resetting usually resolves these. Give that a try and post back to let us know.

  • Cacao agent is not available

    I have installed dsee6.1 binary with agent running on default port 11162. I also installed dscc6.1 on the same server (host).
    When I was trying to create a ds instance through dscc, it occured a problem as following:
    Could not contact the DSCC agent on *****. Use the command cacaoadm to check that the DSCC agent is installed and running on port 11162.
    I am very surprise to see this error message coming out, since I have double check that the agent is running and port is 11162. However, i can't create the ds through the dscc.
    The following is the params output from cacaoadm:
    cacaoadm list-params
    snmp-adaptor-port=11161
    snmp-adaptor-trap-port=11162
    jmxmp-connector-port=11162
    commandstream-adaptor-port=11163
    rmi-registry-port=11164
    secure-webserver-port=11165
    java-flags=-Xms4M -Xmx64M
    java-home=/iplanet/servers/v6/jre
    jdmk-home=/iplanet/servers/v6/dsee6/private/lib/..
    nss-lib-home=/iplanet/servers/v6/dsee6/private/lib
    nss-tools-home=/iplanet/servers/v6/dsee6/bin
    retries=4
    enable-instrumentation=false
    network-bind-address=0.0.0.0
    And, the following is message when I check the cacao agent:
    /dscc6/bin>./dsccsetup status
    Sun Java (TM) Web Console is not installed
    DSCC Agent is registered in Cacao
    DSCC Registry has been created
    Path of DSCC registry is /iplanet/servers/v6/var/dscc6/dcc/ads
    Port of DSCC registry is 3998
    Any help will be appriciated!
    Thanks!

    So the question is why the dscc can't find the running cacao instance using default port?
    After I read the installation guide again, I have something unsure in my mind:
    I have 5.2 running on this server, which I did not shutdown when installing 6.1. Do I have to shut them down while installing? Can two different version of DS/DPS co-exist in one box?
    As I knew, older version than 6.x can co-host, but not sure about the new version.
    I have successfully tested v6.1 in my sandbox, then I moved to the dev region. In sandbox server, I don't have any other version running, so I can't tell if the v5.2 would be the factor.
    Please shed anylight.
    Thanks!

  • Personalise value icon not available when query executed from portal

    Hi
    The personalise value icon is not available when the query is executed from the portal.
    Is this the norm or is there anything i must change?
    Regards
    Shalini

    Since you have not given us much information to go on, I will start with the basics...
    Have you activated the Personalization DSOs (0PERS_WTE, 0PERS_BOD & 0PERS_VAR)? 
    If not, you will need to do so in tcode SPRO (SAP Reference IMG --> SAP NetWeaver --> Business Intelligence --> Settings for Reporting and Analysis --> General Settings for Reporting and Analysis --> Activate Personalization in BEx)
    Are you running a BW 3.5 query or a BI 7.0 query?
    If BI 7.0, the way to set personalization values has changed. You have to open the personalization section of the variables refresh screen by clicking on "Show Personalization Values" at the top of the screen, then move the desired variables down to that section.
    Hope some of this helps...
    Bob

  • Oracle 11g Installation DBCA ORA-01034: Oracle Not Available

    I have two oracle databases installed (using the default settings in dbca) and running on a Windows 2003 R2 server, already.
    However, when I try and create a third using the defaults, after 2%, dbca receives an ORA-01034: Oracle Not Available and stops the creation process.
    I've searched 16 hours or so for a similar issue, but haven't had much luck other than to check sqlnet.ora for OS authentication, the registry keys, and tnsnames.ora.
    I've looked at the trace log for dbca a few times for the different attempts at creating a new database, and they tend to differ when the error is generated.
    Looking at Event Viewer, I noticed that connection events occurred during installation, and that for at least one case, CONNECT / AS SYSDBA occured 3 times before the error was generated.
    I've made more than 3 databases on another machine (XP - not server 2003), without any issue, but it seems odd that this error would be because of server 2003.
    I'm fairly inexperienced with configuring databases for Oracle, so please excuse my ignorance.
    Here's a log file to accompany the issue:
    [main] [14:59:16:460] [DBCAWizard.removePageFromList:1318] DBCAWizard->removePageFromList: The page to be removed = NetworkConfPage
    [main] [14:59:16:460] [DBCAWizard.removePageFromList:1318] DBCAWizard->removePageFromList: The page to be removed = NetworkConfPage
    [main] [14:59:16:460] [DBCAWizard.removePageFromList:1318] DBCAWizard->removePageFromList: The page to be removed = NetworkConfPage
    [main] [14:59:16:460] [DBCAWizard.removePageFromList:1318] DBCAWizard->removePageFromList: The page to be removed = NetworkConfPage
    [main] [14:59:16:460] [DBCAWizard.removePageFromList:1318] DBCAWizard->removePageFromList: The page to be removed = NetworkConfPage
    [main] [14:59:16:460] [DBCAWizard.removePageFromList:1318] DBCAWizard->removePageFromList: The page to be removed = NetworkConfPage
    [main] [14:59:16:460] [DBCAWizard.removePageFromList:1318] DBCAWizard->removePageFromList: The page to be removed = NetworkConfPage
    [main] [14:59:16:460] [CommandLineArguments.process:639] CommandLineArguments->process: number of arguments = 0
    [main] [14:59:16:976] [Host.checkOPS:2290] Inside checkOPS
    [main] [14:59:16:976] [Host.checkOPS:2302] Trying to check cluster existence
    [main] [14:59:17:38] [Library.getInstance:106] Created instance of Library.
    [main] [14:59:17:38] [Library.load:206] Loading orauts.dll...
    [main] [14:59:17:38] [Library.load:212] oracleHome D:\app\Administrator\product\11.1.0\db_1
    [main] [14:59:17:38] [Library.load:227] Property oracle.installer.library_loc is set to value=D:\app\Administrator\product\11.1.0\db_1\oui\lib\win32
    [main] [14:59:17:54] [Library.load:229] Loading library D:\app\Administrator\product\11.1.0\db_1\oui\lib\win32\orauts.dll
    [main] [14:59:17:54] [Library.load:262] Loaded library D:\app\Administrator\product\11.1.0\db_1\oui\lib\win32\orauts.dll from path=
    D:\app\Administrator\product\11.1.0\db_1\oui\lib\win32
    [main] [14:59:17:54] [Library.load:206] Loading MSVCR71.dll...
    [main] [14:59:17:54] [Library.load:212] oracleHome D:\app\Administrator\product\11.1.0\db_1
    [main] [14:59:17:54] [Library.load:227] Property oracle.installer.library_loc is set to value=D:\app\Administrator\product\11.1.0\db_1\oui\lib\win32
    [main] [14:59:17:54] [Library.load:229] Loading library D:\app\Administrator\product\11.1.0\db_1\oui\lib\win32\MSVCR71.dll
    [main] [14:59:17:54] [Library.load:262] Loaded library D:\app\Administrator\product\11.1.0\db_1\oui\lib\win32\MSVCR71.dll from path=
    D:\app\Administrator\product\11.1.0\db_1\oui\lib\win32
    [main] [14:59:17:54] [Library.load:206] Loading orawsec11.dll...
    [main] [14:59:17:54] [Library.load:212] oracleHome D:\app\Administrator\product\11.1.0\db_1
    [main] [14:59:17:54] [Library.load:227] Property oracle.installer.library_loc is set to value=D:\app\Administrator\product\11.1.0\db_1\oui\lib\win32
    [main] [14:59:17:54] [Library.load:229] Loading library D:\app\Administrator\product\11.1.0\db_1\oui\lib\win32\orawsec11.dll
    [main] [14:59:17:54] [Library.load:262] Loaded library D:\app\Administrator\product\11.1.0\db_1\oui\lib\win32\orawsec11.dll from path=
    D:\app\Administrator\product\11.1.0\db_1\oui\lib\win32
    [main] [14:59:17:54] [Library.load:206] Loading orasrvm11.dll...
    [main] [14:59:17:54] [Library.load:212] oracleHome D:\app\Administrator\product\11.1.0\db_1
    [main] [14:59:17:54] [Library.load:227] Property oracle.installer.library_loc is set to value=D:\app\Administrator\product\11.1.0\db_1\oui\lib\win32
    [main] [14:59:17:54] [Library.load:229] Loading library D:\app\Administrator\product\11.1.0\db_1\oui\lib\win32\orasrvm11.dll
    [main] [14:59:17:54] [Library.load:262] Loaded library D:\app\Administrator\product\11.1.0\db_1\oui\lib\win32\orasrvm11.dll from path=
    D:\app\Administrator\product\11.1.0\db_1\oui\lib\win32
    [main] [14:59:17:54] [Version.isPre10i:213] isPre10i.java: Returning FALSE
    [main] [14:59:17:54] [WindowsSystem.regKeyExists:1104] WindowsSystem.regKeyExists: mainkey= HKEY_LOCAL_MACHINE subkey = Software\Oracle\Ocr
    [main] [14:59:17:648] [WindowsSystem.getCSSConfigType:1271] configType=null
    [main] [14:59:17:648] [Host.checkOPS:2308] cluster existence:false
    [main] [14:59:17:648] [Host.checkOPS:2346] Cluster installed=false
    [AWT-EventQueue-0] [14:59:19:429] [OsUtilsBase.getOracleBaseFromInventory:413] Getting ORACLE_BASE from inventory.
    [AWT-EventQueue-0] [14:59:19:445] [OsUtilsBase.getOracleBaseFromInventory:427] Setting oracle.installer.oui_loc = D:\app\Administrator\product\11.1.0\db_1\oui
    [AWT-EventQueue-0] [14:59:19:445] [OsUtilsBase.getOracleBaseFromInventory:430] Oracle Inventory property invptrloc = null
    [AWT-EventQueue-0] [14:59:19:445] [OsUtilsBase.getOracleBaseFromInventory:447] Inventory Home Size = 1
    [AWT-EventQueue-0] [14:59:19:445] [OsUtilsBase.getOracleBaseFromInventory:473] ORACLE_BASE returned from inventory:= D:\app\Administrator
    [AWT-EventQueue-0] [14:59:19:445] [Host.displayOracleBaseWarning:1115] oracleBaseFromEnv null
    [AWT-EventQueue-0] [14:59:19:445] [Host.displayOracleBaseWarning:1121] oracle_base from OUI D:\app\Administrator oracle_base from env
    [AWT-EventQueue-0] [14:59:24:538] [InitParamHandler.endElement:517] CustomSGA flag: false
    [AWT-EventQueue-0] [14:59:24:538] [InitParamHandler.endElement:518] Database Type: MULTIPURPOSE
    [AWT-EventQueue-0] [14:59:24:538] [InitParamHandler.endElement:519] Mem Percentage: 40
    [AWT-EventQueue-0] [14:59:24:538] [InitParamHandler.endElement:552] Total memory MB: 818
    [AWT-EventQueue-0] [14:59:24:538] [OsUtilsBase.is64Bit:294] architecture is 64 bit: false
    [AWT-EventQueue-0] [14:59:24:538] [MemoryCalculator.<clinit>:92] setting memory minimums for 32 bit
    [AWT-EventQueue-0] [14:59:24:554] [StorageAttributes.setAttribute:237] IN threadID:1 group#=1
    [AWT-EventQueue-0] [14:59:24:554] [StorageAttributes.setAttribute:237] IN threadID:1 group#=2
    [AWT-EventQueue-0] [14:59:24:554] [StorageAttributes.setAttribute:246] Current threadID=1
    [AWT-EventQueue-0] [14:59:24:554] [StorageAttributes.setAttribute:253] Current threadID=1 ==> redoGroups[0]=1
    [AWT-EventQueue-0] [14:59:24:554] [StorageAttributes.setAttribute:263] vRedoGroups:[1]
    [AWT-EventQueue-0] [14:59:24:554] [StorageAttributes.setAttribute:293] setAttribute: bExists=false
    [AWT-EventQueue-0] [14:59:24:554] [StorageAttributes.setAttribute:237] IN threadID:1 group#=3
    [AWT-EventQueue-0] [14:59:24:554] [StorageAttributes.setAttribute:246] Current threadID=1
    [AWT-EventQueue-0] [14:59:24:554] [StorageAttributes.setAttribute:253] Current threadID=1 ==> redoGroups[0]=1
    [AWT-EventQueue-0] [14:59:24:554] [StorageAttributes.setAttribute:253] Current threadID=1 ==> redoGroups[1]=2
    [AWT-EventQueue-0] [14:59:24:554] [StorageAttributes.setAttribute:263] vRedoGroups:[1, 2]
    [AWT-EventQueue-0] [14:59:24:554] [StorageAttributes.setAttribute:293] setAttribute: bExists=false
    [AWT-EventQueue-0] [14:59:24:554] [TemplateManager.parseCloneTemplate:1509] See for any transportable datafiles in TemplateManager.....
    [AWT-EventQueue-0] [14:59:36:804] [DatabaseNamePage.onNameChanged:751] m_oldGDBName =
    [AWT-EventQueue-0] [14:59:36:819] [OracleHome.isRacEnabled:177] bRacOn = false
    [AWT-EventQueue-0] [14:59:36:897] [ManagementOptionsPage.initializePage:692] Enumerating Agents
    [AWT-EventQueue-0] [14:59:41:350] [ManagementOptionsPage.validate:1049] emConfigChkBox State: true
    [AWT-EventQueue-0] [14:59:41:350] [ManagementOptionsPage.validate:1050] switchChkBox State: false
    [AWT-EventQueue-0] [14:59:41:350] [ManagementOptionsPage.validateAndSetParams:1090] In validateAndSetParams
    [AWT-EventQueue-0] [14:59:41:350] [NetworkUtils.checkListenerStatus:350] Checking listener status: LISTENER
    [AWT-EventQueue-0] [14:59:41:569] [NetworkUtils.checkListenerStatus:356] lsnrct status output
    LSNRCTL for 32-bit Windows: Version 11.1.0.6.0 - Production on 13-JUN-2008 14:59:41
    Copyright (c) 1991, 2007, Oracle. All rights reserved.
    Connecting to (DESCRIPTION=(ADDRESS=(PROTOCOL=TCP)(HOST=XXX)(PORT=1521)))
    STATUS of the LISTENER
    Alias LISTENER
    Version TNSLSNR for 32-bit Windows: Version 11.1.0.6.0 - Production
    Start Date 13-JUN-2008 10:32:10
    Uptime 0 days 4 hr. 27 min. 33 sec
    Trace Level user
    Security ON: Local OS Authentication
    SNMP OFF
    Listener Parameter File D:\app\Administrator\product\11.1.0\db_1\network\admin\listener.ora
    Listener Log File d:\app\administrator\diag\tnslsnr\XXX\listener\alert\log.xml
    Listener Trace File d:\app\administrator\diag\tnslsnr\XXX\listener\trace\ora_1792_1876.trc
    Listening Endpoints Summary...
    (DESCRIPTION=(ADDRESS=(PROTOCOL=tcp)(HOST=XXX)(PORT=1521)))
    Services Summary...
    Service "Oracle8" has 2 instance(s).
    Instance "ORCL", status UNKNOWN, has 1 handler(s) for this service...
    Instance "XXX", status UNKNOWN, has 1 handler(s) for this service...
    Service "XXX" has 1 instance(s).
    Instance "XXX", status READY, has 1 handler(s) for this service...
    Service "XXX" has 1 instance(s).
    Instance "XXX", status READY, has 1 handler(s) for this service...
    Service "XXX" has 1 instance(s).
    Instance "XXX", status READY, has 1 handler(s) for this service...
    Service "XXX" has 1 instance(s).
    Instance "XXX", status READY, has 1 handler(s) for this service...
    Service "XXX" has 1 instance(s).
    Instance "XXX", status READY, has 1 handler(s) for this service...
    Service "XXX" has 1 instance(s).
    Instance "XXX", status READY, has 1 handler(s) for this service...
    The command completed successfully
    [AWT-EventQueue-0] [15:0:14:475] [TemplateManager.isInstallTemplate:2226] Selected Template by user:=General Purpose
    [AWT-EventQueue-0] [15:0:14:475] [TemplateManager.isInstallTemplate:2233] The Message Id to be searched:=GENERAL_PURPOSE
    [AWT-EventQueue-0] [15:0:14:490] [TemplateManager.isInstallTemplate:2226] Selected Template by user:=General Purpose
    [AWT-EventQueue-0] [15:0:14:490] [TemplateManager.isInstallTemplate:2233] The Message Id to be searched:=GENERAL_PURPOSE
    [AWT-EventQueue-0] [15:0:14:506] [TemplateManager.isInstallTemplate:2226] Selected Template by user:=General Purpose
    [AWT-EventQueue-0] [15:0:14:506] [TemplateManager.isInstallTemplate:2233] The Message Id to be searched:=GENERAL_PURPOSE
    [AWT-EventQueue-0] [15:0:15:84] [Verifier.processRawConfigFile:3709] StorageType == 0
    [AWT-EventQueue-0] [15:0:15:84] [DatabaseOptionPage.doNext:266] processRawConfigFile=false
    [AWT-EventQueue-0] [15:0:17:724] [InitParametersPage.validate:225] InitParametersPage->validate: in SPFile condition
    [AWT-EventQueue-0] [15:0:17:724] [InitParametersPage.validate:229] original spFileName with variables ={ORACLE_HOME}\database\spfile{SID}.ora
    [AWT-EventQueue-0] [15:0:17:724] [Verifier.validateSPFile:4853] The SPFILE={ORACLE_HOME}\database\spfile{SID}.ora
    [AWT-EventQueue-0] [15:0:17:724] [ClusterUtils.validateRawDevice:784] Inside validateRawDevice
    [AWT-EventQueue-0] [15:0:17:724] [ClusterUtils.validateRawDevice:796] Device Exception
    [AWT-EventQueue-0] [15:0:17:724] [ClusterUtils.isRaw:762] The handle is invalid.
    [AWT-EventQueue-0] [15:0:17:724] [InitParametersPage.validate:337] InitParametersPage->validate: calling createUndoAttributes
    [AWT-EventQueue-0] [15:0:19:131] [StoragePage.selectAndExpandItem:333] item=oracle.sysman.emSDK.client.dataComponent.dataDrivenTree.TreeParentNode[label=Storage,index=0]
    [AWT-EventQueue-0] [15:0:19:178] [StoragePage.selectAndExpandItem:342] item.isExpanded=true
    [AWT-EventQueue-0] [15:0:19:193] [StoragePage.selectAndExpandItem:354] selection.isSelected(item)=false
    [AWT-EventQueue-0] [15:0:20:771] [StoragePage.validate:934] StoragePage: In validate function
    [AWT-EventQueue-0] [15:0:20:787] [Verifier.validateTemplate:1780] StorageType == 0
    [AWT-EventQueue-0] [15:0:20:818] [CreationOptionPage.setSaveTemplateVisible:465] Setting save template visible = true
    [AWT-EventQueue-0] [15:0:23:177] [DBCAWizard.onFinish:1203] m_bFinishClicked: false
    [TaskScheduler timer] [15:0:23:177] [OracleHome.isRacEnabled:177] bRacOn = false
    [TaskScheduler timer] [15:0:23:193] [NetworkUtils.checkListenerStatus:350] Checking listener status: LISTENER
    [TaskScheduler timer] [15:0:23:412] [NetworkUtils.checkListenerStatus:356] lsnrct status output
    LSNRCTL for 32-bit Windows: Version 11.1.0.6.0 - Production on 13-JUN-2008 15:00:23
    Copyright (c) 1991, 2007, Oracle. All rights reserved.
    Connecting to (DESCRIPTION=(ADDRESS=(PROTOCOL=TCP)(HOST=XXX)(PORT=1521)))
    STATUS of the LISTENER
    Alias LISTENER
    Version TNSLSNR for 32-bit Windows: Version 11.1.0.6.0 - Production
    Start Date 13-JUN-2008 10:32:10
    Uptime 0 days 4 hr. 28 min. 14 sec
    Trace Level user
    Security ON: Local OS Authentication
    SNMP OFF
    Listener Parameter File D:\app\Administrator\product\11.1.0\db_1\network\admin\listener.ora
    Listener Log File d:\app\administrator\diag\tnslsnr\XXX\listener\alert\log.xml
    Listener Trace File d:\app\administrator\diag\tnslsnr\XXX\listener\trace\ora_1792_1876.trc
    Listening Endpoints Summary...
    (DESCRIPTION=(ADDRESS=(PROTOCOL=tcp)(HOST=XXX)(PORT=1521)))
    Services Summary...
    Service "Oracle8" has 2 instance(s).
    Instance "ORCL", status UNKNOWN, has 1 handler(s) for this service...
    Instance "XXX", status UNKNOWN, has 1 handler(s) for this service...
    Service "XXX" has 1 instance(s).
    Instance "XXX", status READY, has 1 handler(s) for this service...
    Service "XXX" has 1 instance(s).
    Instance "XXX", status READY, has 1 handler(s) for this service...
    Service "XXX" has 1 instance(s).
    Instance "XXX", status READY, has 1 handler(s) for this service...
    Service "XXX" has 1 instance(s).
    Instance "XXX", status READY, has 1 handler(s) for this service...
    Service "XXX" has 1 instance(s).
    Instance "XXX", status READY, has 1 handler(s) for this service...
    Service "XXX" has 1 instance(s).
    Instance "XXX", status READY, has 1 handler(s) for this service...
    The command completed successfully
    [TaskScheduler timer] [15:0:23:412] [Verifier.validateTemplate:1780] StorageType == 0
    [TaskScheduler timer] [15:0:23:412] [Verifier.calculateCloneDatafilePathsAndSizes:3137] canonicalPath=D:\app\Administrator\oradata\
    [TaskScheduler timer] [15:0:23:412] [Verifier.calculateCloneDatafilePathsAndSizes:3137] canonicalPath=D:\app\Administrator\oradata\
    [TaskScheduler timer] [15:0:23:412] [Verifier.calculateCloneDatafilePathsAndSizes:3137] canonicalPath=D:\app\Administrator\oradata\
    [TaskScheduler timer] [15:0:23:412] [Verifier.calculateCloneDatafilePathsAndSizes:3137] canonicalPath=D:\app\Administrator\oradata\
    [TaskScheduler timer] [15:0:23:412] [Verifier.calculateCloneDatafilePathsAndSizes:3137] canonicalPath=D:\app\Administrator\oradata\
    [TaskScheduler timer] [15:0:23:443] [Verifier.calculateRedoLogGroupFileSizes:3269] canonicalPath=D:\app\Administrator\oradata\
    [TaskScheduler timer] [15:0:23:443] [Verifier.calculateRedoLogGroupFileSizes:3269] canonicalPath=D:\app\Administrator\oradata\
    [TaskScheduler timer] [15:0:23:443] [Verifier.calculateRedoLogGroupFileSizes:3269] canonicalPath=D:\app\Administrator\oradata\
    [TaskScheduler timer] [15:0:23:443] [Verifier.getControlfFileSizes:3187] No. of Control files:=3
    [Thread-15] [15:0:23:474] [UIHost.getHtmlSummary:1055] UIHost:getHtmlSummary: running createTemplateDocument NOT for SHOW_TEMPLATE condition
    [Thread-15] [15:0:23:490] [UIHost.getHtmlSummary:1061] UIHost:getHtmlSummary: after running createTemplateDocument NOT for SHOW_TEMPLATE condition
    [Thread-15] [15:0:23:615] [XSLConst.generateHeaderXSL:211] NLS:Template Name:= null
    [Thread-15] [15:0:23:615] [XSLConst.generateHeaderXSL:216] Template Description:= Use this database template to create a pre-configured database optimized for general purpose or transaction processing usage.
    [Thread-15] [15:0:23:865] [UIHost.getHtmlSummary:1100] UIHost->getHtmlSummary: start printing of transformed document
    [TaskScheduler timer] [15:0:26:302] [Host.executeSteps:4426] Executing steps....
    [TaskScheduler timer] [15:0:26:302] [Host.setUpForOperation:3198] setUpForOperation: Mode = 128
    [TaskScheduler timer] [15:0:26:302] [OsUtilsBase.copyFile:715] OsUtilsBase.copyFile:
    [TaskScheduler timer] [15:0:26:318] [OsUtilsBase.copyFile:763] **write of file at destination complete...
    [TaskScheduler timer] [15:0:26:318] [OsUtilsBase.copyFile:798] **file copy status:= true
    [TaskScheduler timer] [15:1:13:879] [OsUtilsBase.copyFile:763] **write of file at destination complete...
    [TaskScheduler timer] [15:1:13:879] [OsUtilsBase.copyFile:798] **file copy status:= true
    [TaskScheduler timer] [15:1:13:895] [Host.executeSteps:4605] setupForOperation returned: true
    [TaskScheduler timer] [15:1:13:895] [Host.createStepSQLInterface:6444] sid =Mapping
    [TaskScheduler timer] [15:1:13:895] [SQLEngine.getEnvParams:446] NLS_LANG: AMERICAN_AMERICA.WE8MSWIN1252
    [TaskScheduler timer] [15:1:13:895] [SQLEngine.initialize:253] Execing SQLPLUS/SVRMGR process...
    [TaskScheduler timer] [15:1:14:35] [SQLEngine.initialize:281] m_bReaderStarted: false
    [TaskScheduler timer] [15:1:14:35] [SQLEngine.initialize:285] Starting Reader Thread...
    [Thread-35] [15:1:14:35] [StepContext$ModeRunner.run:2487] ---- Progress Needed:=true
    [Thread-35] [15:1:14:67] [BasicStep.execute:202] Executing Step : CLONE_DB_CREATION_RMAN_RESTORE
    [Thread-35] [15:1:14:67] [StepErrorHandler.setFatalErrors:322] setting Fatal Error: ORA-01092
    [Thread-35] [15:1:14:67] [StepErrorHandler.setFatalErrors:322] setting Fatal Error: ORA-01034
    [Thread-35] [15:1:14:67] [StepErrorHandler.setFatalErrors:322] setting Fatal Error: ORA-03114
    [Thread-35] [15:1:14:67] [StepErrorHandler.setFatalErrors:322] setting Fatal Error: ORA-12560
    [Thread-35] [15:1:14:67] [StepErrorHandler.setIgnorableErrors:250] setting Ignorable Error: ORA-01109
    [Thread-35] [15:1:14:67] [BasicStep.configureSettings:304] messageHandler being set=oracle.sysman.assistants.util.UIMessageHandler@e1666
    [Thread-35] [15:1:14:67] [BasicStep.execute:202] Executing Step : INSTANCE_CREATION
    [Thread-35] [15:1:14:67] [BasicStep.configureSettings:304] messageHandler being set=oracle.sysman.assistants.util.UIMessageHandler@e1666
    [Thread-35] [15:1:14:67] [InitParamAttributes.sortParams:3562] m_sortOn:4 sortOn:4
    [Thread-35] [15:1:14:67] [InitParamAttributes.createInitFile:1698] Processing init param db_block_size
    [Thread-35] [15:1:14:67] [InitParamAttributes.createInitFile:1703] The value is 8192
    [Thread-35] [15:1:14:67] [InitParamAttributes.createInitFile:1698] Processing init param open_cursors
    [Thread-35] [15:1:14:67] [InitParamAttributes.createInitFile:1703] The value is 300
    [Thread-35] [15:1:14:67] [InitParamAttributes.createInitFile:1698] Processing init param db_domain
    [Thread-35] [15:1:14:67] [InitParamAttributes.createInitFile:1703] The value is ""
    [Thread-35] [15:1:14:192] [InitParamAttributes.createInitFile:1698] Processing init param db_name
    [Thread-35] [15:1:14:192] [InitParamAttributes.createInitFile:1703] The value is Mapping
    [Thread-35] [15:1:14:192] [InitParamAttributes.createInitFile:1698] Processing init param control_files
    [Thread-35] [15:1:14:192] [InitParamAttributes.createInitFile:1703] The value is ("D:\app\Administrator\oradata\Mapping\control01.ctl", "D:\app\Administrator\oradata\Mapping\control02.ctl", "D:\app\Administrator\oradata\Mapping\control03.ctl")
    [Thread-35] [15:1:14:192] [InitParamAttributes.createInitFile:1698] Processing init param db_recovery_file_dest
    [Thread-35] [15:1:14:192] [InitParamAttributes.createInitFile:1703] The value is D:\app\Administrator\flash_recovery_area
    [Thread-35] [15:1:14:192] [InitParamAttributes.createInitFile:1698] Processing init param db_recovery_file_dest_size
    [Thread-35] [15:1:14:192] [InitParamAttributes.createInitFile:1703] The value is 2147483648
    [Thread-35] [15:1:14:192] [InitParamAttributes.createInitFile:1698] Processing init param compatible
    [Thread-35] [15:1:14:192] [InitParamAttributes.createInitFile:1703] The value is 11.1.0.0.0
    [Thread-35] [15:1:14:192] [InitParamAttributes.createInitFile:1698] Processing init param diagnostic_dest
    [Thread-35] [15:1:14:192] [InitParamAttributes.createInitFile:1703] The value is D:\app\Administrator
    [Thread-35] [15:1:14:192] [InitParamAttributes.createInitFile:1698] Processing init param memory_target
    [Thread-35] [15:1:14:192] [InitParamAttributes.createInitFile:1703] The value is 857735168
    [Thread-35] [15:1:14:192] [InitParamAttributes.createInitFile:1698] Processing init param processes
    [Thread-35] [15:1:14:192] [InitParamAttributes.createInitFile:1703] The value is 150
    [Thread-35] [15:1:14:192] [InitParamAttributes.createInitFile:1698] Processing init param audit_file_dest
    [Thread-35] [15:1:14:192] [InitParamAttributes.createInitFile:1703] The value is D:\app\Administrator\admin\Mapping\adump
    [Thread-35] [15:1:14:192] [InitParamAttributes.createInitFile:1698] Processing init param audit_trail
    [Thread-35] [15:1:14:192] [InitParamAttributes.createInitFile:1703] The value is db
    [Thread-35] [15:1:14:192] [InitParamAttributes.createInitFile:1698] Processing init param remote_login_passwordfile
    [Thread-35] [15:1:14:192] [InitParamAttributes.createInitFile:1703] The value is EXCLUSIVE
    [Thread-35] [15:1:14:192] [InitParamAttributes.createInitFile:1698] Processing init param dispatchers
    [Thread-35] [15:1:14:192] [InitParamAttributes.createInitFile:1703] The value is "(PROTOCOL=TCP) (SERVICE=MappingXDB)"
    [Thread-35] [15:1:14:192] [InitParamAttributes.createInitFile:1698] Processing init param undo_tablespace
    [Thread-35] [15:1:14:192] [InitParamAttributes.createInitFile:1703] The value is UNDOTBS1
    [Thread-35] [15:1:14:473] [OracleHome.isRacEnabled:177] bRacOn = false
    [Thread-35] [15:1:14:473] [Host.noEntryinOratab:5562] Check made for oratab arg passed............
    [Thread-35] [15:1:14:473] [Oradim.getAddEntryCommand:355] AddEntry=[D:\app\Administrator\product\11.1.0\db_1\bin\oradim.exe, -new, -sid, MAPPING, -startmode, manual, -spfile]
    [Thread-35] [15:1:19:35] [Oradim.getEditEntryCommand:436] getEditEntry cmd=[D:\app\Administrator\product\11.1.0\db_1\bin\oradim.exe, -edit, -sid, MAPPING, -startmode, auto, -srvcstart, system]
    [Thread-35] [15:1:19:520] [Oradim.addSidToRegistry:884] oracleHomeKey: SOFTWARE\ORACLE\KEY_OraDb11g_home1
    [Thread-35] [15:1:19:520] [OracleHome.getVersion:974] OracleHome.getVersion called. Current Version: 11.1.0.6.0
    [Thread-35] [15:1:19:520] [OracleHome.getVersion:1046] Current Version From Inventory: 11.1.0.6.0
    [Thread-35] [15:1:19:520] [OracleHome.getVersion:1051] using sqlplus: D:\app\Administrator\product\11.1.0\db_1\bin\sqlplus.exe
    [Thread-35] [15:1:19:520] [CommonUtils.createPasswordFile:515] calling new orapwd for version : 11.1.0.6.0
    [Thread-35] [15:1:19:520] [OsUtilsBase.execProg:1335] beginning execProg with input array.
    [Thread-35] [15:1:19:520] [OsUtilsBase.execProg:1347] finished execProg with input array.
    [Thread-35] [15:1:19:520] [BasicStep.configureSettings:304] messageHandler being set=oracle.sysman.assistants.util.UIMessageHandler@e1666
    [Thread-35] [15:1:19:520] [CloneRmanRestoreStep.executeImpl:218] Instance Creation went fine..........
    [Thread-35] [15:1:19:520] [CloneRmanRestoreStep.executeImpl:225] db_recovery_file_dest=D:\app\Administrator\flash_recovery_area
    [Thread-35] [15:1:19:520] [CloneRmanRestoreStep.executeImpl:228] db_recovery_file_dest_size=2147483648
    [Thread-35] [15:1:23:67] [SQLEngine.done:2014] Done called
    oracle.sysman.assistants.util.sqlEngine.SQLFatalErrorException: ORA-01034: ORACLE not available
         at oracle.sysman.assistants.util.sqlEngine.SQLEngine.executeImpl(SQLEngine.java:1530)
         at oracle.sysman.assistants.util.sqlEngine.SQLEngine.connect(SQLEngine.java:853)
         at oracle.sysman.assistants.dbca.backend.CloneRmanRestoreStep.executeImpl(CloneRmanRestoreStep.java:245)
         at oracle.sysman.assistants.util.step.BasicStep.execute(BasicStep.java:210)
         at oracle.sysman.assistants.util.step.Step.execute(Step.java:140)
         at oracle.sysman.assistants.util.step.StepContext$ModeRunner.run(StepContext.java:2497)
         at java.lang.Thread.run(Thread.java:595)
    [Thread-35] [15:1:23:207] [BasicStep.configureSettings:304] messageHandler being set=oracle.sysman.assistants.util.UIMessageHandler@e1666
    [Thread-35] [15:1:23:207] [StepContext$ModeRunner.run:2513] oracle.sysman.assistants.dbca.backend.CloneRmanRestoreStep.executeImpl(CloneRmanRestoreStep.java:448)
    oracle.sysman.assistants.util.step.BasicStep.execute(BasicStep.java:210)
    oracle.sysman.assistants.util.step.Step.execute(Step.java:140)
    oracle.sysman.assistants.util.step.StepContext$ModeRunner.run(StepContext.java:2497)
    java.lang.Thread.run(Thread.java:595)
    [Thread-35] [15:1:24:191] [SQLEngine.done:2014] Done called
    [Thread-35] [15:1:24:191] [Host.createStepSQLInterface:6444] sid =Mapping
    [Thread-35] [15:1:24:207] [SQLEngine.getEnvParams:446] NLS_LANG: AMERICAN_AMERICA.WE8MSWIN1252
    [Thread-35] [15:1:24:207] [SQLEngine.initialize:253] Execing SQLPLUS/SVRMGR process...
    [Thread-35] [15:1:24:207] [SQLEngine.initialize:281] m_bReaderStarted: false
    [Thread-35] [15:1:24:207] [SQLEngine.initialize:285] Starting Reader Thread...
    [Thread-35] [15:1:24:348] [CloneRmanRestoreStep.cancel:614] Cleaning up partially extracted files failed
    [Thread-35] [15:1:24:535] [SQLEngine.done:2014] Done called
    [Thread-35] [15:1:24:754] [CloneRmanRestoreStep.cancel:629] RMAN cleanup failed
    [Thread-35] [15:1:24:754] [CloneRmanRestoreStep.cancel:630] oracle.sysman.assistants.util.sqlEngine.SQLEngine.executeImpl(SQLEngine.java:1530)
    oracle.sysman.assistants.util.sqlEngine.SQLEngine.connect(SQLEngine.java:853)
    oracle.sysman.assistants.dbca.backend.CloneRmanRestoreStep.cancel(CloneRmanRestoreStep.java:617)
    oracle.sysman.assistants.util.step.StepContext$ModeRunner.cancel(StepContext.java:2590)
    oracle.sysman.assistants.util.step.StepContext$ModeRunner.run(StepContext.java:2552)
    java.lang.Thread.run(Thread.java:595)
    [Thread-35] [15:1:24:754] [CloneDBCreationStep.cancel:895] CloneDBCreationStep.cancel(): nodeList= null
    [Thread-35] [15:1:24:754] [SQLEngine.reInitialize:624] Reinitializing SQLEngine...
    [Thread-35] [15:1:24:754] [SQLEngine.getEnvParams:446] NLS_LANG: AMERICAN_AMERICA.WE8MSWIN1252
    [Thread-35] [15:1:24:754] [SQLEngine.initialize:253] Execing SQLPLUS/SVRMGR process...
    [Thread-35] [15:1:24:754] [SQLEngine.initialize:281] m_bReaderStarted: false
    [Thread-35] [15:1:24:863] [SQLEngine.initialize:285] Starting Reader Thread...
    [Thread-35] [15:1:24:863] [SQLEngine.initialize:334] Waiting for m_bReaderStarted to be true
    [Thread-35] [15:1:25:254] [SQLEngine.done:2014] Done called
    [Thread-35] [15:1:25:473] [OracleHome.isRacEnabled:177] bRacOn = false
    [Thread-35] [15:1:25:473] [Oradim.getRemoveEntryCommand:477] Execing D:\app\Administrator\product\11.1.0\db_1\bin\oradim.exe -delete -sid MAPPING
    [Thread-35] [15:1:27:316] [SQLEngine.done:2014] Done called
    [AWT-EventQueue-0] [15:1:30:379] [SQLEngine.done:2014] Done called
    Any help will be greatly appreciated.

    I found out from another dba that oracle takes up 1 gb of memory by default. So creating 2 databases with default settings means that 2 gb of memory was taken on a windows 32 bit server. So that means only 2 gb is available to the machine, and the server takes up more than half of that making the memory available scarce for another oracle database to be created.

  • AC20000027 - Element PURCHASINGCONTRACT is not available in the container

    Hi Experts,
    I created the simple WF. Start event - RELEASESTEPCREATED, object - BUS2012. WF have one step, task - "TS20000166 Release of purchase order". When I start my workflow in test run the status is set to "error" - "Element PURCHASINGCONTRACT is not available in the container", Message no. OL356.
    Testing the WF in transaction SWUD don't found any errors.
    Can anybody help me with this?
    Thanks a lot,
    Andrei

    Hi Andru,
    I worked on Release strategy for PO.
    I am SURE the Problem is with RULE, RULE is not returning any agents.
    The problem is that the rule (20000027) your using is not returning any agents.
    1. Just go to transaction PFAC and enter the RULE. go to display mode.
    Click on the second button in the application tool bar (SIMULATE RULE RESOLUTION - shift + F8 ) a popup sort of screen will be displayed.
    Now enter the PO number and Release code.
    Execute the DETAIL button in the middle of the popup screen.
    See if the rule is returning any agents.
    2. If the above mentioned test is successfully executed,
    Check in your workflow SWDD.
    Open your workflow.
    Double click on the RELEASE CODE workflow container element.
    Check if the structure defined under D.Type is - T16FC and Field is - FRGCO.
    3. If even the 2nd point is correct,
    Check the binding for the step '0000000004'
    In the binding, Pass the Busines Object BUS2012 (Purchase order) and RELEASE CODE (T16FC-FRGCO) that are present in the Workflow container. and check if they both are import and export parameters.
    3. You are sure that the agents are specified in SPRO (under Material management >> PURCHASING >> Purchase order/CONTRACT >> Release Procedure for PO/CONTRACT >> Define release procedure for PO / CONTRACT >> WORKFLOW )
    If you double click on WORKFLOW, you will see that the AGENT IDs are assigned to the Release Group and release codes of Purchase Order.
    Eg: Release group       Release code  Agent Type        Agent ID
                01                          L1               US                SAPUSER1
                01                          L2               US                SAPUSER2
                01                          L3               US                SAPUSER3
                01                          L4               US                SAPUSER4
    Try all this and tell me the result.
    Regards,
    Gautham Paspala

  • J2EE database not available via test

    Hi Experts,
    I have installed NW 7.0 portal Ehp2 system on Solaris cluster with logical name  . Now when I try to start SAP with the startsap script . it is not starting the database (Oracle) . It just checks whether DB is running and giving message J2EE database not available via test , and asking to check JdbCon.log .
    This is Bootstrap log
    com.sap.engine.frame.core.configuration.ConfigurationException: Error while connecting to DB.
            at com.sap.engine.core.configuration.impl.persistence.rdbms.DBConnectionPool.createConnection(DBConnectionPool.java:384)
            at com.sap.engine.core.configuration.impl.persistence.rdbms.DBConnectionPool.<init>(DBConnectionPool.java:132)
            at com.sap.engine.core.configuration.impl.persistence.rdbms.PersistenceHandler.<init>(PersistenceHandler.java:38)
            at com.sap.engine.core.configuration.impl.cache.ConfigurationCache.<init>(ConfigurationCache.java:150)
            at com.sap.engine.core.configuration.bootstrap.ConfigurationManagerBootstrapImpl.init(ConfigurationManagerBootstrapImpl.java:214)
            at com.sap.engine.core.configuration.bootstrap.ConfigurationManagerBootstrapImpl.<init>(ConfigurationManagerBootstrapImpl.java:46)
            at com.sap.engine.bootstrap.Synchronizer.<init>(Synchronizer.java:76)
            at com.sap.engine.bootstrap.Bootstrap.initDatabaseConnection(Bootstrap.java:474)
            at com.sap.engine.bootstrap.Bootstrap.<init>(Bootstrap.java:156)
            at com.sap.engine.bootstrap.Bootstrap.main(Bootstrap.java:946)
            at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
            at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
            at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
            at java.lang.reflect.Method.invoke(Method.java:331)
            at com.sap.engine.offline.OfflineToolStart.main(OfflineToolStart.java:81)
    Caused by: java.sql.SQLException: ORA-01034: ORACLE not available
    ORA-27101: shared memory realm does not exist
    SVR4 Error: 2: No such file or directory
    If I start db manually , then runnig startsap , it runs fine . And also stopsap script , wont stop database .
    I checked listener.ora,tnsnames.ora, user profiles, environment variables, permissions etc , nothing found wrong .
    When I execute startj2eedb as sidadm , it starts db .
    Can you please throw some light on this issue
    Regards,
    Nibu Antony

    Yes, it is all logical hostname .
    For more clarification on the issue
    bash-3.2$ startsap <logical hostname>
    Checking <SID> Database
    J2EE Database is not available via  test
    See logfile /usr/sap/<SID>//JdbcCon.log
    Starting Startup Agent sapstartsrv
    Instance Service on host hostname<SID>p1 started
    Starting SAP Instance SCS00
    Startup-Log is written to /usr/sap/<SID>//startsap_SCS00_hostname<SID>p1.log
    Instance on host hostname<SID>p1 started
    Starting Startup Agent sapstartsrv
    Instance Service on host hostname<SID>p1 started
    Starting SAP Instance JC01
    Startup-Log is written to /usr/sap/<SID>//startsap_JC01_hostname<SID>p1.log
    Instance on host hostname<SID>p1 started
    Starting SAP-Collector Daemon
    12:38:20 13.03.2014     LOG: Using PerfDir (DIR_PERF) = /usr/sap/tmp
    * This is Saposcol Version COLL 21.04 720 - V3.83 64Bit
    * Usage:  saposcol -l: Start OS Collector
    *         saposcol -k: Stop  OS Collector
    *         saposcol -d: OS Collector Dialog Mode
    *         saposcol -s: OS Collector Status
    * Starting collector (create new process)
    saposcol on host hostname<SID>p1 started
    bash-3.2$ sqlplus / as sysdba
    SQL*Plus: Release 11.2.0.3.0 Production on Thu Mar 13 12:38:34 2014
    Copyright (c) 1982, 2011, Oracle.  All rights reserved.
    Connected to an idle instance.
    here is the logs
    JdbcCon.log
    Start: Thu Mar 13 12:36:48 AST 2014
    Error: Couldn't connect to DB
    java.sql.SQLException: ORA-01034: ORACLE not available
    ORA-27101: shared memory realm does not exist
    SVR4 Error: 2: No such file or directory

  • Log Reader Agent is not able to read Transaction Log of Publisher database.

    Hi,
    There is no restore or change in recovery model or detach-attach action performed on my production database but still I am seeing below error message from Log Reader Agent-
    Error messages:
    The process could not execute 'sp_repldone/sp_replcounters' on 'ProdInstance'. (Source: MSSQL_REPL, Error number: MSSQL_REPL20011)
    Get help:
    An error occurred while processing the log for database 'MyDatabase'.  If possible, restore from backup. If a backup is not available, it might be necessary to rebuild the log. (Source: MSSQLServer, Error number: 9004)
    The process could not set the last distributed transaction. (Source: MSSQL_REPL, Error number: MSSQL_REPL22017)
    Get help: The process could not execute 'sp_repldone/sp_replcounters' on 'ProdInstance'. (Source: MSSQL_REPL, Error number: MSSQL_REPL22037)
    Note- CheckDB on production and distribution database executed successfully. Also, I need subscriber to be a true copy of publisher so I think sp_replrestart is not an option for me.
    My question is how to resolve this issue? I am thinking that reinitialization should resolve the problem but what if this is not going to resolve? Do I need to reconfigure the transaction replication? Please sugegst.

    Hi,
    Please check out this link on how to resolve “The process could not execute 'sp_repldone/sp_replcounters'” error.
    http://blogs.msdn.com/b/repltalk/archive/2010/02/19/the-process-could-not-execute-sp-repldone-sp-replcounters.aspx
    The possible cause could be:
    1.The last LSN in Transaction Log is less than what the LSN Log Reader is trying to find. An old backup may have been restored on top of Published Database. After the restore, the new Transaction Log doesn't contain the data now distributor & subscriber(s)
    have.
    2.Database corruption.
    Since you have not restored the published database, I suggest you run
    DBCC CHECKDB to confirm the consistency of the database. Refer to the How to fix in the above link.
    Thanks.
    Tracy Cai
    TechNet Community Support

  • Parner Profile not available

    Hi,
    Could anybody give some help on my problem?
    I am using generic datasource to populate master data.After executing the package , I am geeting an error " EDI: Partner profile not available" .
    Even though the data is getting populating into data target.
    With rgds,
    Anil

    Hi Anil,
    use transaction WE20 to define your partner profile...but what you are facing is very weird...are you already loading something from the same source system ? If this is the first time set in WE20 this settings:
    Partner no.= the source system logical name
    partn type = 'LS'
    typ = US
    agent = (enter a user and language)
    outbound = RSRQST
    inbound = RSINFO and RSSEND
    patrner class = BIW
    if you re using a flat file you don't need outbound parameters.
    If you need more help, let me know (and please don't forget to assign some points by clickin'on the yellow star for each reply to the contributors that help you !!!)
    Bye,
    Roberto

  • 503 Service not available

    Hi.
    Could you help me ....
    I modified a bsp to display a PDF file into a new popup window, and I tested into my quality server using SE80 transaction and with my portal application after and evething was OK but when a released the new BSP into production server the BSP crashed.
    This is the error message that throws:
    503 Service not Available
    Erro 6
    Version 6040
    MOdule http_j2ee_mt.c
    Line  795
    Server EMKIOAIX103_CRP_00
    Details Cannot reach external application server on localhost:50000
    I explain you my program does it:
    1.-  the user chooses the a record into a Inbox
    2.-  IC display its data with BSP IRecReason
    3.- the user chooses a Action link
    4.- IC display a action PRINT PDF
    5.- the user selects a print pdf
    This is my code to EH_ONGO than create the PDF spool
    METHOD EH_ONGO .
      DATA: function_name TYPE rs38l_fnam,
            dummy(254)        TYPE c,
            ls_archive_index  TYPE toa_dara,
            ls_orderadm_h     TYPE crmt_output_orderadm_h_com,
            ls_activity_h     TYPE crmt_output_activity_h_com,
            ls_lead_h         TYPE crmt_output_lead_h_com,
            ls_opport_h       TYPE crmt_output_opport_h_com,
            ls_orgman_h       TYPE crmt_output_orgman_h_com,
            lt_partner_h      TYPE crmt_output_partner_h_comt,
            ls_pricingdata_h  TYPE crmt_output_pricingdata_h_com,
            ls_sales_h        TYPE crmt_output_sales_h_com,
            ls_shipping_h     TYPE crmt_output_shipping_h_com,
            lt_payplan_d_h    TYPE crmt_output_payplan_d_h_comt,
            ls_customer_h     TYPE crmt_output_customer_h_com,
            ls_cumulat_h     TYPE crmt_output_cumulat_h_com,
            lt_billing_h     TYPE   crmt_output_billing_h_comt,
            lt_cancel_h      TYPE   crmt_output_cancel_h_comt,
            lt_appointment_h TYPE   crmt_output_appointment_h_comt,
            lt_billplan_d_h  TYPE   crmt_output_billplan_d_h_comt,
            lt_billplan_h    TYPE   crmt_output_billplan_h_comt,
            lt_status_d_h    TYPE   crmt_output_status_d_h_comt,
            lt_status_h      TYPE   crmt_output_status_h_comt,
            lt_srv_subject_h TYPE   crmt_output_srv_subject_h_comt,
            lt_srv_reason_h  TYPE   crmt_output_srv_reason_h_comt,
            lt_srv_result_h  TYPE   crmt_output_srv_result_h_comt,
            ls_acs_h         TYPE   crmt_acs_h_com,
            lt_orderadm_i    TYPE   crmt_output_orderadm_i_comt,
            lt_orgman_i      TYPE   crmt_output_orgman_i_comt,
            lt_pricingdata_i TYPE   crmt_output_pricingdata_i_comt,
            lt_pricing_i     TYPE   crmt_output_pricing_i_comt,
            lt_product_i     TYPE   crmt_output_product_i_comt,
            lt_sales_i       TYPE   crmt_output_sales_i_comt,
            lt_shipping_i    TYPE   crmt_output_shipping_i_comt,
            lt_schedlin_i    TYPE   crmt_output_schedlin_i_comt,
            lt_customer_i    TYPE   crmt_output_customer_i_comt,
            lt_partner_i     TYPE   crmt_output_partner_i_comt,
            lt_item_cstics_i TYPE   crmt_item_cstics_tab,
            lt_billing_i     TYPE   crmt_output_billing_i_comt,
            lt_cancel_i      TYPE   crmt_output_cancel_i_comt,
            lt_finprod_i     TYPE   crmt_output_finprod_i_comt,
            lt_ordprp_i      TYPE   crmt_output_ordprp_i_comt,
            lt_appointment_i TYPE   crmt_output_appointment_i_comt,
            lt_billplan_d_i  TYPE   crmt_output_billplan_d_i_comt,
            lt_billplan_i    TYPE   crmt_output_billplan_i_comt,
            lt_orderadm_i_qt TYPE   crmt_output_orderadm_i_qt_comt,
            lt_orderadm_i_in TYPE   crmt_output_orderadm_i_in_comt,
            lt_schedlin_i_cf TYPE   crmt_output_schedlin_i_cf_comt,
            lt_status_i      TYPE   crmt_output_status_i_comt,
            lt_working_set_e_s_bbp TYPE /1cn/working_set_e_s_bbp_t,
            lv_language LIKE  sy-langu,
            lo_order TYPE REF TO cl_doc_crm_order,
            lv_object_guid TYPE crmt_object_guid,
            lv_object_kind TYPE crmt_object_kind,
            lv_status TYPE jstat,
            lt_status TYPE TABLE OF jstat,
            lv_error_tab_wrk TYPE ssferror,
            ls_output_options TYPE ssfcompop.
    DATA: IP_SMART_FORM  Type     TDSFNAME,
          IO_APPL_OBJECT Type Ref To OBJECT,
         IO_APPL_OBJECT Type Ref to CL_DOC_CRM_ORDER_H,
          IO_PARTNER Type Ref to CL_PARTNER_PPF,
          IS_ARCHIVE_INDEX  Type TOA_DARA,
          CT_ARCHIVE_INDEX_TAB Type TSFDARA,
          IS_ARCHIVE_PARAMETERS  Type ARC_PARAMS,
          control_parameters TYPE ssfctrlop,
          IS_MAIL_APPL_OBJ  Type SWOTOBJID,
          IS_MAIL_RECIPIENT  Type SWOTOBJID,
          IS_MAIL_SENDER  Type SWOTOBJID,
          IP_USER_SETTINGS type TDBOOL,
          ES_DOCUMENT_OUTPUT_INFO type SSFCRESPD,
          ES_JOB_OUTPUT_INFO type SSFCRESCL,
          ES_JOB_OUTPUT_OPTIONS type SSFCRESOP.
    Resultado generado: Formato PDF
    data: l_pdf_xstring  type xstring,
          lt_lines       type table of tline,
          ls_line        type tline,
          l_pdf_len      type i,
          lv_exist       type n,
          output_data    type ssfcrescl,
          lp_proc_stat TYPE ppfdtstat.
    Explanation: this method executes the action,
    gets the medium, gets the log, reads the messages from
    the log and gets the spool id from the messages.     JLARIN 27/DIC/06
    DATA: lv_event TYPE REF TO cl_htmlb_event_tableview,
          lv_index TYPE int4,
          wa_action TYPE LINE OF crmt_action_table_tab,
          ls_action_obj TYPE REF TO if_action_ppf,
          lr_action TYPE REF TO cl_trigger_ppf,
          lr_medium type Ref to     if_medium_ppf,
         ls_action TYPE zcrmt_action,
          lv_spoolid TYPE symsgv.
    DATA:
         grid   type ref to cl_crm_ic_gridlayout,
         GLOBAL_EVENT     type string.
    data: lo_trigger type ref to CL_TRIGGER_PPF,
          lo_sfprint type ref to CL_SF_PRINT_PPF,
          lo_typed_context type ref to CL_CRM_IC_IRECACTIONPLAN_CTXT,
          lt_all_data type PPFSINPRRU.
    data: lo_rp_rc type i.
    DATA: lv_head_guid      TYPE crmt_object_guid.
      DATA: adminh     TYPE REF TO cl_crm_ic_irecactionplan_cn02,
            guidstring TYPE string,
            guid       TYPE crmt_object_guid,
            action_obj   TYPE crmt_action_obj_tab,
            lv_object_id TYPE crmt_object_id.
       DATA: request_event TYPE REF TO cl_htmlb_event.
       DATA: event_data    TYPE REF TO if_htmlb_data.
       DATA: event_value  type string.
        event_data = cl_htmlb_manager=>get_event_ex( request ).
        if event_data is bound.
          try.
              request_event ?= event_data.
            catch cx_sy_move_cast_error.
              clear request_event.
          endtry.
          event_value = event_data->event_id.
          translate event_value to lower case.             "#EC TRANSLANG
       endif.
       if event_value <> 'x1_x3_x17_x21_actionitems'.
          return.
       endif.
    TRY.
    create object lo_sfprint.
    create object lo_trigger.
        lv_exist = 1.
      Get the header GUID
        adminh = typed_context->btadminh.
        guidstring = adminh->get_actionitem( 'GUID' ).
        If guidstring is initial.
           return.
        endif.
        lv_head_guid = guidstring.
        CALL FUNCTION 'CRM_ORDERADM_H_READ_OW'
          EXPORTING
            iv_orderadm_h_guid     = lv_head_guid
          IMPORTING
            ev_object_id           = lv_object_id
          EXCEPTIONS
            admin_header_not_found = 1
            OTHERS                 = 2.
           CALL FUNCTION 'CRM_ACTION_GET_TOOLBAR_IL'
          EXPORTING
            iv_ref_guid     = lv_head_guid
            iv_ref_kind     = 'A'
          IMPORTING
            et_action_obj   = action_obj
            et_action_fld   = action_fld
          EXCEPTIONS
            parameter_error = 1
            error_occurred  = 2
            error_message   = 3
            OTHERS          = 4.
        IF sy-subrc EQ 0.
          LOOP AT action_fld INTO wa_action where applkey = lv_object_id.
          ENDLOOP.
          IF sy-subrc EQ 0.
             ls_action_obj = wa_action-action.
             lv_exist = 0.
          ENDIF.
        ENDIF.
        if lv_exist is initial.
        CALL METHOD ls_action_obj->get_processing
         IMPORTING
           ei_processing = lr_medium.
        CALL METHOD lo_trigger->get_appl
          RECEIVING
            result = io_appl_object.
          call function 'CRM_ACTION_GET_APPL_OBJECT'
            exporting
              iv_ref_guid          = lv_head_guid
              iv_ref_kind          = 'A'
            IMPORTING
              EV_APPL_OBJECT       = io_appl_object.
        CALL METHOD lr_medium->execute
          EXPORTING
            io_appl_ref       = io_appl_object   "lo_appl_ref
          RECEIVING
            rp_status         = lp_proc_stat
          EXCEPTIONS
            missing_arguments = 1
            OTHERS            = 2.
      ip_smart_form =  lr_medium->get_smartform( ).
        ip_smart_form = 'ZIMPR_ORDEN_RASTREO'.
      get the function name for this smart form
        CALL FUNCTION 'SSF_FUNCTION_MODULE_NAME'
          EXPORTING
            formname           = ip_smart_form
          IMPORTING
            fm_name            = function_name
          EXCEPTIONS
            no_form            = 1
            no_function_module = 2
            OTHERS             = 3.
    determine the tables of the application
      CALL FUNCTION 'CRM_OUTPUT_SINGLE_READ'
        EXPORTING
          iv_object                 = io_appl_object
          iv_smart_form             = ip_smart_form
          ir_partner                = io_partner
        IMPORTING
          es_output_orderadm_h      = ls_orderadm_h
          es_output_activity_h      = ls_activity_h
          es_output_lead_h          = ls_lead_h
          es_output_opport_h        = ls_opport_h
          es_output_orgman_h        = ls_orgman_h
          es_output_pricingdata_h   = ls_pricingdata_h
          es_output_sales_h         = ls_sales_h
          es_output_shipping_h      = ls_shipping_h
          et_output_partner_h       = lt_partner_h
          et_output_payplan_d_h     = lt_payplan_d_h
          es_output_customer_h      = ls_customer_h
          es_output_cumulat_h       = ls_cumulat_h
          et_output_billing_h       = lt_billing_h
          et_output_cancel_h        = lt_cancel_h
          et_output_appointment_h   = lt_appointment_h
          et_output_billplan_d_h    = lt_billplan_d_h
          et_output_billplan_h      = lt_billplan_h
          et_output_status_d_h      = lt_status_d_h
          et_output_status_h        = lt_status_h
          et_output_srv_subject_h   = lt_srv_subject_h
          et_output_srv_reason_h    = lt_srv_reason_h
          et_output_srv_result_h    = lt_srv_result_h
          es_output_acs_h           = ls_acs_h
          et_output_orderadm_i      = lt_orderadm_i
          et_output_orgman_i        = lt_orgman_i
          et_output_pricingdata_i   = lt_pricingdata_i
          et_output_pricing_i       = lt_pricing_i
          et_output_product_i       = lt_product_i
          et_output_sales_i         = lt_sales_i
          et_output_shipping_i      = lt_shipping_i
          et_output_schedlin_i      = lt_schedlin_i
          et_output_partner_i       = lt_partner_i
          et_item_cstics_i          = lt_item_cstics_i
          et_output_customer_i      = lt_customer_i
          et_output_billing_i       = lt_billing_i
          et_output_cancel_i        = lt_cancel_i
          et_output_finprod_i       = lt_finprod_i
          et_output_ordprp_i        = lt_ordprp_i
          et_output_appointment_i   = lt_appointment_i
          et_output_billplan_d_i    = lt_billplan_d_i
          et_output_billplan_i      = lt_billplan_i
          et_output_orderadm_i_qt   = lt_orderadm_i_qt
          et_output_orderadm_i_in   = lt_orderadm_i_in
          et_output_schedlin_i_cf   = lt_schedlin_i_cf
          et_output_status_i        = lt_status_i
          et_output_wrk_set_e_s_bbp = lt_working_set_e_s_bbp
          ev_language               = lv_language.
      control_parameters-langu = sy-langu.
      control_parameters-no_dialog = 'X'.
      control_parameters-getotf    = 'X'.
    call function to process smart form
      CALL FUNCTION function_name
        EXPORTING
          archive_index        = is_archive_index
          archive_index_tab    = ct_archive_index_tab
          archive_parameters   = is_archive_parameters
          control_parameters   = control_parameters
          mail_appl_obj        = is_mail_appl_obj
          mail_recipient       = is_mail_recipient
          mail_sender          = is_mail_sender
          output_options       = ls_output_options
          user_settings        = ip_user_settings
          orderadm_h           = ls_orderadm_h
          activity_h           = ls_activity_h
          lead_h               = ls_lead_h
          opport_h             = ls_opport_h
          orgman_h             = ls_orgman_h
          partner_h            = lt_partner_h
          pricingdata_h        = ls_pricingdata_h
          sales_h              = ls_sales_h
          shipping_h           = ls_shipping_h
          payplan_d_h          = lt_payplan_d_h
          cumulat_h            = ls_cumulat_h
          customer_h           = ls_customer_h
          acs_h                = ls_acs_h
          billing_h            = lt_billing_h
          cancel_h             = lt_cancel_h
          appointment_h        = lt_appointment_h
          billplan_d_h         = lt_billplan_d_h
          billplan_h           = lt_billplan_h
          status_d_h           = lt_status_d_h
          status_h             = lt_status_h
          srv_subject_h        = lt_srv_subject_h
          srv_reason_h         = lt_srv_reason_h
          srv_result_h         = lt_srv_result_h
          orderadm_i           = lt_orderadm_i
          orderadm_i_qt        = lt_orderadm_i_qt
          orderadm_i_in        = lt_orderadm_i_in
          orgman_i             = lt_orgman_i
          pricingdata_i        = lt_pricingdata_i
          pricing_i            = lt_pricing_i
          product_i            = lt_product_i
          sales_i              = lt_sales_i
          schedlin_i           = lt_schedlin_i
          schedlin_i_cf        = lt_schedlin_i_cf
          shipping_i           = lt_shipping_i
          partner_i            = lt_partner_i
          item_cstics_i        = lt_item_cstics_i
          customer_i           = lt_customer_i
          billing_i            = lt_billing_i
          cancel_i             = lt_cancel_i
          finprod_i            = lt_finprod_i
          ordprp_i             = lt_ordprp_i
          appointment_i        = lt_appointment_i
          billplan_d_i         = lt_billplan_d_i
          billplan_i           = lt_billplan_i
          status_i             = lt_status_i
          working_set_e_s_bbp  = lt_working_set_e_s_bbp
          language             = lv_language
        IMPORTING
          document_output_info = es_document_output_info
          job_output_info      = output_data
          job_output_options   = es_job_output_options
        EXCEPTIONS
          output_canceled      = 1
          parameter_error      = 2
          OTHERS               = 3.
        call function 'CONVERT_OTF'
          EXPORTING
            format                = 'PDF'
          IMPORTING
            bin_filesize          = l_pdf_len
            bin_file              = l_pdf_xstring
          TABLES
            OTF                   = OUTPUT_DATA-OTFDATA
            LINES                 = LT_LINES
          EXCEPTIONS
            err_max_linewidth     = 1
            err_format            = 2
            err_conv_not_possible = 3
            err_bad_otf           = 4
            others                = 5.
          IF SY-SUBRC = 0.
            DATA: lv_string TYPE string,
                  lv_numguid(10) TYPE N.
            lv_string = lv_spoolid.
            lv_numguid = lv_object_id.
           CALL METHOD op_previewspool( iv_spoolid = lv_string ).
            me->op_previewspool( iv_spoolid     = lv_string
                                 lv_pdf_xstring = l_pdf_xstring
                                 lv_guid        = lv_numguid ).
          ENDIF.
        endif.
        CATCH cx_sy_move_cast_error.
      ENDTRY.
    ENDMETHOD.
    <b>This part send the result to a new windows and refresh a main page, but thow a error 503 ...</b>
    method OP_PREVIEWSPOOL .
    *{   INSERT         CRDK902447                                        1
      DATA: lv_url TYPE string.
    DATA: cached_response TYPE REF TO if_http_response,
          new_navegator   TYPE REF TO CL_WB_NAVIGATOR_NEW_WINDOW,
          new_navigation   TYPE REF TO CL_BSP_NAVIGATION.
    DATA: guid TYPE guid_32.
    Resultado generado: Formato PDF
      data: l_pdf_xstring  type xstring,
            display_url    type string,
            pagina_url    type string,
            display_pag    type string,
            V_TARGET_URL   TYPE STRING,
            W_APP_NAME     TYPE STRING,
            l_pdf_len      type i.
    DATA:  out TYPE REF TO IF_BSP_WRITER,
            lv_out TYPE REF TO cl_BSP_WRITER,
            lv_context Type Ref to IF_BSP_PAGE_CONTEXT,
            lv_body type string,
            lv_body2 type string,
            lv_guidmem(10) type n,
            lv_doscmd(200),
            lv_pag   type xstring,
            lv_back(2) type n.
    IF not lv_pdf_xstring is initial.
        CREATE OBJECT cached_response TYPE CL_HTTP_RESPONSE EXPORTING add_c_msg = 1.
        l_pdf_len = xstrlen( lv_pdf_xstring ).
        cached_response->set_data( data   = lv_pdf_xstring
                              length = l_pdf_len ).
        cached_response->set_header_field( name  = if_http_header_fields=>content_type
                                           value = 'application/pdf' ).
         cached_response->set_header_field(
                            name  = 'content-disposition'
                            value = 'attachment; filename = test.pdf' ).
        cached_response->set_status( code = 200 reason = 'OK' ).
        cached_response->server_cache_expire_rel( expires_rel = 180 ).
        CALL FUNCTION 'GUID_CREATE'
          IMPORTING
            ev_guid_32 = guid.
    DATA:
        LV_HOST     TYPE STRING,     "Host name
        LV_PORT     TYPE STRING,     "Port Number
        LV_LOCAL_URL     TYPE STRING,     "URL (Relative to the Current Server)
        LV_ABS_URL     TYPE STRING,     "URL (Absolute, incl. Protocol, Host, Port)
        LV_PROTOCOL     TYPE STRING,     "Available Protocol
        LV_servidor     TYPE STRING.     "Available Protocol
    Construct the Target URL
          CALL METHOD RUNTIME->CONSTRUCT_BSP_URL
            EXPORTING
              IN_PROTOCOL       = 'http'
              IN_APPLICATION_NS = RUNTIME->APPLICATION_NAMESPACE
              IN_APPLICATION    = 'crm_ic'
              IN_PAGE           = '**/crm_ic'
            IMPORTING
              OUT_ABS_URL       = V_TARGET_URL
              OUT_HOST          = LV_HOST
              OUT_PORT          = LV_PORT
              OUT_LOCAL_URL     = LV_LOCAL_URL
              OUT_PROTOCOL      = LV_PROTOCOL.
            CONCATENATE 'crm_ic' '/**/' INTO W_APP_NAME.
          REPLACE W_APP_NAME IN V_TARGET_URL WITH SPACE.
        CONCATENATE V_TARGET_URL '/' guid '.pdf' INTO display_url.
        cl_http_server=>server_cache_upload( url      = display_url
                                             response = cached_response ).
        lv_url = adapter->if_bsp_page~get_page_url( ).
        if LV_PORT is initial.
           concatenate 'http://' LV_HOST lv_url+0(61) 'default.do' into lv_url.
        else.
           concatenate 'http://' LV_HOST ':' LV_PORT lv_url+0(61) 'default.do' into lv_url.
        endif.
          lv_body2 = '<html><head><script language=JavaScript>'.
          concatenate lv_body2 'window.open("' display_url '"); ' INTO lv_body2.
          concatenate lv_body2 ' location.href = "' lv_url '";  </script></head>' INTO lv_body2.
          concatenate lv_body2 '<body></body></html>' INTO lv_body2.
         concatenate lv_body2 ' function back(){ history.back( ); } </script></head>' INTO lv_body2.
         concatenate lv_body2 '<body><a href="javascript:back();">regresar</a></body></html>' INTO lv_body2.
          response->append_cdata( data = lv_body2 ) .
          response->set_header_field( name  = 'content-type'
                                      value = 'text/html' ). "#EC NOTEXT
          response->set_header_field( name = 'cache-control' value = 'max-age=0' ). "#EC NOTEXT
         response->set_header_field( name = 'Expires' value = '0' ). "#EC NOTEXT
          navigation->response_complete( ).
    return.
    ENDIF.
    endmethod.

    Using the search bar, it seems it's a proxy issue, but finding a simple solution did not seem apparent.  Have you tried a reboot?  To reboot, hold the power and home buttons at the same time for at least 10 seconds (hold past the swipe to shut down bar, should it appear) until the Apple Logo appears, then let go fo both buttons and your iPad will restart.  Hopefully, this will solve the problem......

  • Office 365 : SharePoint designer workflow platform not available

    Hi,
    We have Office 365 E3 subscription and I am facing issue creating any SP2013 designer workflow for the sites created via Custom Web Templates ?
    The option to create designer workflow with "SharePoint 2013 Workflow platform" is unavailable with message..
    "The option for SharePoint 2013 Workflow platform is not available because the workflow service is not configured on the server, please contact your server administrator"
    To add to the issue even if we create 2010 designer workflow, the workflow does not get associated with the list/library. I can see the workflow in SharePoint designer but not with library settings from the browser and of course the workflow is not executed.
    Can anyone please suggest any solution ?
    Thanks in advance.
    Rajesh

    Hi,
    If you are a tenant administrator please log into your admin account
    navigate to this  "https://youraccount-admin.sharepoint.com/_layouts/15/online/TenantSettings.aspx"
    In this page,under connected services, deselect the "Block
    SharePoint 2013 workflows".
    Murugesa Pandian.,SharePoint 2010, MCPD | MCTS - Configure

  • LVM Volumes not available after update

    Hi All!
    I haven't updated my system for about two months and today I updated it. Now I have the problem that I cannot boot properly. I have my root partition in an LVM volume and on boot I get the message
    ERROR: device 'UUID=xxx' not found. Skipping fs
    ERROR: Unable to find root device 'UUID=xxx'
    After that I land in the recovery shell. After some research I found, that "lvm lvdisplay" showed that my volumes where not available and I had to reenable them with "lvm vgchange -a y".
    Issuing any lvm command also produced the following warning:
    WARNING: lvmetad is running but disabled. Restart lvmetad before enabling it!
    Anyway, after issuing the commands and exiting the recovery shell, the system booted again. However, I would prefer being able to boot without manual actions.
    Thanks in advance!
    Further information:
    vgdisplay
    --- Volume group ---
    VG Name ArchLVM
    System ID
    Format lvm2
    Metadata Areas 1
    Metadata Sequence No 3
    VG Access read/write
    VG Status resizable
    MAX LV 0
    Cur LV 2
    Open LV 1
    Max PV 0
    Cur PV 1
    Act PV 1
    VG Size 232.69 GiB
    PE Size 4.00 MiB
    Total PE 59568
    Alloc PE / Size 59568 / 232.69 GiB
    Free PE / Size 0 / 0
    VG UUID SoB3M1-v1fD-1abI-PNJ3-6IOn-FfdI-0RoLK5
    lvdisplay (LV Status was 'not available' right after booting)
    --- Logical volume ---
    LV Path /dev/ArchLVM/Swap
    LV Name Swap
    VG Name ArchLVM
    LV UUID XRYBrz-LojR-k6SD-XIxV-wHnY-f3VG-giKL6V
    LV Write Access read/write
    LV Creation host, time archiso, 2014-05-16 14:43:06 +0200
    LV Status available
    # open 0
    LV Size 8.00 GiB
    Current LE 2048
    Segments 1
    Allocation inherit
    Read ahead sectors auto
    - currently set to 256
    Block device 254:0
    --- Logical volume ---
    LV Path /dev/ArchLVM/Root
    LV Name Root
    VG Name ArchLVM
    LV UUID lpjDl4-Jqzu-ZWkq-Uphc-IaOo-6Rzd-cIh5yv
    LV Write Access read/write
    LV Creation host, time archiso, 2014-05-16 14:43:27 +0200
    LV Status available
    # open 1
    LV Size 224.69 GiB
    Current LE 57520
    Segments 1
    Allocation inherit
    Read ahead sectors auto
    - currently set to 256
    Block device 254:1
    /etc/fstab
    # /etc/fstab: static file system information
    # <file system> <dir> <type> <options> <dump> <pass>
    # /dev/mapper/ArchLVM-Root
    UUID=2db82d1a-47a4-4e30-a819-143e8fb75199 / ext4 rw,relatime,data=ordered 0 1
    #/dev/mapper/ArchLVM-Root / ext4 rw,relatime,data=ordered 0 1
    # /dev/sda1
    UUID=72691888-a781-4cdd-a98e-2613d87925d0 /boot ext2 rw,relatime 0 2
    /etc/mkinitcpio.conf
    # vim:set ft=sh
    # MODULES
    # The following modules are loaded before any boot hooks are
    # run. Advanced users may wish to specify all system modules
    # in this array. For instance:
    # MODULES="piix ide_disk reiserfs"
    MODULES=""
    # BINARIES
    # This setting includes any additional binaries a given user may
    # wish into the CPIO image. This is run last, so it may be used to
    # override the actual binaries included by a given hook
    # BINARIES are dependency parsed, so you may safely ignore libraries
    BINARIES=""
    # FILES
    # This setting is similar to BINARIES above, however, files are added
    # as-is and are not parsed in any way. This is useful for config files.
    FILES=""
    # HOOKS
    # This is the most important setting in this file. The HOOKS control the
    # modules and scripts added to the image, and what happens at boot time.
    # Order is important, and it is recommended that you do not change the
    # order in which HOOKS are added. Run 'mkinitcpio -H <hook name>' for
    # help on a given hook.
    # 'base' is _required_ unless you know precisely what you are doing.
    # 'udev' is _required_ in order to automatically load modules
    # 'filesystems' is _required_ unless you specify your fs modules in MODULES
    # Examples:
    ## This setup specifies all modules in the MODULES setting above.
    ## No raid, lvm2, or encrypted root is needed.
    # HOOKS="base"
    ## This setup will autodetect all modules for your system and should
    ## work as a sane default
    # HOOKS="base udev autodetect block filesystems"
    ## This setup will generate a 'full' image which supports most systems.
    ## No autodetection is done.
    # HOOKS="base udev block filesystems"
    ## This setup assembles a pata mdadm array with an encrypted root FS.
    ## Note: See 'mkinitcpio -H mdadm' for more information on raid devices.
    # HOOKS="base udev block mdadm encrypt filesystems"
    ## This setup loads an lvm2 volume group on a usb device.
    # HOOKS="base udev block lvm2 filesystems"
    ## NOTE: If you have /usr on a separate partition, you MUST include the
    # usr, fsck and shutdown hooks.
    HOOKS="base udev autodetect modconf block lvm2 filesystems keyboard fsck"
    # COMPRESSION
    # Use this to compress the initramfs image. By default, gzip compression
    # is used. Use 'cat' to create an uncompressed image.
    #COMPRESSION="gzip"
    #COMPRESSION="bzip2"
    #COMPRESSION="lzma"
    #COMPRESSION="xz"
    #COMPRESSION="lzop"
    #COMPRESSION="lz4"
    # COMPRESSION_OPTIONS
    # Additional options for the compressor
    #COMPRESSION_OPTIONS=""
    /boot/grub/grub.cfg
    # DO NOT EDIT THIS FILE
    # It is automatically generated by grub-mkconfig using templates
    # from /etc/grub.d and settings from /etc/default/grub
    ### BEGIN /etc/grub.d/00_header ###
    insmod part_gpt
    insmod part_msdos
    if [ -s $prefix/grubenv ]; then
    load_env
    fi
    if [ "${next_entry}" ] ; then
    set default="${next_entry}"
    set next_entry=
    save_env next_entry
    set boot_once=true
    else
    set default="0"
    fi
    if [ x"${feature_menuentry_id}" = xy ]; then
    menuentry_id_option="--id"
    else
    menuentry_id_option=""
    fi
    export menuentry_id_option
    if [ "${prev_saved_entry}" ]; then
    set saved_entry="${prev_saved_entry}"
    save_env saved_entry
    set prev_saved_entry=
    save_env prev_saved_entry
    set boot_once=true
    fi
    function savedefault {
    if [ -z "${boot_once}" ]; then
    saved_entry="${chosen}"
    save_env saved_entry
    fi
    function load_video {
    if [ x$feature_all_video_module = xy ]; then
    insmod all_video
    else
    insmod efi_gop
    insmod efi_uga
    insmod ieee1275_fb
    insmod vbe
    insmod vga
    insmod video_bochs
    insmod video_cirrus
    fi
    if [ x$feature_default_font_path = xy ] ; then
    font=unicode
    else
    insmod part_msdos
    insmod lvm
    insmod ext2
    set root='lvmid/SoB3M1-v1fD-1abI-PNJ3-6IOn-FfdI-0RoLK5/lpjDl4-Jqzu-ZWkq-Uphc-IaOo-6Rzd-cIh5yv'
    if [ x$feature_platform_search_hint = xy ]; then
    search --no-floppy --fs-uuid --set=root --hint='lvmid/SoB3M1-v1fD-1abI-PNJ3-6IOn-FfdI-0RoLK5/lpjDl4-Jqzu-ZWkq-Uphc-IaOo-6Rzd-cIh5yv' 2db82d1a-47a4-4e30-a819-143e8fb75199
    else
    search --no-floppy --fs-uuid --set=root 2db82d1a-47a4-4e30-a819-143e8fb75199
    fi
    font="/usr/share/grub/unicode.pf2"
    fi
    if loadfont $font ; then
    set gfxmode=auto
    load_video
    insmod gfxterm
    fi
    terminal_input console
    terminal_output gfxterm
    if [ x$feature_timeout_style = xy ] ; then
    set timeout_style=menu
    set timeout=5
    # Fallback normal timeout code in case the timeout_style feature is
    # unavailable.
    else
    set timeout=5
    fi
    ### END /etc/grub.d/00_header ###
    ### BEGIN /etc/grub.d/10_linux ###
    menuentry 'Arch Linux' --class arch --class gnu-linux --class gnu --class os $menuentry_id_option 'gnulinux-simple-2db82d1a-47a4-4e30-a819-143e8fb75199' {
    load_video
    set gfxpayload=keep
    insmod gzio
    insmod part_msdos
    insmod ext2
    set root='hd0,msdos1'
    if [ x$feature_platform_search_hint = xy ]; then
    search --no-floppy --fs-uuid --set=root --hint-bios=hd0,msdos1 --hint-efi=hd0,msdos1 --hint-baremetal=ahci0,msdos1 72691888-a781-4cdd-a98e-2613d87925d0
    else
    search --no-floppy --fs-uuid --set=root 72691888-a781-4cdd-a98e-2613d87925d0
    fi
    echo 'Loading Linux linux ...'
    linux /vmlinuz-linux root=UUID=2db82d1a-47a4-4e30-a819-143e8fb75199 rw quiet
    echo 'Loading initial ramdisk ...'
    initrd /initramfs-linux.img
    submenu 'Advanced options for Arch Linux' $menuentry_id_option 'gnulinux-advanced-2db82d1a-47a4-4e30-a819-143e8fb75199' {
    menuentry 'Arch Linux, with Linux linux' --class arch --class gnu-linux --class gnu --class os $menuentry_id_option 'gnulinux-linux-advanced-2db82d1a-47a4-4e30-a819-143e8fb75199' {
    load_video
    set gfxpayload=keep
    insmod gzio
    insmod part_msdos
    insmod ext2
    set root='hd0,msdos1'
    if [ x$feature_platform_search_hint = xy ]; then
    search --no-floppy --fs-uuid --set=root --hint-bios=hd0,msdos1 --hint-efi=hd0,msdos1 --hint-baremetal=ahci0,msdos1 72691888-a781-4cdd-a98e-2613d87925d0
    else
    search --no-floppy --fs-uuid --set=root 72691888-a781-4cdd-a98e-2613d87925d0
    fi
    echo 'Loading Linux linux ...'
    linux /vmlinuz-linux root=UUID=2db82d1a-47a4-4e30-a819-143e8fb75199 rw quiet
    echo 'Loading initial ramdisk ...'
    initrd /initramfs-linux.img
    menuentry 'Arch Linux, with Linux linux (fallback initramfs)' --class arch --class gnu-linux --class gnu --class os $menuentry_id_option 'gnulinux-linux-fallback-2db82d1a-47a4-4e30-a819-143e8fb75199' {
    load_video
    set gfxpayload=keep
    insmod gzio
    insmod part_msdos
    insmod ext2
    set root='hd0,msdos1'
    if [ x$feature_platform_search_hint = xy ]; then
    search --no-floppy --fs-uuid --set=root --hint-bios=hd0,msdos1 --hint-efi=hd0,msdos1 --hint-baremetal=ahci0,msdos1 72691888-a781-4cdd-a98e-2613d87925d0
    else
    search --no-floppy --fs-uuid --set=root 72691888-a781-4cdd-a98e-2613d87925d0
    fi
    echo 'Loading Linux linux ...'
    linux /vmlinuz-linux root=UUID=2db82d1a-47a4-4e30-a819-143e8fb75199 rw quiet
    echo 'Loading initial ramdisk ...'
    initrd /initramfs-linux-fallback.img
    ### END /etc/grub.d/10_linux ###
    ### BEGIN /etc/grub.d/20_linux_xen ###
    ### END /etc/grub.d/20_linux_xen ###
    ### BEGIN /etc/grub.d/30_os-prober ###
    ### END /etc/grub.d/30_os-prober ###
    ### BEGIN /etc/grub.d/40_custom ###
    # This file provides an easy way to add custom menu entries. Simply type the
    # menu entries you want to add after this comment. Be careful not to change
    # the 'exec tail' line above.
    ### END /etc/grub.d/40_custom ###
    ### BEGIN /etc/grub.d/41_custom ###
    if [ -f ${config_directory}/custom.cfg ]; then
    source ${config_directory}/custom.cfg
    elif [ -z "${config_directory}" -a -f $prefix/custom.cfg ]; then
    source $prefix/custom.cfg;
    fi
    ### END /etc/grub.d/41_custom ###
    ### BEGIN /etc/grub.d/60_memtest86+ ###
    ### END /etc/grub.d/60_memtest86+ ###
    Last edited by Kirodema (2014-07-16 07:31:34)

    use_lvmetad = 0
    lvm2-lvmetad is not enabled or running on my system. Shall I activate it?
    # This is an example configuration file for the LVM2 system.
    # It contains the default settings that would be used if there was no
    # /etc/lvm/lvm.conf file.
    # Refer to 'man lvm.conf' for further information including the file layout.
    # To put this file in a different directory and override /etc/lvm set
    # the environment variable LVM_SYSTEM_DIR before running the tools.
    # N.B. Take care that each setting only appears once if uncommenting
    # example settings in this file.
    # This section allows you to set the way the configuration settings are handled.
    config {
    # If enabled, any LVM2 configuration mismatch is reported.
    # This implies checking that the configuration key is understood
    # by LVM2 and that the value of the key is of a proper type.
    # If disabled, any configuration mismatch is ignored and default
    # value is used instead without any warning (a message about the
    # configuration key not being found is issued in verbose mode only).
    checks = 1
    # If enabled, any configuration mismatch aborts the LVM2 process.
    abort_on_errors = 0
    # Directory where LVM looks for configuration profiles.
    profile_dir = "/etc/lvm/profile"
    # This section allows you to configure which block devices should
    # be used by the LVM system.
    devices {
    # Where do you want your volume groups to appear ?
    dir = "/dev"
    # An array of directories that contain the device nodes you wish
    # to use with LVM2.
    scan = [ "/dev" ]
    # If set, the cache of block device nodes with all associated symlinks
    # will be constructed out of the existing udev database content.
    # This avoids using and opening any inapplicable non-block devices or
    # subdirectories found in the device directory. This setting is applied
    # to udev-managed device directory only, other directories will be scanned
    # fully. LVM2 needs to be compiled with udev support for this setting to
    # take effect. N.B. Any device node or symlink not managed by udev in
    # udev directory will be ignored with this setting on.
    obtain_device_list_from_udev = 1
    # If several entries in the scanned directories correspond to the
    # same block device and the tools need to display a name for device,
    # all the pathnames are matched against each item in the following
    # list of regular expressions in turn and the first match is used.
    preferred_names = [ ]
    # Try to avoid using undescriptive /dev/dm-N names, if present.
    # preferred_names = [ "^/dev/mpath/", "^/dev/mapper/mpath", "^/dev/[hs]d" ]
    # A filter that tells LVM2 to only use a restricted set of devices.
    # The filter consists of an array of regular expressions. These
    # expressions can be delimited by a character of your choice, and
    # prefixed with either an 'a' (for accept) or 'r' (for reject).
    # The first expression found to match a device name determines if
    # the device will be accepted or rejected (ignored). Devices that
    # don't match any patterns are accepted.
    # Be careful if there there are symbolic links or multiple filesystem
    # entries for the same device as each name is checked separately against
    # the list of patterns. The effect is that if the first pattern in the
    # list to match a name is an 'a' pattern for any of the names, the device
    # is accepted; otherwise if the first pattern in the list to match a name
    # is an 'r' pattern for any of the names it is rejected; otherwise it is
    # accepted.
    # Don't have more than one filter line active at once: only one gets used.
    # Run vgscan after you change this parameter to ensure that
    # the cache file gets regenerated (see below).
    # If it doesn't do what you expect, check the output of 'vgscan -vvvv'.
    # If lvmetad is used, then see "A note about device filtering while
    # lvmetad is used" comment that is attached to global/use_lvmetad setting.
    # By default we accept every block device:
    filter = [ "a/.*/" ]
    # Exclude the cdrom drive
    # filter = [ "r|/dev/cdrom|" ]
    # When testing I like to work with just loopback devices:
    # filter = [ "a/loop/", "r/.*/" ]
    # Or maybe all loops and ide drives except hdc:
    # filter =[ "a|loop|", "r|/dev/hdc|", "a|/dev/ide|", "r|.*|" ]
    # Use anchors if you want to be really specific
    # filter = [ "a|^/dev/hda8$|", "r/.*/" ]
    # Since "filter" is often overridden from command line, it is not suitable
    # for system-wide device filtering (udev rules, lvmetad). To hide devices
    # from LVM-specific udev processing and/or from lvmetad, you need to set
    # global_filter. The syntax is the same as for normal "filter"
    # above. Devices that fail the global_filter are not even opened by LVM.
    # global_filter = []
    # The results of the filtering are cached on disk to avoid
    # rescanning dud devices (which can take a very long time).
    # By default this cache is stored in the /etc/lvm/cache directory
    # in a file called '.cache'.
    # It is safe to delete the contents: the tools regenerate it.
    # (The old setting 'cache' is still respected if neither of
    # these new ones is present.)
    # N.B. If obtain_device_list_from_udev is set to 1 the list of
    # devices is instead obtained from udev and any existing .cache
    # file is removed.
    cache_dir = "/etc/lvm/cache"
    cache_file_prefix = ""
    # You can turn off writing this cache file by setting this to 0.
    write_cache_state = 1
    # Advanced settings.
    # List of pairs of additional acceptable block device types found
    # in /proc/devices with maximum (non-zero) number of partitions.
    # types = [ "fd", 16 ]
    # If sysfs is mounted (2.6 kernels) restrict device scanning to
    # the block devices it believes are valid.
    # 1 enables; 0 disables.
    sysfs_scan = 1
    # By default, LVM2 will ignore devices used as component paths
    # of device-mapper multipath devices.
    # 1 enables; 0 disables.
    multipath_component_detection = 1
    # By default, LVM2 will ignore devices used as components of
    # software RAID (md) devices by looking for md superblocks.
    # 1 enables; 0 disables.
    md_component_detection = 1
    # By default, if a PV is placed directly upon an md device, LVM2
    # will align its data blocks with the md device's stripe-width.
    # 1 enables; 0 disables.
    md_chunk_alignment = 1
    # Default alignment of the start of a data area in MB. If set to 0,
    # a value of 64KB will be used. Set to 1 for 1MiB, 2 for 2MiB, etc.
    # default_data_alignment = 1
    # By default, the start of a PV's data area will be a multiple of
    # the 'minimum_io_size' or 'optimal_io_size' exposed in sysfs.
    # - minimum_io_size - the smallest request the device can perform
    # w/o incurring a read-modify-write penalty (e.g. MD's chunk size)
    # - optimal_io_size - the device's preferred unit of receiving I/O
    # (e.g. MD's stripe width)
    # minimum_io_size is used if optimal_io_size is undefined (0).
    # If md_chunk_alignment is enabled, that detects the optimal_io_size.
    # This setting takes precedence over md_chunk_alignment.
    # 1 enables; 0 disables.
    data_alignment_detection = 1
    # Alignment (in KB) of start of data area when creating a new PV.
    # md_chunk_alignment and data_alignment_detection are disabled if set.
    # Set to 0 for the default alignment (see: data_alignment_default)
    # or page size, if larger.
    data_alignment = 0
    # By default, the start of the PV's aligned data area will be shifted by
    # the 'alignment_offset' exposed in sysfs. This offset is often 0 but
    # may be non-zero; e.g.: certain 4KB sector drives that compensate for
    # windows partitioning will have an alignment_offset of 3584 bytes
    # (sector 7 is the lowest aligned logical block, the 4KB sectors start
    # at LBA -1, and consequently sector 63 is aligned on a 4KB boundary).
    # But note that pvcreate --dataalignmentoffset will skip this detection.
    # 1 enables; 0 disables.
    data_alignment_offset_detection = 1
    # If, while scanning the system for PVs, LVM2 encounters a device-mapper
    # device that has its I/O suspended, it waits for it to become accessible.
    # Set this to 1 to skip such devices. This should only be needed
    # in recovery situations.
    ignore_suspended_devices = 0
    # ignore_lvm_mirrors: Introduced in version 2.02.104
    # This setting determines whether logical volumes of "mirror" segment
    # type are scanned for LVM labels. This affects the ability of
    # mirrors to be used as physical volumes. If 'ignore_lvm_mirrors'
    # is set to '1', it becomes impossible to create volume groups on top
    # of mirror logical volumes - i.e. to stack volume groups on mirrors.
    # Allowing mirror logical volumes to be scanned (setting the value to '0')
    # can potentially cause LVM processes and I/O to the mirror to become
    # blocked. This is due to the way that the "mirror" segment type handles
    # failures. In order for the hang to manifest itself, an LVM command must
    # be run just after a failure and before the automatic LVM repair process
    # takes place OR there must be failures in multiple mirrors in the same
    # volume group at the same time with write failures occurring moments
    # before a scan of the mirror's labels.
    # Note that these scanning limitations do not apply to the LVM RAID
    # types, like "raid1". The RAID segment types handle failures in a
    # different way and are not subject to possible process or I/O blocking.
    # It is encouraged that users set 'ignore_lvm_mirrors' to 1 if they
    # are using the "mirror" segment type. Users that require volume group
    # stacking on mirrored logical volumes should consider using the "raid1"
    # segment type. The "raid1" segment type is not available for
    # active/active clustered volume groups.
    # Set to 1 to disallow stacking and thereby avoid a possible deadlock.
    ignore_lvm_mirrors = 1
    # During each LVM operation errors received from each device are counted.
    # If the counter of a particular device exceeds the limit set here, no
    # further I/O is sent to that device for the remainder of the respective
    # operation. Setting the parameter to 0 disables the counters altogether.
    disable_after_error_count = 0
    # Allow use of pvcreate --uuid without requiring --restorefile.
    require_restorefile_with_uuid = 1
    # Minimum size (in KB) of block devices which can be used as PVs.
    # In a clustered environment all nodes must use the same value.
    # Any value smaller than 512KB is ignored.
    # Ignore devices smaller than 2MB such as floppy drives.
    pv_min_size = 2048
    # The original built-in setting was 512 up to and including version 2.02.84.
    # pv_min_size = 512
    # Issue discards to a logical volumes's underlying physical volume(s) when
    # the logical volume is no longer using the physical volumes' space (e.g.
    # lvremove, lvreduce, etc). Discards inform the storage that a region is
    # no longer in use. Storage that supports discards advertise the protocol
    # specific way discards should be issued by the kernel (TRIM, UNMAP, or
    # WRITE SAME with UNMAP bit set). Not all storage will support or benefit
    # from discards but SSDs and thinly provisioned LUNs generally do. If set
    # to 1, discards will only be issued if both the storage and kernel provide
    # support.
    # 1 enables; 0 disables.
    issue_discards = 0
    # This section allows you to configure the way in which LVM selects
    # free space for its Logical Volumes.
    allocation {
    # When searching for free space to extend an LV, the "cling"
    # allocation policy will choose space on the same PVs as the last
    # segment of the existing LV. If there is insufficient space and a
    # list of tags is defined here, it will check whether any of them are
    # attached to the PVs concerned and then seek to match those PV tags
    # between existing extents and new extents.
    # Use the special tag "@*" as a wildcard to match any PV tag.
    # Example: LVs are mirrored between two sites within a single VG.
    # PVs are tagged with either @site1 or @site2 to indicate where
    # they are situated.
    # cling_tag_list = [ "@site1", "@site2" ]
    # cling_tag_list = [ "@*" ]
    # Changes made in version 2.02.85 extended the reach of the 'cling'
    # policies to detect more situations where data can be grouped
    # onto the same disks. Set this to 0 to revert to the previous
    # algorithm.
    maximise_cling = 1
    # Whether to use blkid library instead of native LVM2 code to detect
    # any existing signatures while creating new Physical Volumes and
    # Logical Volumes. LVM2 needs to be compiled with blkid wiping support
    # for this setting to take effect.
    # LVM2 native detection code is currently able to recognize these signatures:
    # - MD device signature
    # - swap signature
    # - LUKS signature
    # To see the list of signatures recognized by blkid, check the output
    # of 'blkid -k' command. The blkid can recognize more signatures than
    # LVM2 native detection code, but due to this higher number of signatures
    # to be recognized, it can take more time to complete the signature scan.
    use_blkid_wiping = 1
    # Set to 1 to wipe any signatures found on newly-created Logical Volumes
    # automatically in addition to zeroing of the first KB on the LV
    # (controlled by the -Z/--zero y option).
    # The command line option -W/--wipesignatures takes precedence over this
    # setting.
    # The default is to wipe signatures when zeroing.
    wipe_signatures_when_zeroing_new_lvs = 1
    # Set to 1 to guarantee that mirror logs will always be placed on
    # different PVs from the mirror images. This was the default
    # until version 2.02.85.
    mirror_logs_require_separate_pvs = 0
    # Set to 1 to guarantee that cache_pool metadata will always be
    # placed on different PVs from the cache_pool data.
    cache_pool_metadata_require_separate_pvs = 0
    # Specify the minimal chunk size (in kiB) for cache pool volumes.
    # Using a chunk_size that is too large can result in wasteful use of
    # the cache, where small reads and writes can cause large sections of
    # an LV to be mapped into the cache. However, choosing a chunk_size
    # that is too small can result in more overhead trying to manage the
    # numerous chunks that become mapped into the cache. The former is
    # more of a problem than the latter in most cases, so we default to
    # a value that is on the smaller end of the spectrum. Supported values
    # range from 32(kiB) to 1048576 in multiples of 32.
    # cache_pool_chunk_size = 64
    # Set to 1 to guarantee that thin pool metadata will always
    # be placed on different PVs from the pool data.
    thin_pool_metadata_require_separate_pvs = 0
    # Specify chunk size calculation policy for thin pool volumes.
    # Possible options are:
    # "generic" - if thin_pool_chunk_size is defined, use it.
    # Otherwise, calculate the chunk size based on
    # estimation and device hints exposed in sysfs:
    # the minimum_io_size. The chunk size is always
    # at least 64KiB.
    # "performance" - if thin_pool_chunk_size is defined, use it.
    # Otherwise, calculate the chunk size for
    # performance based on device hints exposed in
    # sysfs: the optimal_io_size. The chunk size is
    # always at least 512KiB.
    # thin_pool_chunk_size_policy = "generic"
    # Specify the minimal chunk size (in KB) for thin pool volumes.
    # Use of the larger chunk size may improve performance for plain
    # thin volumes, however using them for snapshot volumes is less efficient,
    # as it consumes more space and takes extra time for copying.
    # When unset, lvm tries to estimate chunk size starting from 64KB
    # Supported values are in range from 64 to 1048576.
    # thin_pool_chunk_size = 64
    # Specify discards behaviour of the thin pool volume.
    # Select one of "ignore", "nopassdown", "passdown"
    # thin_pool_discards = "passdown"
    # Set to 0, to disable zeroing of thin pool data chunks before their
    # first use.
    # N.B. zeroing larger thin pool chunk size degrades performance.
    # thin_pool_zero = 1
    # This section that allows you to configure the nature of the
    # information that LVM2 reports.
    log {
    # Controls the messages sent to stdout or stderr.
    # There are three levels of verbosity, 3 being the most verbose.
    verbose = 0
    # Set to 1 to suppress all non-essential messages from stdout.
    # This has the same effect as -qq.
    # When this is set, the following commands still produce output:
    # dumpconfig, lvdisplay, lvmdiskscan, lvs, pvck, pvdisplay,
    # pvs, version, vgcfgrestore -l, vgdisplay, vgs.
    # Non-essential messages are shifted from log level 4 to log level 5
    # for syslog and lvm2_log_fn purposes.
    # Any 'yes' or 'no' questions not overridden by other arguments
    # are suppressed and default to 'no'.
    silent = 0
    # Should we send log messages through syslog?
    # 1 is yes; 0 is no.
    syslog = 1
    # Should we log error and debug messages to a file?
    # By default there is no log file.
    #file = "/var/log/lvm2.log"
    # Should we overwrite the log file each time the program is run?
    # By default we append.
    overwrite = 0
    # What level of log messages should we send to the log file and/or syslog?
    # There are 6 syslog-like log levels currently in use - 2 to 7 inclusive.
    # 7 is the most verbose (LOG_DEBUG).
    level = 0
    # Format of output messages
    # Whether or not (1 or 0) to indent messages according to their severity
    indent = 1
    # Whether or not (1 or 0) to display the command name on each line output
    command_names = 0
    # A prefix to use before the message text (but after the command name,
    # if selected). Default is two spaces, so you can see/grep the severity
    # of each message.
    prefix = " "
    # To make the messages look similar to the original LVM tools use:
    # indent = 0
    # command_names = 1
    # prefix = " -- "
    # Set this if you want log messages during activation.
    # Don't use this in low memory situations (can deadlock).
    # activation = 0
    # Some debugging messages are assigned to a class and only appear
    # in debug output if the class is listed here.
    # Classes currently available:
    # memory, devices, activation, allocation, lvmetad, metadata, cache,
    # locking
    # Use "all" to see everything.
    debug_classes = [ "memory", "devices", "activation", "allocation",
    "lvmetad", "metadata", "cache", "locking" ]
    # Configuration of metadata backups and archiving. In LVM2 when we
    # talk about a 'backup' we mean making a copy of the metadata for the
    # *current* system. The 'archive' contains old metadata configurations.
    # Backups are stored in a human readable text format.
    backup {
    # Should we maintain a backup of the current metadata configuration ?
    # Use 1 for Yes; 0 for No.
    # Think very hard before turning this off!
    backup = 1
    # Where shall we keep it ?
    # Remember to back up this directory regularly!
    backup_dir = "/etc/lvm/backup"
    # Should we maintain an archive of old metadata configurations.
    # Use 1 for Yes; 0 for No.
    # On by default. Think very hard before turning this off.
    archive = 1
    # Where should archived files go ?
    # Remember to back up this directory regularly!
    archive_dir = "/etc/lvm/archive"
    # What is the minimum number of archive files you wish to keep ?
    retain_min = 10
    # What is the minimum time you wish to keep an archive file for ?
    retain_days = 30
    # Settings for the running LVM2 in shell (readline) mode.
    shell {
    # Number of lines of history to store in ~/.lvm_history
    history_size = 100
    # Miscellaneous global LVM2 settings
    global {
    # The file creation mask for any files and directories created.
    # Interpreted as octal if the first digit is zero.
    umask = 077
    # Allow other users to read the files
    #umask = 022
    # Enabling test mode means that no changes to the on disk metadata
    # will be made. Equivalent to having the -t option on every
    # command. Defaults to off.
    test = 0
    # Default value for --units argument
    units = "h"
    # Since version 2.02.54, the tools distinguish between powers of
    # 1024 bytes (e.g. KiB, MiB, GiB) and powers of 1000 bytes (e.g.
    # KB, MB, GB).
    # If you have scripts that depend on the old behaviour, set this to 0
    # temporarily until you update them.
    si_unit_consistency = 1
    # Whether or not to display unit suffix for sizes. This setting has
    # no effect if the units are in human-readable form (global/units="h")
    # in which case the suffix is always displayed.
    suffix = 1
    # Whether or not to communicate with the kernel device-mapper.
    # Set to 0 if you want to use the tools to manipulate LVM metadata
    # without activating any logical volumes.
    # If the device-mapper kernel driver is not present in your kernel
    # setting this to 0 should suppress the error messages.
    activation = 1
    # If we can't communicate with device-mapper, should we try running
    # the LVM1 tools?
    # This option only applies to 2.4 kernels and is provided to help you
    # switch between device-mapper kernels and LVM1 kernels.
    # The LVM1 tools need to be installed with .lvm1 suffices
    # e.g. vgscan.lvm1 and they will stop working after you start using
    # the new lvm2 on-disk metadata format.
    # The default value is set when the tools are built.
    # fallback_to_lvm1 = 0
    # The default metadata format that commands should use - "lvm1" or "lvm2".
    # The command line override is -M1 or -M2.
    # Defaults to "lvm2".
    # format = "lvm2"
    # Location of proc filesystem
    proc = "/proc"
    # Type of locking to use. Defaults to local file-based locking (1).
    # Turn locking off by setting to 0 (dangerous: risks metadata corruption
    # if LVM2 commands get run concurrently).
    # Type 2 uses the external shared library locking_library.
    # Type 3 uses built-in clustered locking.
    # Type 4 uses read-only locking which forbids any operations that might
    # change metadata.
    # N.B. Don't use lvmetad with locking type 3 as lvmetad is not yet
    # supported in clustered environment. If use_lvmetad=1 and locking_type=3
    # is set at the same time, LVM always issues a warning message about this
    # and then it automatically disables lvmetad use.
    locking_type = 1
    # Set to 0 to fail when a lock request cannot be satisfied immediately.
    wait_for_locks = 1
    # If using external locking (type 2) and initialisation fails,
    # with this set to 1 an attempt will be made to use the built-in
    # clustered locking.
    # If you are using a customised locking_library you should set this to 0.
    fallback_to_clustered_locking = 1
    # If an attempt to initialise type 2 or type 3 locking failed, perhaps
    # because cluster components such as clvmd are not running, with this set
    # to 1 an attempt will be made to use local file-based locking (type 1).
    # If this succeeds, only commands against local volume groups will proceed.
    # Volume Groups marked as clustered will be ignored.
    fallback_to_local_locking = 1
    # Local non-LV directory that holds file-based locks while commands are
    # in progress. A directory like /tmp that may get wiped on reboot is OK.
    locking_dir = "/run/lock/lvm"
    # Whenever there are competing read-only and read-write access requests for
    # a volume group's metadata, instead of always granting the read-only
    # requests immediately, delay them to allow the read-write requests to be
    # serviced. Without this setting, write access may be stalled by a high
    # volume of read-only requests.
    # NB. This option only affects locking_type = 1 viz. local file-based
    # locking.
    prioritise_write_locks = 1
    # Other entries can go here to allow you to load shared libraries
    # e.g. if support for LVM1 metadata was compiled as a shared library use
    # format_libraries = "liblvm2format1.so"
    # Full pathnames can be given.
    # Search this directory first for shared libraries.
    # library_dir = "/lib"
    # The external locking library to load if locking_type is set to 2.
    # locking_library = "liblvm2clusterlock.so"
    # Treat any internal errors as fatal errors, aborting the process that
    # encountered the internal error. Please only enable for debugging.
    abort_on_internal_errors = 0
    # Check whether CRC is matching when parsed VG is used multiple times.
    # This is useful to catch unexpected internal cached volume group
    # structure modification. Please only enable for debugging.
    detect_internal_vg_cache_corruption = 0
    # If set to 1, no operations that change on-disk metadata will be permitted.
    # Additionally, read-only commands that encounter metadata in need of repair
    # will still be allowed to proceed exactly as if the repair had been
    # performed (except for the unchanged vg_seqno).
    # Inappropriate use could mess up your system, so seek advice first!
    metadata_read_only = 0
    # 'mirror_segtype_default' defines which segtype will be used when the
    # shorthand '-m' option is used for mirroring. The possible options are:
    # "mirror" - The original RAID1 implementation provided by LVM2/DM. It is
    # characterized by a flexible log solution (core, disk, mirrored)
    # and by the necessity to block I/O while reconfiguring in the
    # event of a failure.
    # There is an inherent race in the dmeventd failure handling
    # logic with snapshots of devices using this type of RAID1 that
    # in the worst case could cause a deadlock.
    # Ref: https://bugzilla.redhat.com/show_bug.cgi?id=817130#c10
    # "raid1" - This implementation leverages MD's RAID1 personality through
    # device-mapper. It is characterized by a lack of log options.
    # (A log is always allocated for every device and they are placed
    # on the same device as the image - no separate devices are
    # required.) This mirror implementation does not require I/O
    # to be blocked in the kernel in the event of a failure.
    # This mirror implementation is not cluster-aware and cannot be
    # used in a shared (active/active) fashion in a cluster.
    # Specify the '--type <mirror|raid1>' option to override this default
    # setting.
    mirror_segtype_default = "raid1"
    # 'raid10_segtype_default' determines the segment types used by default
    # when the '--stripes/-i' and '--mirrors/-m' arguments are both specified
    # during the creation of a logical volume.
    # Possible settings include:
    # "raid10" - This implementation leverages MD's RAID10 personality through
    # device-mapper.
    # "mirror" - LVM will layer the 'mirror' and 'stripe' segment types. It
    # will do this by creating a mirror on top of striped sub-LVs;
    # effectively creating a RAID 0+1 array. This is suboptimal
    # in terms of providing redundancy and performance. Changing to
    # this setting is not advised.
    # Specify the '--type <raid10|mirror>' option to override this default
    # setting.
    raid10_segtype_default = "raid10"
    # The default format for displaying LV names in lvdisplay was changed
    # in version 2.02.89 to show the LV name and path separately.
    # Previously this was always shown as /dev/vgname/lvname even when that
    # was never a valid path in the /dev filesystem.
    # Set to 1 to reinstate the previous format.
    # lvdisplay_shows_full_device_path = 0
    # Whether to use (trust) a running instance of lvmetad. If this is set to
    # 0, all commands fall back to the usual scanning mechanisms. When set to 1
    # *and* when lvmetad is running (automatically instantiated by making use of
    # systemd's socket-based service activation or run as an initscripts service
    # or run manually), the volume group metadata and PV state flags are obtained
    # from the lvmetad instance and no scanning is done by the individual
    # commands. In a setup with lvmetad, lvmetad udev rules *must* be set up for
    # LVM to work correctly. Without proper udev rules, all changes in block
    # device configuration will be *ignored* until a manual 'pvscan --cache'
    # is performed. These rules are installed by default.
    # If lvmetad has been running while use_lvmetad was 0, it MUST be stopped
    # before changing use_lvmetad to 1 and started again afterwards.
    # If using lvmetad, the volume activation is also switched to automatic
    # event-based mode. In this mode, the volumes are activated based on
    # incoming udev events that automatically inform lvmetad about new PVs
    # that appear in the system. Once the VG is complete (all the PVs are
    # present), it is auto-activated. The activation/auto_activation_volume_list
    # setting controls which volumes are auto-activated (all by default).
    # A note about device filtering while lvmetad is used:
    # When lvmetad is updated (either automatically based on udev events
    # or directly by pvscan --cache <device> call), the devices/filter
    # is ignored and all devices are scanned by default. The lvmetad always
    # keeps unfiltered information which is then provided to LVM commands
    # and then each LVM command does the filtering based on devices/filter
    # setting itself.
    # To prevent scanning devices completely, even when using lvmetad,
    # the devices/global_filter must be used.
    # N.B. Don't use lvmetad with locking type 3 as lvmetad is not yet
    # supported in clustered environment. If use_lvmetad=1 and locking_type=3
    # is set at the same time, LVM always issues a warning message about this
    # and then it automatically disables lvmetad use.
    use_lvmetad = 0
    # Full path of the utility called to check that a thin metadata device
    # is in a state that allows it to be used.
    # Each time a thin pool needs to be activated or after it is deactivated
    # this utility is executed. The activation will only proceed if the utility
    # has an exit status of 0.
    # Set to "" to skip this check. (Not recommended.)
    # The thin tools are available as part of the device-mapper-persistent-data
    # package from https://github.com/jthornber/thin-provisioning-tools.
    # thin_check_executable = "/usr/bin/thin_check"
    # Array of string options passed with thin_check command. By default,
    # option "-q" is for quiet output.
    # With thin_check version 2.1 or newer you can add "--ignore-non-fatal-errors"
    # to let it pass through ignorable errors and fix them later.
    # thin_check_options = [ "-q" ]
    # Full path of the utility called to repair a thin metadata device
    # is in a state that allows it to be used.
    # Each time a thin pool needs repair this utility is executed.
    # See thin_check_executable how to obtain binaries.
    # thin_repair_executable = "/usr/bin/thin_repair"
    # Array of extra string options passed with thin_repair command.
    # thin_repair_options = [ "" ]
    # Full path of the utility called to dump thin metadata content.
    # See thin_check_executable how to obtain binaries.
    # thin_dump_executable = "/usr/bin/thin_dump"
    # If set, given features are not used by thin driver.
    # This can be helpful not just for testing, but i.e. allows to avoid
    # using problematic implementation of some thin feature.
    # Features:
    # block_size
    # discards
    # discards_non_power_2
    # external_origin
    # metadata_resize
    # external_origin_extend
    # thin_disabled_features = [ "discards", "block_size" ]
    activation {
    # Set to 1 to perform internal checks on the operations issued to
    # libdevmapper. Useful for debugging problems with activation.
    # Some of the checks may be expensive, so it's best to use this
    # only when there seems to be a problem.
    checks = 0
    # Set to 0 to disable udev synchronisation (if compiled into the binaries).
    # Processes will not wait for notification from udev.
    # They will continue irrespective of any possible udev processing
    # in the background. You should only use this if udev is not running
    # or has rules that ignore the devices LVM2 creates.
    # The command line argument --nodevsync takes precedence over this setting.
    # If set to 1 when udev is not running, and there are LVM2 processes
    # waiting for udev, run 'dmsetup udevcomplete_all' manually to wake them up.
    udev_sync = 1
    # Set to 0 to disable the udev rules installed by LVM2 (if built with
    # --enable-udev_rules). LVM2 will then manage the /dev nodes and symlinks
    # for active logical volumes directly itself.
    # N.B. Manual intervention may be required if this setting is changed
    # while any logical volumes are active.
    udev_rules = 1
    # Set to 1 for LVM2 to verify operations performed by udev. This turns on
    # additional checks (and if necessary, repairs) on entries in the device
    # directory after udev has completed processing its events.
    # Useful for diagnosing problems with LVM2/udev interactions.
    verify_udev_operations = 0
    # If set to 1 and if deactivation of an LV fails, perhaps because
    # a process run from a quick udev rule temporarily opened the device,
    # retry the operation for a few seconds before failing.
    retry_deactivation = 1
    # How to fill in missing stripes if activating an incomplete volume.
    # Using "error" will make inaccessible parts of the device return
    # I/O errors on access. You can instead use a device path, in which
    # case, that device will be used to in place of missing stripes.
    # But note that using anything other than "error" with mirrored
    # or snapshotted volumes is likely to result in data corruption.
    missing_stripe_filler = "error"
    # The linear target is an optimised version of the striped target
    # that only handles a single stripe. Set this to 0 to disable this
    # optimisation and always use the striped target.
    use_linear_target = 1
    # How much stack (in KB) to reserve for use while devices suspended
    # Prior to version 2.02.89 this used to be set to 256KB
    reserved_stack = 64
    # How much memory (in KB) to reserve for use while devices suspended
    reserved_memory = 8192
    # Nice value used while devices suspended
    process_priority = -18
    # If volume_list is defined, each LV is only activated if there is a
    # match against the list.
    # "vgname" and "vgname/lvname" are matched exactly.
    # "@tag" matches any tag set in the LV or VG.
    # "@*" matches if any tag defined on the host is also set in the LV or VG
    # If any host tags exist but volume_list is not defined, a default
    # single-entry list containing "@*" is assumed.
    # volume_list = [ "vg1", "vg2/lvol1", "@tag1", "@*" ]
    # If auto_activation_volume_list is defined, each LV that is to be
    # activated with the autoactivation option (--activate ay/-a ay) is
    # first checked against the list. There are two scenarios in which
    # the autoactivation option is used:
    # - automatic activation of volumes based on incoming PVs. If all the
    # PVs making up a VG are present in the system, the autoactivation
    # is triggered. This requires lvmetad (global/use_lvmetad=1) and udev
    # to be running. In this case, "pvscan --cache -aay" is called
    # automatically without any user intervention while processing
    # udev events. Please, make sure you define auto_activation_volume_list
    # properly so only the volumes you want and expect are autoactivated.
    # - direct activation on command line with the autoactivation option.
    # In this case, the user calls "vgchange --activate ay/-a ay" or
    # "lvchange --activate ay/-a ay" directly.
    # By default, the auto_activation_volume_list is not defined and all
    # volumes will be activated either automatically or by using --activate ay/-a ay.
    # N.B. The "activation/volume_list" is still honoured in all cases so even
    # if the VG/LV passes the auto_activation_volume_list, it still needs to
    # pass the volume_list for it to be activated in the end.
    # If auto_activation_volume_list is defined but empty, no volumes will be
    # activated automatically and --activate ay/-a ay will do nothing.
    # auto_activation_volume_list = []
    # If auto_activation_volume_list is defined and it's not empty, only matching
    # volumes will be activated either automatically or by using --activate ay/-a ay.
    # "vgname" and "vgname/lvname" are matched exactly.
    # "@tag" matches any tag set in the LV or VG.
    # "@*" matches if any tag defined on the host is also set in the LV or VG
    # auto_activation_volume_list = [ "vg1", "vg2/lvol1", "@tag1", "@*" ]
    # If read_only_volume_list is defined, each LV that is to be activated
    # is checked against the list, and if it matches, it as activated
    # in read-only mode. (This overrides '--permission rw' stored in the
    # metadata.)
    # "vgname" and "vgname/lvname" are matched exactly.
    # "@tag" matches any tag set in the LV or VG.
    # "@*" matches if any tag defined on the host is also set in the LV or VG
    # read_only_volume_list = [ "vg1", "vg2/lvol1", "@tag1", "@*" ]
    # Each LV can have an 'activation skip' flag stored persistently against it.
    # During activation, this flag is used to decide whether such an LV is skipped.
    # The 'activation skip' flag can be set during LV creation and by default it
    # is automatically set for thin snapshot LVs. The 'auto_set_activation_skip'
    # enables or disables this automatic setting of the flag while LVs are created.
    # auto_set_activation_skip = 1
    # For RAID or 'mirror' segment types, 'raid_region_size' is the
    # size (in KiB) of each:
    # - synchronization operation when initializing
    # - each copy operation when performing a 'pvmove' (using 'mirror' segtype)
    # This setting has replaced 'mirror_region_size' since version 2.02.99
    raid_region_size = 512
    # Setting to use when there is no readahead value stored in the metadata.
    # "none" - Disable readahead.
    # "auto" - Use default value chosen by kernel.
    readahead = "auto"
    # 'raid_fault_policy' defines how a device failure in a RAID logical
    # volume is handled. This includes logical volumes that have the following
    # segment types: raid1, raid4, raid5*, and raid6*.
    # In the event of a failure, the following policies will determine what
    # actions are performed during the automated response to failures (when
    # dmeventd is monitoring the RAID logical volume) and when 'lvconvert' is
    # called manually with the options '--repair' and '--use-policies'.
    # "warn" - Use the system log to warn the user that a device in the RAID
    # logical volume has failed. It is left to the user to run
    # 'lvconvert --repair' manually to remove or replace the failed
    # device. As long as the number of failed devices does not
    # exceed the redundancy of the logical volume (1 device for
    # raid4/5, 2 for raid6, etc) the logical volume will remain
    # usable.
    # "allocate" - Attempt to use any extra physical volumes in the volume
    # group as spares and replace faulty devices.
    raid_fault_policy = "warn"
    # 'mirror_image_fault_policy' and 'mirror_log_fault_policy' define
    # how a device failure affecting a mirror (of "mirror" segment type) is
    # handled. A mirror is composed of mirror images (copies) and a log.
    # A disk log ensures that a mirror does not need to be re-synced
    # (all copies made the same) every time a machine reboots or crashes.
    # In the event of a failure, the specified policy will be used to determine
    # what happens. This applies to automatic repairs (when the mirror is being
    # monitored by dmeventd) and to manual lvconvert --repair when
    # --use-policies is given.
    # "remove" - Simply remove the faulty device and run without it. If
    # the log device fails, the mirror would convert to using
    # an in-memory log. This means the mirror will not
    # remember its sync status across crashes/reboots and
    # the entire mirror will be re-synced. If a
    # mirror image fails, the mirror will convert to a
    # non-mirrored device if there is only one remaining good
    # copy.
    # "allocate" - Remove the faulty device and try to allocate space on
    # a new device to be a replacement for the failed device.
    # Using this policy for the log is fast and maintains the
    # ability to remember sync state through crashes/reboots.
    # Using this policy for a mirror device is slow, as it
    # requires the mirror to resynchronize the devices, but it
    # will preserve the mirror characteristic of the device.
    # This policy acts like "remove" if no suitable device and
    # space can be allocated for the replacement.
    # "allocate_anywhere" - Not yet implemented. Useful to place the log device
    # temporarily on same physical volume as one of the mirror
    # images. This policy is not recommended for mirror devices
    # since it would break the redundant nature of the mirror. This
    # policy acts like "remove" if no suitable device and space can
    # be allocated for the replacement.
    mirror_log_fault_policy = "allocate"
    mirror_image_fault_policy = "remove"
    # 'snapshot_autoextend_threshold' and 'snapshot_autoextend_percent' define
    # how to handle automatic snapshot extension. The former defines when the
    # snapshot should be extended: when its space usage exceeds this many
    # percent. The latter defines how much extra space should be allocated for
    # the snapshot, in percent of its current size.
    # For example, if you set snapshot_autoextend_threshold to 70 and
    # snapshot_autoextend_percent to 20, whenever a snapshot exceeds 70% usage,
    # it will be extended by another 20%. For a 1G snapshot, using up 700M will
    # trigger a resize to 1.2G. When the usage exceeds 840M, the snapshot will
    # be extended to 1.44G, and so on.
    # Setting snapshot_autoextend_threshold to 100 disables automatic
    # extensions. The minimum value is 50 (A setting below 50 will be treated
    # as 50).
    snapshot_autoextend_threshold = 100
    snapshot_autoextend_percent = 20
    # 'thin_pool_autoextend_threshold' and 'thin_pool_autoextend_percent' define
    # how to handle automatic pool extension. The former defines when the
    # pool should be extended: when its space usage exceeds this many
    # percent. The latter defines how much extra space should be allocated for
    # the pool, in percent of its current size.
    # For example, if you set thin_pool_autoextend_threshold to 70 and
    # thin_pool_autoextend_percent to 20, whenever a pool exceeds 70% usage,
    # it will be extended by another 20%. For a 1G pool, using up 700M will
    # trigger a resize to 1.2G. When the usage exceeds 840M, the pool will
    # be extended to 1.44G, and so on.
    # Setting thin_pool_autoextend_threshold to 100 disables automatic
    # extensions. The minimum value is 50 (A setting below 50 will be treated
    # as 50).
    thin_pool_autoextend_threshold = 100
    thin_pool_autoextend_percent = 20
    # While activating devices, I/O to devices being (re)configured is
    # suspended, and as a precaution against deadlocks, LVM2 needs to pin
    # any memory it is using so it is not paged out. Groups of pages that
    # are known not to be accessed during activation need not be pinned
    # into memory. Each string listed in this setting is compared against
    # each line in /proc/self/maps, and the pages corresponding to any
    # lines that match are not pinned. On some systems locale-archive was
    # found to make up over 80% of the memory used by the process.
    # mlock_filter = [ "locale/locale-archive", "gconv/gconv-modules.cache" ]
    # Set to 1 to revert to the default behaviour prior to version 2.02.62
    # which used mlockall() to pin the whole process's memory while activating
    # devices.
    use_mlockall = 0
    # Monitoring is enabled by default when activating logical volumes.
    # Set to 0 to disable monitoring or use the --ignoremonitoring option.
    monitoring = 1
    # When pvmove or lvconvert must wait for the kernel to finish
    # synchronising or merging data, they check and report progress
    # at intervals of this number of seconds. The default is 15 seconds.
    # If this is set to 0 and there is only one thing to wait for, there
    # are no progress reports, but the process is awoken immediately the
    # operation is complete.
    polling_interval = 15
    # Report settings.
    # report {
    # Align columns on report output.
    # aligned=1
    # When buffered reporting is used, the report's content is appended
    # incrementally to include each object being reported until the report
    # is flushed to output which normally happens at the end of command
    # execution. Otherwise, if buffering is not used, each object is
    # reported as soon as its processing is finished.
    # buffered=1
    # Show headings for columns on report.
    # headings=1
    # A separator to use on report after each field.
    # separator=" "
    # Use a field name prefix for each field reported.
    # prefixes=0
    # Quote field values when using field name prefixes.
    # quoted=1
    # Output each column as a row. If set, this also implies report/prefixes=1.
    # colums_as_rows=0
    # Comma separated list of columns to sort by when reporting 'lvm devtypes' command.
    # See 'lvm devtypes -o help' for the list of possible fields.
    # devtypes_sort="devtype_name"
    # Comma separated list of columns to report for 'lvm devtypes' command.
    # See 'lvm devtypes -o help' for the list of possible fields.
    # devtypes_cols="devtype_name,devtype_max_partitions,devtype_description"
    # Comma separated list of columns to report for 'lvm devtypes' command in verbose mode.
    # See 'lvm devtypes -o help' for the list of possible fields.
    # devtypes_cols_verbose="devtype_name,devtype_max_partitions,devtype_description"
    # Comma separated list of columns to sort by when reporting 'lvs' command.
    # See 'lvs -o help' for the list of possible fields.
    # lvs_sort="vg_name,lv_name"
    # Comma separated list of columns to report for 'lvs' command.
    # See 'lvs -o help' for the list of possible fields.
    # lvs_cols="lv_name,vg_name,lv_attr,lv_size,pool_lv,origin,data_percent,move_pv,mirror_log,copy_percent,convert_lv"
    # Comma separated list of columns to report for 'lvs' command in verbose mode.
    # See 'lvs -o help' for the list of possible fields.
    # lvs_cols_verbose="lv_name,vg_name,seg_count,lv_attr,lv_size,lv_major,lv_minor,lv_kernel_major,lv_kernel_minor,pool_lv,origin,data_percent,metadata_percent,move_pv,copy_percent,mirror_log,convert
    # Comma separated list of columns to sort by when reporting 'vgs' command.
    # See 'vgs -o help' for the list of possible fields.
    # vgs_sort="vg_name"
    # Comma separated list of columns to report for 'vgs' command.
    # See 'vgs -o help' for the list of possible fields.
    # vgs_cols="vg_name,pv_count,lv_count,snap_count,vg_attr,vg_size,vg_free"
    # Comma separated list of columns to report for 'vgs' command in verbose mode.
    # See 'vgs -o help' for the list of possible fields.
    # vgs_cols_verbose="vg_name,vg_attr,vg_extent_size,pv_count,lv_count,snap_count,vg_size,vg_free,vg_uuid,vg_profile"
    # Comma separated list of columns to sort by when reporting 'pvs' command.
    # See 'pvs -o help' for the list of possible fields.
    # pvs_sort="pv_name"
    # Comma separated list of columns to report for 'pvs' command.
    # See 'pvs -o help' for the list of possible fields.
    # pvs_cols="pv_name,vg_name,pv_fmt,pv_attr,pv_size,pv_free"
    # Comma separated list of columns to report for 'pvs' command in verbose mode.
    # See 'pvs -o help' for the list of possible fields.
    # pvs_cols_verbose="pv_name,vg_name,pv_fmt,pv_attr,pv_size,pv_free,dev_size,pv_uuid"
    # Comma separated list of columns to sort by when reporting 'lvs --segments' command.
    # See 'lvs --segments -o help' for the list of possible fields.
    # segs_sort="vg_name,lv_name,seg_start"
    # Comma separated list of columns to report for 'lvs --segments' command.
    # See 'lvs --segments -o help' for the list of possible fields.
    # segs_cols="lv_name,vg_name,lv_attr,stripes,segtype,seg_size"
    # Comma separated list of columns to report for 'lvs --segments' command in verbose mode.
    # See 'lvs --segments -o help' for the list of possible fields.
    # segs_cols_verbose="lv_name,vg_name,lv_attr,seg_start,seg_size,stripes,segtype,stripesize,chunksize"
    # Comma separated list of columns to sort by when reporting 'pvs --segments' command.
    # See 'pvs --segments -o help' for the list of possible fields.
    # pvsegs_sort="pv_name,pvseg_start"
    # Comma separated list of columns to sort by when reporting 'pvs --segments' command.
    # See 'pvs --segments -o help' for the list of possible fields.
    # pvsegs_cols="pv_name,vg_name,pv_fmt,pv_attr,pv_size,pv_free,pvseg_start,pvseg_size"
    # Comma separated list of columns to sort by when reporting 'pvs --segments' command in verbose mode.
    # See 'pvs --segments -o help' for the list of possible fields.
    # pvsegs_cols_verbose="pv_name,vg_name,pv_fmt,pv_attr,pv_size,pv_free,pvseg_start,pvseg_size,lv_name,seg_start_pe,segtype,seg_pe_ranges"
    # Advanced section #
    # Metadata settings
    # metadata {
    # Default number of copies of metadata to hold on each PV. 0, 1 or 2.
    # You might want to override it from the command line with 0
    # when running pvcreate on new PVs which are to be added to large VGs.
    # pvmetadatacopies = 1
    # Default number of copies of metadata to maintain for each VG.
    # If set to a non-zero value, LVM automatically chooses which of
    # the available metadata areas to use to achieve the requested
    # number of copies of the VG metadata. If you set a value larger
    # than the the total number of metadata areas available then
    # metadata is stored in them all.
    # The default value of 0 ("unmanaged") disables this automatic
    # management and allows you to control which metadata areas
    # are used at the individual PV level using 'pvchange
    # --metadataignore y/n'.
    # vgmetadatacopies = 0
    # Approximate default size of on-disk metadata areas in sectors.
    # You should increase this if you have large volume groups or
    # you want to retain a large on-disk history of your metadata changes.
    # pvmetadatasize = 255
    # List of directories holding live copies of text format metadata.
    # These directories must not be on logical volumes!
    # It's possible to use LVM2 with a couple of directories here,
    # preferably on different (non-LV) filesystems, and with no other
    # on-disk metadata (pvmetadatacopies = 0). Or this can be in
    # addition to on-disk metadata areas.
    # The feature was originally added to simplify testing and is not
    # supported under low memory situations - the machine could lock up.
    # Never edit any files in these directories by hand unless you
    # you are absolutely sure you know what you are doing! Use
    # the supplied toolset to make changes (e.g. vgcfgrestore).
    # dirs = [ "/etc/lvm/metadata", "/mnt/disk2/lvm/metadata2" ]
    # Event daemon
    dmeventd {
    # mirror_library is the library used when monitoring a mirror device.
    # "libdevmapper-event-lvm2mirror.so" attempts to recover from
    # failures. It removes failed devices from a volume group and
    # reconfigures a mirror as necessary. If no mirror library is
    # provided, mirrors are not monitored through dmeventd.
    mirror_library = "libdevmapper-event-lvm2mirror.so"
    # snapshot_library is the library used when monitoring a snapshot device.
    # "libdevmapper-event-lvm2snapshot.so" monitors the filling of
    # snapshots and emits a warning through syslog when the use of
    # the snapshot exceeds 80%. The warning is repeated when 85%, 90% and
    # 95% of the snapshot is filled.
    snapshot_library = "libdevmapper-event-lvm2snapshot.so"
    # thin_library is the library used when monitoring a thin device.
    # "libdevmapper-event-lvm2thin.so" monitors the filling of
    # pool and emits a warning through syslog when the use of
    # the pool exceeds 80%. The warning is repeated when 85%, 90% and
    # 95% of the pool is filled.
    thin_library = "libdevmapper-event-lvm2thin.so"
    # Full path of the dmeventd binary.
    # executable = "/usr/sbin/dmeventd"

  • Grid Control performance tab not available when on the target home of ASM

    I noticed recently that the "Performance" tab/link is not available when you are using 10g Grid Control and you are on the home page of an ASM instance. Can this be explained?

    Hi MarGur
    I try to help you. The OEM Grid Control in some times not discover the Oracle 9i databases and you must to add manually.
    First, check in your target machine that the user owner of agent software have rights of read and write in the ORACLE_HOME directories of Oracle 9i databases.
    Go to Grid Control Console
    Select Targets ---> Databases
    Press Add --> Select the host ebisdb2.oneshield.com and press continue
    The automatic discover process begin. When finished is very normal that nor discover the Oracle 9i databases. You press the Add manually button
    in the page that appear you must entry all the configuration information of database:
    ORACLE_HOME directory
    DBSNMP USER
    DBSNMP PASSWORD
    HOSTNAME
    PORT
    SID
    Check that the dbsnmp user in your database is unlock and press the TEST button to verify the connection work fine.
    After this many windows appear of configuration process, and you will put the sys password and host credentials.
    If you have problems, tell me
    Regards

  • My youtube does not work on safari only when i try to play a video it says "This video is not available on mobile add to playlist" But i am using a mac air 11 inch it is not a mobile. Please Help. i have to use the youtube on firefox or chrome

    my youtube does not work on safari only when i try to play a video it says "This video is not available on mobile add to playlist" But i am using a mac air 11 inch it is not a mobile. Please Help. i have to use the youtube on firefox or chrome. At times it also say QuickTime Player can't open "video.3gp". and
    The file may be damaged or may not be a movie file that is compatible with QuickTime Player. when i try to open a video on youtube on safari

    I too am having the same issue as the OP.
    Your USER AGENT information is Mozilla/5.0 (Macintosh; Intel Mac OS X 10_10_2) AppleWebKit/600.3.18 (KHTML, like Gecko) Version/8.0.3 Safari/600.3.18
    Every webserver that receives a request from your browser is able to determine the HTTP USER AGENT information unless it has been removed by some software (e.g. firewall) before the request was trasmitted.

  • Error "Object 'DM' (APCO) of type 'Application' is not available in version 'A'" while transporting process chain to production system.

    Hi Gurus,
    Whenever I try to transport a process chain ZFI_0FIGL_C10 from dev. to qua. & prod. the error “Object 'DM' (APCO) of type 'Application' is not available in version 'A'” is displayed in the end of creating transport package.
    The process chain was created to load data to 0figl_c10 and it works all right on all systems (dev. qua. prod.).
    The process chain contains processes for:
    - starting the process chain
    - executing infopackage (for loading data to 0figl_o10),
    - ACR (activate data in 0figl_o10)
    - delete index 0figl_c10
    - DTP loading data from 0figl_o14 to 0figl_c10    
    - executing infopackage 80FIGL_O10
    - creating index in 0figl_c10
    - building DB statistics
    - rolling up aggregates in 0figl_c10
    - deleting of requests from the changelog 0figl_o10
    - deleting of requests from PSA (0fi_gl_10)
    How to find what cause that error.
    Regards,
    Leszek

    Hi,
    A SAP consultant told me to ignore that error. 
    Indeed, after creating transport with some new process chains (again I was informed about that error "APCO ...") and transporting it to production everything works all right.
    The problem is described in sap note: 883745.
    Regards, Leszek  

Maybe you are looking for

  • Need help, I seem to have lost 2 months of photos

    Hello, I own PSE5 on Windows XP. I just recently purchased a hard drive and was in the process of moving my photos to it. I was following the article on Adobe's website entitled Move files to another disk using Folder Location Arrangement in Photosho

  • How can I change the return address ?

    I have an email forwarding service with it's own email address. Any email sent there will forward the email to any address I specify. I want to be able to write messages from my Mail account and have the return address (i.e. when they reply it goes t

  • Drill down (Double click) option in BSP

    I need to provide  double click(drill down) option in the  report layout of BSP. How to do that.

  • Can a FM in R/3 be made system independent?

    Hi Experts, My requirement is such that somehow I want to expose a FM already existing in R/3 to be used in some other application such as WD JAVA or BPM. Now I know how I can do that but the question is I do not want to connect to R/3 while executin

  • Is there a setting to remove all duplicates from my library on iPhoto '11? (version 9.1.3)

    Is there a setting that I can use to delete all duplicates in my iPhoto library?  My last iMac was "dying", and I attempted to save photos by burning to CD.  When I downloaded from this CD, have incorrect dates and multiple duplicates.  It will take