Confused about ZFS filesystems created with Solaris 11 Zone

Hello.
Installing a blank Zone in Solaris *10* with "zonepath=/export/zones/TESTvm01" just creates one zfs filesystem:
+"zfs list+
+...+
+rzpool/export/zones/TESTvm01 4.62G 31.3G 4.62G /export/zones/TESTvm01"+
Doing the same steps with Solaris *11* will ?create? more filesystems:
+"zfs list+
+...+
+rpool/export/zones/TESTvm05 335M 156G 32K /export/zones/TESTvm05+
+rpool/export/zones/TESTvm05/rpool 335M 156G 31K /rpool+
+rpool/export/zones/TESTvm05/rpool/ROOT 335M 156G 31K legacy+
+rpool/export/zones/TESTvm05/rpool/ROOT/solaris 335M 156G 310M /export/zones/TESTvm05/root+
+rpool/export/zones/TESTvm05/rpool/ROOT/solaris/var 24.4M 156G 23.5M /export/zones/TESTvm05/root/var+
+rpool/export/zones/TESTvm05/rpool/export 62K 156G 31K /export+
+rpool/export/zones/TESTvm05/rpool/export/home 31K 156G 31K /export/home"+
I dont understand why Solaris 11 is doing that. Just one FS (like in Solaris 10) would be better for my setup. I want to configure all created volumes by myself.
Is it possible to deactivate this automatic "feature"?

There are several reasons that it works like this, all guided by the simple idea "everything in a zone should work exactly like it does in the global zone, unless that is impractical." By having this layout we get:
* The same zfs administrative practices within a zone that are found in the global zone. This allows, for example, compression, encryption, etc. of parts of the zone.
* beadm(1M) and pkg(1) are able to create boot environments within the zone, thus making it easy to keep the global zone software in sync with non-global zone software as the system is updated (equivalent of patching in Solaris 10). Note that when Solaris 11 updates the kernel, core libraries, and perhaps other things, a new boot environment is automatically created (for the global zone and each zone) and the updates are done to the new boot environment(s). Thus, you get the benefits that Live Upgrade offered without the severe headaches that sometimes come with Live Upgrade.
* The ability to have a separate /var file system. This is required by policies at some large customers, such as the US Department of Defense via the DISA STIG.
* The ability to perform a p2v of a global zone into a zone (see solaris(5) for examples) without losing the dataset hierarchy or properties (e.g. compression, etc.) set on datasets in that hierarchy.
When this dataset hierarchy is combined with the fact that the ZFS namespace is virtualized in a zone (a feature called "dataset aliasing"), you see the same hierarchy in the zone that you would see in the global zone. Thus, you don't have confusing output from df saying that / is mounted on / and such.
Because there is integration between pkg, beadm, zones, and zfs, there is no way to disable this behavior. You can remove and optionally replace /export with something else if you wish.
If your goal is to prevent zone administrators from altering the dataset hierarchy, you may be able to accomplish this with immutable zones (see zones admin guide or file-mac-profile in zonecfg(1M)). This will have other effects as well, such as making all or most of the zone unwritable. If needed, you can add fs or dataset resources which will not be subject to file-mac-profile and as such will be writable.

Similar Messages

  • Zfs core dumps in Solaris Zones after patchcluster installation

    Hi
    I've installed the recommended patch cluster for Solaris 10 (SPARC) on my T2000 Server.
    Now after the reboot all the Solaris Zones (got about 10 of them) instantly do zfs core dumps when their booted.
    Some Output:
    Dec 15 15:34:43 XXXXX genunix: NOTICE: core_log: zfs[1711] core dumped: /var/core/core_XXXXXX_zfs_0_0_1292423682_1711
    bash-3.00# svcs
    STATE STIME FMRI
    online 13:58:50 svc:/system/svc/restarter:default
    online 13:58:51 svc:/system/filesystem/root:default
    online 13:58:51 svc:/network/loopback:default
    online 13:58:53 svc:/network/pfil:default
    online 13:58:54 svc:/system/installupdates:default
    online 13:58:55 svc:/system/boot-archive:default
    online 13:58:56 svc:/network/physical:default
    online 13:58:57 svc:/system/filesystem/usr:default
    online 13:58:57 svc:/system/identity:node
    online 13:58:58 svc:/system/device/local:default
    online 13:58:58 svc:/system/keymap:default
    online 13:58:58 svc:/milestone/devices:default
    online 13:58:58 svc:/system/filesystem/minimal:default
    online 13:58:58 svc:/system/rmtmpfiles:default
    online 13:58:58 svc:/system/cryptosvc:default
    online 13:58:59 svc:/system/identity:domain
    online 13:58:59 svc:/system/name-service-cache:default
    online 13:58:59 svc:/system/coreadm:default
    online 13:58:59 svc:/network/ipsec/ipsecalgs:default
    online 13:58:59 svc:/application/print/ppd-cache-update:default
    online 13:58:59 svc:/network/ipsec/policy:default
    online 13:59:00 svc:/milestone/network:default
    online 13:59:00 svc:/network/initial:default
    online 13:59:00 svc:/system/manifest-import:default
    online 13:59:00 svc:/network/service:default
    online 13:59:00 svc:/milestone/single-user:default
    online 13:59:00 svc:/network/dns/client:default
    online 13:59:00 svc:/network/routing-setup:default
    online 13:59:00 svc:/milestone/name-services:default
    online 13:59:03 svc:/milestone/sysconfig:default
    online 13:59:03 svc:/system/utmp:default
    online 13:59:03 svc:/system/console-login:default
    offline 13:58:51 svc:/system/sysidtool:net
    offline 13:58:51 svc:/network/rpc/bind:default
    offline 13:58:51 svc:/system/sysidtool:system
    offline 13:58:52 svc:/network/nfs/status:default
    offline 13:58:52 svc:/network/nfs/nlockmgr:default
    offline 13:58:52 svc:/network/nfs/cbd:default
    offline 13:58:52 svc:/network/nfs/mapid:default
    offline 13:58:52 svc:/network/inetd:default
    offline 13:58:52 svc:/network/nfs/client:default
    offline 13:58:52 svc:/system/filesystem/autofs:default
    offline 13:58:53 svc:/system/system-log:default
    offline 13:58:53 svc:/network/smtp:sendmail
    offline 13:58:53 svc:/system/cron:default
    offline 13:58:53 svc:/milestone/multi-user:default
    offline 13:58:54 svc:/application/management/snmpdx:default
    offline 13:58:54 svc:/application/management/dmi:default
    offline 13:58:54 svc:/application/management/seaport:default
    offline 13:58:54 svc:/network/ssh:default
    offline 13:58:54 svc:/milestone/multi-user-server:default
    offline 13:58:54 svc:/application/font/fc-cache:default
    offline 13:58:55 svc:/application/management/sma:default
    offline 13:58:58 svc:/system/sac:default
    offline 13:59:00 svc:/network/shares/group:default
    offline 13:59:00 svc:/system/boot-archive-update:default
    maintenance 15:34:43 svc:/system/filesystem/local:default
    uninitialized 13:58:52 svc:/network/rpc/gss:default
    uninitialized 13:58:54 svc:/application/font/stfsloader:default
    uninitialized 13:58:55 svc:/network/rpc/rstat:default
    uninitialized 13:58:55 svc:/application/print/rfc1179:default
    uninitialized 13:58:55 svc:/application/x11/xfs:default
    uninitialized 13:58:55 svc:/network/finger:default
    uninitialized 13:58:55 svc:/network/ftp:default
    uninitialized 13:58:56 svc:/network/login:rlogin
    uninitialized 13:58:56 svc:/network/nfs/rquota:default
    uninitialized 13:58:56 svc:/network/rpc/rusers:default
    uninitialized 13:58:56 svc:/network/rpc/smserver:default
    uninitialized 13:58:57 svc:/network/security/ktkt_warn:default
    uninitialized 13:58:57 svc:/network/shell:default
    uninitialized 13:58:57 svc:/network/telnet:default
    uninitialized 13:58:58 svc:/network/rpc-100235_1/rpc_ticotsord:default
    uninitialized 13:58:58 svc:/network/rpc/cde-calendar-manager:default
    uninitialized 13:58:59 svc:/network/rpc/cde-ttdbserver:tcp
    bash-3.00# svcs -x
    svc:/system/filesystem/local:default (local file system mounts)
    State: maintenance since Wed Dec 15 15:34:43 2010
    Reason: Start method exited with $SMF_EXIT_ERR_FATAL.
    See: http://sun.com/msg/SMF-8000-KS
    See: /var/svc/log/system-filesystem-local:default.log
    Impact: 24 dependent services are not running. (Use -v for list.)
    svc:/network/rpc/gss:default (Generic Security Service)
    State: uninitialized since Wed Dec 15 13:58:52 2010
    Reason: Restarter svc:/network/inetd:default is not running.
    See: http://sun.com/msg/SMF-8000-5H
    See: gssd(1M)
    Impact: 11 dependent services are not running. (Use -v for list.)
    svc:/network/rpc/rstat:default (kernel statistics server)
    State: uninitialized since Wed Dec 15 13:58:55 2010
    Reason: Restarter svc:/network/inetd:default is not running.
    See: http://sun.com/msg/SMF-8000-5H
    See: rpc.rstatd(1M)
    See: rstatd(1M)
    Impact: 1 dependent service is not running. (Use -v for list.)
    bash-3.00# cat /var/svc/log/system-filesystem-local:default.log
    [ Jul 10 03:41:35 Enabled. ]
    [ Jul 10 03:43:45 Rereading configuration. ]
    [ Jul 10 03:44:17 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Jul 10 03:44:18 Method "start" exited with status 0 ]
    [ Jul 10 12:44:36 Enabled. ]
    [ Jul 10 12:44:45 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Jul 10 12:44:45 Method "start" exited with status 0 ]
    [ Jul 10 12:45:03 Enabled. ]
    [ Jul 10 12:45:09 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Jul 10 12:45:09 Method "start" exited with status 0 ]
    [ Jul 10 13:54:39 Enabled. ]
    [ Jul 10 13:54:46 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Jul 10 13:54:46 Method "start" exited with status 0 ]
    [ Jul 18 12:34:19 Enabled. ]
    [ Jul 18 12:34:27 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Jul 18 12:34:27 Method "start" exited with status 0 ]
    [ Jul 18 17:29:08 Enabled. ]
    [ Jul 18 17:29:15 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Jul 18 17:29:15 Method "start" exited with status 0 ]
    [ Jul 19 11:13:34 Enabled. ]
    [ Jul 19 11:13:42 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Jul 19 11:13:42 Method "start" exited with status 0 ]
    [ Jul 19 11:32:23 Enabled. ]
    [ Jul 19 11:32:30 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Jul 19 11:32:31 Method "start" exited with status 0 ]
    [ Jul 19 11:57:06 Enabled. ]
    [ Jul 19 11:57:13 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Jul 19 11:57:13 Method "start" exited with status 0 ]
    [ Jul 19 12:40:05 Enabled. ]
    [ Jul 19 12:40:12 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Jul 19 12:40:12 Method "start" exited with status 0 ]
    [ Jul 20 15:51:46 Enabled. ]
    [ Jul 20 15:51:54 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Jul 20 15:51:54 Method "start" exited with status 0 ]
    [ Jul 24 11:42:17 Enabled. ]
    [ Jul 24 11:42:25 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Jul 24 11:42:25 Method "start" exited with status 0 ]
    [ Jul 27 12:05:06 Stopping because service disabled. ]
    [ Jul 27 12:05:07 Executing stop method (null) ]
    [ Jul 27 12:13:10 Enabled. ]
    [ Jul 27 12:13:17 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Jul 27 12:13:17 Method "start" exited with status 0 ]
    [ Jul 27 13:10:26 Stopping because service disabled. ]
    [ Jul 27 13:10:26 Executing stop method (null) ]
    [ Jul 27 13:18:19 Enabled. ]
    [ Jul 27 13:18:26 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Jul 27 13:18:26 Method "start" exited with status 0 ]
    [ Jul 31 10:12:31 Enabled. ]
    [ Jul 31 10:12:38 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Jul 31 10:12:39 Method "start" exited with status 0 ]
    [ Aug  7 10:08:52 Stopping because service disabled. ]
    [ Aug  7 10:08:52 Executing stop method (null) ]
    [ Aug  7 10:34:50 Enabled. ]
    [ Aug  7 10:34:57 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug  7 10:34:58 Method "start" exited with status 0 ]
    [ Aug  9 09:14:50 Stopping because service disabled. ]
    [ Aug  9 09:14:50 Executing stop method (null) ]
    [ Aug  9 09:55:03 Enabled. ]
    [ Aug  9 09:55:10 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug  9 09:55:11 Method "start" exited with status 0 ]
    [ Aug 16 10:46:43 Enabled. ]
    [ Aug 16 10:46:50 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug 16 10:46:50 Method "start" exited with status 0 ]
    [ Aug 16 12:06:56 Enabled. ]
    [ Aug 16 12:07:02 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug 16 12:07:02 Method "start" exited with status 0 ]
    [ Aug 16 13:45:10 Enabled. ]
    [ Aug 16 13:45:16 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug 16 13:45:16 Method "start" exited with status 0 ]
    [ Aug 16 14:36:59 Enabled. ]
    [ Aug 16 14:37:07 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug 16 14:37:07 Method "start" exited with status 0 ]
    [ Aug 16 15:00:40 Enabled. ]
    [ Aug 16 15:00:47 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug 16 15:00:47 Method "start" exited with status 0 ]
    [ Aug 22 16:00:13 Enabled. ]
    [ Aug 22 16:00:21 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug 22 16:00:21 Method "start" exited with status 0 ]
    [ Aug 23 09:49:04 Enabled. ]
    [ Aug 23 09:49:10 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug 23 09:49:11 Method "start" exited with status 0 ]
    [ Aug 23 09:49:23 Stopping because service disabled. ]
    [ Aug 23 09:49:23 Executing stop method (null) ]
    [ Aug 23 10:04:13 Enabled. ]
    [ Aug 23 10:04:20 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug 23 10:04:20 Method "start" exited with status 0 ]
    [ Aug 23 10:06:57 Enabled. ]
    [ Aug 23 10:07:04 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug 23 10:07:05 Method "start" exited with status 0 ]
    [ Oct 23 08:43:05 Enabled. ]
    [ Oct 23 08:43:14 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Oct 23 08:43:14 Method "start" exited with status 0 ]
    [ Nov 21 14:51:50 Enabled. ]
    [ Nov 21 14:51:58 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Nov 21 14:51:59 Method "start" exited with status 0 ]
    [ Nov 29 15:33:19 Enabled. ]
    [ Nov 29 15:33:26 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Nov 29 15:33:27 Method "start" exited with status 0 ]
    [ Dec 18 13:37:35 Enabled. ]
    [ Dec 18 13:37:42 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Dec 18 13:37:42 Method "start" exited with status 0 ]
    [ Dec 19 08:48:07 Enabled. ]
    [ Dec 19 08:48:14 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Dec 19 08:48:14 Method "start" exited with status 0 ]
    [ Dec 19 08:56:07 Enabled. ]
    [ Dec 19 08:56:14 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Dec 19 08:56:14 Method "start" exited with status 0 ]
    [ Dec 19 08:57:41 Enabled. ]
    [ Dec 19 08:57:48 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Dec 19 08:57:48 Method "start" exited with status 0 ]
    [ Feb  1 10:39:48 Enabled. ]
    [ Feb  1 10:39:55 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Feb  1 10:39:56 Method "start" exited with status 0 ]
    [ Feb  8 11:12:50 Enabled. ]
    [ Feb  8 11:12:57 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Feb  8 11:12:57 Method "start" exited with status 0 ]
    [ Jun  6 10:08:34 Enabled. ]
    [ Jun  6 10:08:44 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Jun  6 10:08:46 Method "start" exited with status 0 ]
    [ Aug  8 09:45:48 Enabled. ]
    [ Aug  8 09:46:35 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug  8 09:46:36 Method "start" exited with status 0 ]
    [ Aug 14 17:53:59 Enabled. ]
    [ Aug 14 17:54:06 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug 14 17:54:06 Method "start" exited with status 0 ]
    [ Aug 14 17:55:20 Enabled. ]
    [ Aug 14 17:55:27 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug 14 17:55:27 Method "start" exited with status 0 ]
    [ Aug 18 15:19:17 Enabled. ]
    [ Aug 18 15:19:24 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug 18 15:19:24 Method "start" exited with status 0 ]
    [ Aug 29 12:56:48 Enabled. ]
    [ Aug 29 12:56:55 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug 29 12:56:55 Method "start" exited with status 0 ]
    [ Oct 13 11:59:43 Enabled. ]
    [ Oct 13 11:59:50 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Oct 13 11:59:51 Method "start" exited with status 0 ]
    [ Feb 26 16:47:00 Enabled. ]
    [ Feb 26 16:47:07 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Feb 26 16:47:07 Method "start" exited with status 0 ]
    [ Apr 14 15:54:17 Enabled. ]
    [ Apr 14 15:54:24 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Apr 14 15:54:25 Method "start" exited with status 0 ]
    [ Aug 28 08:49:58 Stopping because service disabled. ]
    [ Aug 28 08:49:58 Executing stop method (null) ]
    [ Aug 28 08:52:54 Enabled. ]
    [ Aug 28 08:53:01 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug 28 08:53:01 Method "start" exited with status 0 ]
    [ Aug 28 09:08:20 Stopping because service disabled. ]
    [ Aug 28 09:08:20 Executing stop method (null) ]
    [ Sep  1 16:05:36 Enabled. ]
    [ Sep  1 16:06:05 Executing start method ("/lib/svc/method/fs-local") ]
    [ Sep  1 16:06:05 Method "start" exited with status 0 ]
    [ Sep  1 19:29:19 Enabled. ]
    [ Sep  1 19:29:27 Executing start method ("/lib/svc/method/fs-local") ]
    [ Sep  1 19:29:27 Method "start" exited with status 0 ]
    [ Sep  1 23:04:00 Enabled. ]
    [ Sep  1 23:04:08 Executing start method ("/lib/svc/method/fs-local") ]
    [ Sep  1 23:04:08 Method "start" exited with status 0 ]
    [ Sep  2 00:23:41 Stopping because service disabled. ]
    [ Sep  2 00:23:42 Executing stop method (null) ]
    [ Sep  2 00:46:46 Enabled. ]
    [ Sep  2 00:46:53 Executing start method ("/lib/svc/method/fs-local") ]
    [ Sep  2 00:46:54 Method "start" exited with status 0 ]
    [ Sep 26 12:04:34 Enabled. ]
    [ Sep 26 12:04:41 Executing start method ("/lib/svc/method/fs-local") ]
    [ Sep 26 12:04:42 Method "start" exited with status 0 ]
    [ Oct 20 13:05:09 Stopping because service disabled. ]
    [ Oct 20 13:05:10 Executing stop method (null) ]
    [ Oct 20 13:24:01 Enabled. ]
    [ Oct 20 13:24:08 Executing start method ("/lib/svc/method/fs-local") ]
    [ Oct 20 13:24:09 Method "start" exited with status 0 ]
    [ Oct 21 16:30:59 Enabled. ]
    [ Oct 21 16:31:09 Executing start method ("/lib/svc/method/fs-local") ]
    [ Oct 21 16:31:09 Method "start" exited with status 0 ]
    [ Oct 21 16:32:42 Stopping because service disabled. ]
    [ Oct 21 16:32:42 Executing stop method (null) ]
    [ Oct 21 16:42:05 Enabled. ]
    [ Oct 21 16:42:15 Executing start method ("/lib/svc/method/fs-local") ]
    [ Oct 21 16:42:15 Method "start" exited with status 0 ]
    [ Oct 21 17:27:37 Enabled. ]
    [ Oct 21 17:27:47 Executing start method ("/lib/svc/method/fs-local") ]
    [ Oct 21 17:27:47 Method "start" exited with status 0 ]
    [ Oct 22 11:08:34 Enabled. ]
    [ Oct 22 11:08:43 Executing start method ("/lib/svc/method/fs-local") ]
    [ Oct 22 11:08:43 Method "start" exited with status 0 ]
    [ Dec  9 09:17:33 Stopping because service disabled. ]
    [ Dec  9 09:17:33 Executing stop method (null) ]
    [ Dec 10 11:47:14 Enabled. ]
    [ Dec 10 11:47:24 Executing start method ("/lib/svc/method/fs-local") ]
    [ Dec 10 11:47:25 Method "start" exited with status 0 ]
    [ Dec 15 10:52:25 Stopping because service disabled. ]
    [ Dec 15 10:52:25 Executing stop method (null) ]
    [ Dec 15 12:42:16 Enabled. ]
    [ Dec 15 12:42:26 Executing start method ("/lib/svc/method/fs-local") ]
    Abort - core dumped
    WARNING: /usr/sbin/zfs mount -a failed: exit status 134
    [ Dec 15 12:42:28 Method "start" exited with status 95 ]
    [ Dec 15 13:13:19 Enabled. ]
    [ Dec 15 13:13:28 Executing start method ("/lib/svc/method/fs-local") ]
    Abort - core dumped
    WARNING: /usr/sbin/zfs mount -a failed: exit status 134
    [ Dec 15 13:13:30 Method "start" exited with status 95 ]
    [ Dec 15 13:58:51 Enabled. ]
    [ Dec 15 13:59:00 Executing start method ("/lib/svc/method/fs-local") ]
    Abort - core dumped
    WARNING: /usr/sbin/zfs mount -a failed: exit status 134
    [ Dec 15 13:59:03 Method "start" exited with status 95 ]
    [ Dec 15 15:34:41 Leaving maintenance because clear requested. ]
    [ Dec 15 15:34:41 Enabled. ]
    [ Dec 15 15:34:41 Executing start method ("/lib/svc/method/fs-local") ]
    Abort - core dumped
    WARNING: /usr/sbin/zfs mount -a failed: exit status 134
    [ Dec 15 15:34:43 Method "start" exited with status 95 ]
    bash-3.00# /usr/sbin/zfs mount -a
    internal error: Unknown error
    Abort (core dumped)
    Before that I've installed the patch cluster with halted solaris zones which produced following output:
    Application of patches finished : 2010.12.15 12:08:43
    Following patches were applied :
    119254-77 119812-10 125719-31 140159-03 142911-01
    141588-04 119900-11 125731-05 142292-01 142933-02
    142251-02 119906-16 126206-05 141444-09 142909-17
    118666-28 120460-17 126363-08 141500-07 143140-04
    118667-28 121012-03 126365-16 141502-02 143502-01
    118777-16 121308-20 126868-04 141506-09 143506-01
    140860-02 122212-40 127724-02 141514-02 143615-02
    119059-56 122261-03 136998-09 141518-12 143731-01
    119213-23 122675-05 137000-07 141552-03 143733-01
    124628-10 123003-04 137080-05 141558-01 143977-01
    119252-29 123590-12 137147-06 141586-01 144053-04
    119280-23 123893-22 138195-04 141590-02 144106-01
    124188-03 124393-11 138822-07 141874-09 144254-01
    120199-15 124457-02 138826-07 141876-07 144488-04
    119534-19 124630-42 138880-02 142084-04 144492-01
    120272-28 125215-03 139099-04 142397-01 145124-01
    119757-18 125388-03 139620-01 142529-01 145796-01
    119783-15 125555-07
    Following patches were skipped :
    Patches already applied
    120900-04 119063-01 119810-05 123005-07 137093-01
    121133-02 119081-25 119986-03 124444-01 138866-03
    119317-01 119130-33 120061-02 124939-03 137137-09
    121296-01 123611-04 120201-05 124943-01 137871-02
    138215-01 140899-01 120292-02 124997-01 138181-01
    127884-01 122640-05 120329-02 125279-05 138361-01
    118712-23 126897-02 121975-01 125539-06 138373-02
    118918-24 127755-01 120719-02 125891-01 138647-01
    138217-01 125503-02 120830-06 126440-01 141016-01
    119578-30 125547-02 121095-02 126540-02 139555-08
    121453-02 140796-01 121606-04 127127-11 139967-01
    121453-02 120011-14 124171-07 136882-02 140455-01
    121118-16 139520-02 123630-03 137032-01 140563-01
    118833-36 119764-06
    Patches obsoleted by one or more patches already applied
    118731-01 124204-04 122660-10
    Patches not applicable to packages on the system
    121181-03 120410-33 121211-02 125533-15 143317-03
    119115-35 120412-11 122259-03 125541-06 143510-01
    119117-52 120414-27 122470-03 125952-20 143725-01
    119315-19 120543-21 122911-24 137004-08 143727-01
    119548-14 120739-06 122958-06 138387-01 143739-01
    119903-02 120811-09 123938-02 138824-07 144325-01
    120094-30 120849-04 125136-24 138876-01 145006-02
    120185-21 121104-11 125137-24 139986-01 145080-01
    119368-04 121136-02 125332-14 142244-02 145200-01
    120286-03
    Installation of patch set complete. PLEASE REBOOT THE SYSTEM.
    when i tried to reboot the system by init 6 the service iscsi hanged and produced syslog errors about a unkown iocall -> killed the process. System rebooted w/o errors.
    Does anyone have got an idea how to find out which patch or whatever do produce these errors?

    Hi,
    I'm having the same issue of diskmon core dumps after upgrading grid and database from 11.2.0.1 to 11.2.0.2 on Solaris 10 x86-64. Following this thread, I'm trying to deinstall the Oracle 11.2.0.1 grid home.
    I answered the following questions asked by the grid deinstall tool:
    1. ASM configuration was not detected in this Oracle home. Was ASM configured in this Oracle home (y|n): y
    2. Specify the ASM Diagnostic Destination: /db01/app/oracle/diag/asm/+ASM
    3. Specify the diskgroups that are managed by this ASM instance: DATA FRA
    4. De-configuring ASM will drop all the diskgroups at cleanup time. Do you want deconfig tool to drop the diskgroups:
    If I answer no to question 1, the tool exits with error.
    How do I answer question #4? (confused about this question of dropping diskgroups, it's already been upgraded)
    Deinstall of Oracle db Home 11.2.0.1 was successful.
    The +ASM instance has been upgraded and is running fine under 11.2.0.2 except for the excessive diskmon core dumps.
    Thanks,
    Lisa
    Edited by: user12018917 on Apr 19, 2011 10:21 AM

  • What about sites NOT created with Dreamweaver templates

    I am using Dreamweaver CS3 (no templates), and code most everything by hand (html, javascript and CSS). Would it work to have people in my office adding content to these existing sites with Contribute CS5 since there are NO Dreamweaver templates in place?

    Html webpages are editable in Contribute.

  • DskPercent not returned for ZFS filesystems?

    Hello.
    I'm trying to monitor the space usage of some ZFS filesystems on a Solaris 10 10/08 (137137-09) Sparc system with SNMP. I'm using the Systems Management Agent (SMA) agent.
    To monitor the stuff, I added the following to /etc/sma/snmp/snmpd.conf and restarted svc:/application/management/sma:default:
    # Bug in SMA?
    # Reporting - NET-SNMP, Solaris 10 - has a bug parsing config file for disk space.
    # -> http://forums.sun.com/thread.jspa?threadID=5366466
    disk /proc 42%  # Dummy Wert; wird fälschlicherweise ignoriert werden...
    disk / 5%
    disk /tmp 10%
    disk /apps 4%
    disk /data 3%Now I'm checking what I get via SNMP:
    --($ ~)-- snmpwalk -v2c -c public 10.0.1.26 dsk
    UCD-SNMP-MIB::dskIndex.1 = INTEGER: 1
    UCD-SNMP-MIB::dskIndex.2 = INTEGER: 2
    UCD-SNMP-MIB::dskIndex.3 = INTEGER: 3
    UCD-SNMP-MIB::dskIndex.4 = INTEGER: 4
    UCD-SNMP-MIB::dskPath.1 = STRING: /
    UCD-SNMP-MIB::dskPath.2 = STRING: /tmp
    UCD-SNMP-MIB::dskPath.3 = STRING: /apps
    UCD-SNMP-MIB::dskPath.4 = STRING: /data
    UCD-SNMP-MIB::dskDevice.1 = STRING: /dev/md/dsk/d200
    UCD-SNMP-MIB::dskDevice.2 = STRING: swap
    UCD-SNMP-MIB::dskDevice.3 = STRING: apps
    UCD-SNMP-MIB::dskDevice.4 = STRING: data
    UCD-SNMP-MIB::dskMinimum.1 = INTEGER: -1
    UCD-SNMP-MIB::dskMinimum.2 = INTEGER: -1
    UCD-SNMP-MIB::dskMinimum.3 = INTEGER: -1
    UCD-SNMP-MIB::dskMinimum.4 = INTEGER: -1
    UCD-SNMP-MIB::dskMinPercent.1 = INTEGER: 5
    UCD-SNMP-MIB::dskMinPercent.2 = INTEGER: 10
    UCD-SNMP-MIB::dskMinPercent.3 = INTEGER: 4
    UCD-SNMP-MIB::dskMinPercent.4 = INTEGER: 3
    UCD-SNMP-MIB::dskTotal.1 = INTEGER: 25821143
    UCD-SNMP-MIB::dskTotal.2 = INTEGER: 7150560
    UCD-SNMP-MIB::dskTotal.3 = INTEGER: 0
    UCD-SNMP-MIB::dskTotal.4 = INTEGER: 0
    UCD-SNMP-MIB::dskAvail.1 = INTEGER: 13584648
    UCD-SNMP-MIB::dskAvail.2 = INTEGER: 6471520
    UCD-SNMP-MIB::dskAvail.3 = INTEGER: 0
    UCD-SNMP-MIB::dskAvail.4 = INTEGER: 0
    UCD-SNMP-MIB::dskUsed.1 = INTEGER: 11978284
    UCD-SNMP-MIB::dskUsed.2 = INTEGER: 679040
    UCD-SNMP-MIB::dskUsed.3 = INTEGER: 0
    UCD-SNMP-MIB::dskUsed.4 = INTEGER: 0
    UCD-SNMP-MIB::dskPercent.1 = INTEGER: 47
    UCD-SNMP-MIB::dskPercent.2 = INTEGER: 9
    UCD-SNMP-MIB::dskPercent.3 = INTEGER: 0
    UCD-SNMP-MIB::dskPercent.4 = INTEGER: 0
    UCD-SNMP-MIB::dskPercentNode.1 = INTEGER: 9
    UCD-SNMP-MIB::dskPercentNode.2 = INTEGER: 0
    UCD-SNMP-MIB::dskPercentNode.3 = INTEGER: 0
    UCD-SNMP-MIB::dskPercentNode.4 = INTEGER: 0
    UCD-SNMP-MIB::dskErrorFlag.1 = INTEGER: noError(0)
    UCD-SNMP-MIB::dskErrorFlag.2 = INTEGER: noError(0)
    UCD-SNMP-MIB::dskErrorFlag.3 = INTEGER: noError(0)
    UCD-SNMP-MIB::dskErrorFlag.4 = INTEGER: noError(0)
    UCD-SNMP-MIB::dskErrorMsg.1 = STRING:
    UCD-SNMP-MIB::dskErrorMsg.2 = STRING:
    UCD-SNMP-MIB::dskErrorMsg.3 = STRING:
    UCD-SNMP-MIB::dskErrorMsg.4 = STRING: As expected, dskPercent.1 and dskPercent.2 (ie. */* and */tmp*) returned good values. But why did Solaris/SNMP/??? return 0 for dskPercent.3 (*/apps*) and dskPercent.4 (*/data*)? Those directories are on two ZFS which are on seperate zpools:
    --($ ~)-- zpool list
    NAME   SIZE   USED  AVAIL    CAP  HEALTH  ALTROOT
    apps  39.2G  20.4G  18.9G    51%  ONLINE  -
    data   136G   121G  15.2G    88%  ONLINE  -
    --($ ~)-- zfs list apps data
    NAME   USED  AVAIL  REFER  MOUNTPOINT
    apps  20.4G  18.3G    20K  /apps
    data   121G  13.1G   101K  /dataOr is it supposed to be that way? I'm pretty much confused, because I found some blog posting by a guy called asyd at http://sysadmin.asyd.net/home/en/blog/asyd/zfs+snmp. Copying from there:
    snmpwalk -v2c -c xxxx katsuragi.global.asyd.net UCD-SNMP-MIB::dskTable
    UCD-SNMP-MIB::dskPath.1 = STRING: /
    UCD-SNMP-MIB::dskPath.2 = STRING: /home
    UCD-SNMP-MIB::dskPath.3 = STRING: /data/pkgsrc
    UCD-SNMP-MIB::dskDevice.1 = STRING: /dev/dsk/c1d0s0
    UCD-SNMP-MIB::dskDevice.2 = STRING: data/home
    UCD-SNMP-MIB::dskDevice.3 = STRING: data/pkgsrc
    UCD-SNMP-MIB::dskTotal.1 = INTEGER: 1017935
    UCD-SNMP-MIB::dskTotal.2 = INTEGER: 0
    UCD-SNMP-MIB::dskTotal.3 = INTEGER: 0
    UCD-SNMP-MIB::dskAvail.1 = INTEGER: 755538
    UCD-SNMP-MIB::dskAvail.2 = INTEGER: 0
    UCD-SNMP-MIB::dskAvail.3 = INTEGER: 0
    UCD-SNMP-MIB::dskPercent.1 = INTEGER: 21
    UCD-SNMP-MIB::dskPercent.2 = INTEGER: 18
    UCD-SNMP-MIB::dskPercent.3 = INTEGER: 5What I find confusing, are his dskPercent.2 and dskPercent.3 outputs - for him, he got back dskPercent for what seems to be directories on ZFS filesystems.
    Because of that I'm wondering about how it is supposed to be - should Solaris return dskPercent values for ZFS?+
    Thanks a lot,
    Alexander

    I don't have the ability to test out my theory, but I suspect that you are using the default mount created for the zpools you've created (apps & data) as opposed to specific ZFS files systems, which is what the asyd blog shows.
    Remember, the elements reported on in the asyd blog ARE zfs file systems; they are not just directories. They are indeed mountpoints, and it is reporting the utilization of those zfs file systems in the pool ("data") on which they are constructed. In the case of /home, the administrator has specifically set the mountpoint of the ZFS file system to be /home instead of the default /data/home.
    To test my theory, instead of using your zpool default mount point, try creating a zfs file system under each of your pools and using that as the entry point for your application to write data into the zpools. I suspect you will be rewarded with the desired result: reporting of "disk" (really, pool) percent usage.

  • Need Best Practice for creating BE in ZFS boot environment with zones

    Good Afternoon -
    I have a Sparc system with ZFS Root File System and Zones. I need to create a BE for whenever we do patching or upgrades to the O/S. I have run into issues when testing booting off of the newBE where the zones did not show up. I tried to go back to the original BE by running the luactivate on it and received errors. I did a fresh install of the O/S from cdrom on a ZFS filesystem. Next ran the following commands to create the zones, and then create the BE, then activate it and boot off of it. Please tell me if there are any steps left out or if the sequence was incorrect.
    # zfs create –o canmount=noauto rpool/ROOT/S10be/zones
    # zfs mount rpool/ROOT/S10be/zones
    # zfs create –o canmount=noauto rpool/ROOT/s10be/zones/z1
    # zfs create –o canmount=noauto rpool/ROOT/s10be/zones/z2
    # zfs mount rpool/ROOT/s10be/zones/z1
    # zfs mount rpool/ROOT/s10be/zones/z2
    # chmod 700 /zones/z1
    # chmod 700 /zones/z2
    # zonecfg –z z1
    Myzone: No such zone configured
    Use ‘create’ to begin configuring a new zone
    Zonecfg:myzone> create
    Zonecfg:myzone> set zonepath=/zones/z1
    Zonecfg:myzone> verify
    Zonecfg:myzone> commit
    Zonecfg:myzone>exit
    # zonecfg –z z2
    Myzone: No such zone configured
    Use ‘create’ to begin configuring a new zone
    Zonecfg:myzone> create
    Zonecfg:myzone> set zonepath=/zones/z2
    Zonecfg:myzone> verify
    Zonecfg:myzone> commit
    Zonecfg:myzone>exit
    # zoneadm –z z1 install
    # zoneadm –z z2 install
    # zlogin –C –e 9. z1
    # zlogin –C –e 9. z2
    Output from zoneadm list -v:
    # zoneadm list -v
    ID NAME STATUS PATH BRAND IP
    0 global running / native shared
    2 z1 running /zones/z1 native shared
    4 z2 running /zones/z2 native shared
    Now for the BE create:
    # lucreate –n newBE
    # zfs list
    rpool/ROOT/newBE 349K 56.7G 5.48G /.alt.tmp.b-vEe.mnt <--showed this same type mount for all f/s
    # zfs inherit -r mountpoint rpool/ROOT/newBE
    # zfs set mountpoint=/ rpool/ROOT/newBE
    # zfs inherit -r mountpoint rpool/ROOT/newBE/var
    # zfs set mountpoint=/var rpool/ROOT/newBE/var
    # zfs inherit -r mountpoint rpool/ROOT/newBE/zones
    # zfs set mountpoint=/zones rpool/ROOT/newBE/zones
    and did it for the zones too.
    When ran the luactivate newBE - it came up with errors, so again changed the mountpoints. Then rebooted.
    Once it came up ran the luactivate newBE again and it completed successfully. Ran the lustatus and got:
    # lustatus
    Boot Environment Is Active Active Can Copy
    Name Complete Now On Reboot Delete Status
    s10s_u8wos_08a yes yes no no -
    newBE yes no yes no -
    Ran init 0
    ok boot -L
    picked item two which was newBE
    then boot.
    Came up - but df showed no zones, zfs list showed no zones and when cd into /zones nothing there.
    Please help!
    thanks julie

    The issue here is that lucreate add's an entry to the vfstab in newBE for the zfs filesystems of the zones. You need to lumount newBE /mnt then edit /mnt/etc/vfstab and remove the entries for any zfs filesystems. Then if you luumount it you can continue. It's my understanding that this has been reported to Sun, and, the fix is in the next release of Solaris.

  • Oc 11gR1 update 3: doesn't show ZFS file systems created on brownfield zone

    Subject is a pretty concise description here. I have several brownfield Solaris 10U10 containers running on M5000s, and I have delegated three zpool to each container for use by Oracle. Below is relevant output from zonecfg export for one of these containers. They were all built in the same manner, then placed under management by OC. (Wish I'd been able to build them as green field with Ops Center, but there just wasn't enough time to learn how to configure OpsCenter the way I needed to use it.)
    set name=Oracle-DB-Instance
    set type=string
    set value="Oracle e-Business Suite PREPROD"
    end
    add dataset
    set name=PREPRODredoPOOL
    end
    add dataset
    set name=PREPRODarchPOOL
    end
    add dataset
    set name=PREPRODdataPOOL
    end
    The problem is, none of the file systems built on these delegated pools in the container appear in the Ops Center File System Utilization charts. Does anyone have a suggestion for how to get OC to monitor the file systems in the zone?
    Here's the output from zfs list within the zone described by the zonecfg output above:
    [root@acdpreprod ~]# zfs list
    NAME USED AVAIL REFER MOUNTPOINT
    PREPRODarchPOOL 8.91G 49.7G 31K none
    PREPRODarchPOOL/d05 8.91G 49.7G 8.91G /d05
    PREPRODdataPOOL 807G 364G 31K none
    PREPRODdataPOOL/d02 13.4G 36.6G 13.4G /d02
    PREPRODdataPOOL/d03 782G 364G 782G /d03
    PREPRODdataPOOL/d06 11.4G 88.6G 11.4G /d06
    PREPRODredoPOOL 7.82G 3.93G 31K none
    PREPRODredoPOOL/d04 7.82G 3.93G 7.82G /d04
    None of the file systems in the delegated datasets appear in Ops Center for this zone. Are there any suggestions for how I correct this?

    Do you mean adopt the zone? That requires the zone be halted and it also says something about copying all file systems to the pool created for the zone. Of the 12 zones I have (four on each of three M5000s), seven of them are already in "production" status, and four of those seven now support 7x24 world-wide operations. A do-over is not an option here.

  • What is the best way to backup ZFS filesystem on solaris 10?

    Normally on Linux environment, I'd use mondorescue to create image (full & incremental) so it can be easily restored (full or file/folders) to a new similar server environment for restore purposes in case of disaster.
    I'd like to know the best way to backup a ZFS filesystem to a SAN storage and to restore it from there with minimal downtime. Preferrably with tools already available on Solaris 10.
    Thanks.

    the plan is to backup whole OS, and configuration files
    2 servers to be backed up
    server A zpool:
    - rootpool
    - usr
    - usrtmp
    server B zpool:
    - rootpool
    - usr
    - usrtmp
    if we were to cut hardware cost, it is possible to back up to samba share?
    any suggestions?

  • Mount options for ZFS filesystem on Solaris 10

    Do you know some recomendation
    about mount options for SAP on Oracle
    with data on ZFS filesystem?
    Also recomended block size required.
    We assume that file system with datafiles has 8kb block size
    and offline redologs has default (128kB).
    But what about ONLINE REDOLOGS?
    Best regards
    Andy

    SUN Czech installed new production HW for one Czech customer with ZFS filesystem on data-, redo- and archivelog files.
    Now we have performance problem and currently there is no SAP recomendation
    for ZFS file system.
    The HW which are by benchmark about tvice power has worst responses than
    old hardware.
    a) There is bug in Solaris 10 - ZFS buffers once allocated are not released
        (generally we do not want to use buffering due to prevence of double
         buffering)
    b) ZFS buffers takes about 20GB (32GB total) of memory on DB server
    and we are not able to define huge shared pool and db cache. (it may be possible
    to set special parameter in /etc/system to reduce maximum size of ZFS buffers to e.g. 4GB )
    c) We are looking for proven mount option for ZFS to enable asynchronious/concurent io for database filesystems
    d) There is no proven clear answer for support of ZFS/SOLARIS/Oracle/SAP.
    SAP says It is Oracle problem, Oracle does not certify filesystems from Jan2007
    any more and says ask your OS provider and SUN looks happy, but performance
    goes down and it is not so funny for system with 1TG DB with over 30GB grow
    per month.
    Andy

  • Does SAP support Solaris 10 ZFS filesystem when using DB2 V9.5 FP4?

    Hi,
    I'm installing NW7 (BI usage). SAPINST has failed in the step "ABAP LOAD due to the DB2 error message
    "Unsupported file system type zfs for Direct I/O". It appears my Unix Admin must have decided to set these filesystems as ZFS on this new server.
    I  have several questions requiring your expertise.
    1) Does SAP support ZFS filesystems on Solaris 10 (SPARC hardware)? I can not find any reference in SDN or Service Market Place? Any reference will be much appreciated.
    2) How can I confirm my sapdata fielsystems are ZFS?
    3) What actions do you recommend for me to resolve the SAPINST errors? Do I follow the note "Note 995050 - DB6: NO FILE SYSTEM CACHING for Tablespaces" to disable "Direct I/O" for all DB2 tablespaces? I have seen Markus Doehr's forum Link:[ thread|Re: DB2 on Solaris x64 - ZFS as filesystem possible?; but it does not state exactly how he overcame the error.
    regards
    Benny

    Hi Frank,
    Thanks for your input.
    I have also found  the command "zfs list" that would display any ZFS filesystems.
    We have also gone back to UFS as the ZFS deployment schedule does not meet this particular SAP BW implementation timeline.
    Has any one come across any SAP statement that states NW7 can be deployed with ZFS for DB2 database on Solaris SPARC platform. If not, I'll open an OSS message.
    regards
    Benny

  • Solaris 10 Cluster 3.2 with  2 zones in a failover scenario

    Hi
    Looking for the best way to set up things for the following scenario
    I have 2 M5000 servers with internal storage and a 6140 array for shared storage
    I need to create 2 zones on each in a failover scenario (active /standby)
    On Server1 3 out of 4 cpus for Oracle Database Server 11g and 1 out of 4 cpus for Oracle Application Server
    On Server2 3 out of 4 cpus for Oracle Application Server and 1 out of 4 cpus for Oracle Database Server 11g
    Database files will be placed on the shared storage. In case of failure of Server1 Oracle Database will fail over to Server2 and in case Server2 is down Oracle Application Server will fail to Server1.
    Would a zone cluster using clzonecluster be better?if yes how can i achieve the difference in cpu power in case of failure.
    where is best to keep the zone root path on the internal storage or on the shared storage?
    What about the swap space for both zones?
    Better use exclusive ips or shared will be fine?
    Will it be better to have sparse zone installation for the zone or do the whole thing?
    What is the best way to achieve the cpu assignments needed and how much should be left for the global zone?
    Thanks in advance
    vangelis

    Hi Vangelis,
    Building a cluster, requires some planning and understanding the concepts.
    A good start would be reading some of the documents linked to in this url: http://docs.sun.com/app/docs/doc/819-2969/gcbkf?a=view
    Regards,
    Davy

  • Sun Live Upgrade with local zones Solaris 10

    I have M800 server running global root (/) fs on local disk and running 6 local zones on another local disk. I am running solaris 5.10 8/07.
    I used live upgrade to patch the system and created new BE (lucreate). Both root fs are mirror as RAID-1.
    When I ran lucreate, it copies all 6 local zones root fs to the global root fs and failed no enogh space.
    What is the best procedure to use lu with local zones.
    Note: I used lu with global zone only, and worked without any problem.
    regards,

    I have been trying to use luupgrade for Solaris10 on Sparc, 05/09 -> 10/09.
    lucreate is successful, but luactivate directs me to install 'the rest of the packages' in order to make the BE stable enough to activate. I try to find the packages indicated , but find only "virtual packages" which contain only pkgmap.
    I installed upgrade 6 on a spare disk to make sure my u7 installation was not defective, but got similar results.
    I got beyond luactivate on x86 a while ago, but had other snags which I left unattended.

  • Monitoring services in zones with Solaris container Manager

    I need to know how to Manage Solaris services (SMF) in sparse zone with Solaris container manager.
    I have navigated all the documentation and I have not found any clue.
    I installed the Sun management center (SMC) server on a server box and the agents on others. I can manage the SMF of the global zone by drilling down via the console GUI. But having access to the container manager, I have to go via https connection. And drilling down on the zone did not reveal that SMF can be monitored.
    Please if you have any idea, share it with me.

    Hi,
    check these:
    - Version of webserver latest is 3.1
    # smcwebserver -V
    - check webconsole is started and running
    # smcwebserver status
    Sun Java(TM) Web Console is running
    # svcs webconsole
    STATE STIME FMRI
    online 19:38:06 svc:/system/webconsole:console
    # svcs -pl webconsole
    fmri svc:/system/webconsole:console
    name java web console
    enabled true
    state online
    next_state none
    state_time Wed Feb 10 19:38:06 2010
    logfile /var/svc/log/system-webconsole:console.log
    restarter svc:/system/svc/restarter:default
    contract_id 64
    dependency require_all/none svc:/milestone/network (online)
    dependency require_all/refresh svc:/milestone/name-services (online)
    dependency require_all/none svc:/system/filesystem/local (online)
    dependency optional_all/none svc:/system/filesystem/autofs (online) svc:/network/nfs/client (online)
    dependency require_all/none svc:/system/system-log (online)
    process 843 /usr/java/bin/java -server -Xmx128m -XX:+UseParallelGC -XX:ParallelGCThreads=4
    - check port 6789 is listen mode
    # netstat -an | grep 6789
    *.6789 *.* 0 0 49152 0 LISTEN
    if the output show
    localhost .6789 *.* 0 0 49152 0 LISTEN
    than do these:
    - check that the tcp_listen of webconsole service is true, default is false
    # svccfg -s webconsole listprop options/tcp_listen
    options/tcp_listen boolean false
    # svcadm disable svc:/system/webconsole:console
    # svccfg -s webconsole setprop options/tcp_listen=true
    # svccfg -s webconsole listprop options/tcp_listen
    options/tcp_listen boolean true
    # svcadm enable svc:/system/webconsole:console
    Regards

  • I want to use ZFS filesystem compression on a history database with read-only tablespaces.  What is the best way to get the datafiles compressed?

    I have 5 history databases that total about 2.2 Terabytes. They use about 15 filesystems, but I can start with 2 that are isolated to only one database.  I have asked for ZFS compression to be set on the filesystems, but since these are read-only tablespaces, I do not think any compression will happen.  Can I simply offline, copy and rename the files from one filesystem to the other to make the compression happen?

    just rsync the files to a compressed Zpool. do this using shadow migration, and you only loose access to the data for a few seconds.
    1) make new dataset with compression
    2) enable shadow migration between the new and old
    3) change the database to use the new location
    4) watch as data is automatically copied and compressed :-)
    the the down side, you need extra space to pull the off.

  • I have 4 equal oblong shape created with borders How do I go about knowing what size the selection area is so that I can crop an image to fit. I don't want to use paste in then adjust the bounding box to suit

    I have 4 equal oblong shape created with borders How do I go about knowing what size the selection area is so that I can crop an image to fit. I don't want to use paste in then adjust the bounding box to suit

    What do you mean a moderator

  • Cannot create guest virtual machine with Solaris 10 SPARC in Oracle VM 2.2

    Dear All,
    I try many times to create guest virtual machine with Solaris 10 SPARC but not get success.
    When I created it already and then I start the Power On button, it run normally but the Console button cannot click on (hide). After that the virtual machine running abit then it change status to Powered Off.
    Please kindly give advice for this.
    Thanks and regards,
    Heng

    I storgly doubt, VMware player is available for Solaris 10 SPARC version.
    Solaris 10 is not a comercial product, like ubuntu, windows, MAC etc... It's really for a server.
    That's the reason still I hate Solaris 11 as a desktop. You have very limited product/software.... This may change for Solaris 11, but U can't expect one for Solaris 10.

Maybe you are looking for

  • 'Error 49 bad bind variable' error message.

    When I copied a form that has been working from our Linux server to my PC and compiled it, I received many error messages like this in several procedures, functions and triggers: "Error 49 at line 69, column 10, bad bind variable 'edit_Dealer.corpora

  • Is there any way to import dvds/dvd tv seasons onto itunes and ur ipod?

    ???

  • FBICR3 - Intercompany reconciliation

    Dear Gurus, We have an issue with inter-company reconciliation. When we run FBICR3 txn the company data is pulling through but the partner data is not pulling because of which we can't make manual reconciliations. For example If Company code 1000 and

  • Comparing file content

    Good morning, I am trying to determine if records in one text file are a subset of records in a second file. For example the 1st file(A) contains the names jan eva john And the second(B) contains eva richard jan paul I've started by formulating the p

  • Trying to Clone new SSD, keeps 'Ejecting' itself mid-copy

    I have bought a Crucial M4 SSD Sata 2. I am using a macbook 5.1 Mid 2008 Unibody, running mountain lion. I downloaded 'SuperDuper' and have renamed my SSD to 'My Mac'. I Then attempt to copy all files using SuperDuper, from my old HD to the New, usin