Problem with exporting devices to non-global zone

Hi,
I've problem with exporting devices to my solaris zones (i try do add support to mount /dev/lofi/* in my non-global zone).
A create cfg for my zone.
Here it is:
$ zonecfg -z sapdev info
zonename: sapdev
zonepath: /export/home/zones/sapdev
brand: native
autoboot: true
bootargs:
pool:
limitpriv: default,sys_time
scheduling-class:
ip-type: shared
fs:
dir: /sap
special: /dev/dsk/c1t44d0s0
raw: /dev/rdsk/c1t44d0s0
type: ufs
options: []
net:
address: 194.29.128.45
physical: ce0
device
match: /dev/lofi/1
device
match: /dev/rlofi/1
device
match: /dev/lofi/2
device
match: /dev/rlofi/2
attr:
name: comment
type: string
value: "This is SAP developement zone"
global# lofiadm
Block Device File
/dev/lofi/1 /root/SAP_DB2_9_LUW.iso
/dev/lofi/2 /usr/tmp/fsfile
I reboot the non-global zone, even reboot global-zone, and after that, in sapdev zone, there is no /dev/*lofi/* files.
What i do wrong? Maybe I reduce my sol 10 u4 sparc instalation too much.
Can anybody help me?
Thanks for help,
Marek

I experienced the same problem on my system Sol 10 08/07.
Normally, when the zone enters the READY state during boot, it's zoneadmd will run devfsadm -z <zone>. In my understanding this is to create the necessary device files in ZONEPATH/dev.
This worked well until recently. Now only the directories are still created.
It seems as if devfsadm -z is broken. Somebody should issue a call to sun.
As a workaround you can easily copy the device files into the zone. It is important not to copy the symbolic link but the target.
# cp /dev/lofi/1 ZONEPATH/dev/lofi
Hope this helps,
Konstantin Gremliza

Similar Messages

  • Add tape device to non-global zone

    Hi,
    I have a SCSI attached Ultrium tape device attached and configured against the global zone.
    The /dev/rmt/0* definitions in the global zone are links to ../../devices/pci@2*
    I need to be able to use this tape device from the non-global zones.
    To enable this, I have done the following:
    zonecfg -z <zone name>
    add device
    set match=/dev/rmt/0
    end
    verify
    commit
    exit
    I repeated the above for /dev/rmt/0m and /dev/rmt/0mn
    Then I restarted the zone with the command:
    zoneadm -z <zone name> reboot
    After the reboot, I can see the device when using "mt -f /dev/rmt/0 status", but whenever I try to write a SAP brbackup to the new (initialised and not write protected) tape within the drive I get the following error:
    BR0278E Command output of 'LANG=C cd /oracle/<SID>/sapbackup && /usr/sap/<SID>/SYS/exe/run/brtools -f detach LANG=C cpio -iuvB .tape
    sh: /dev/rmt/0mn: cannot open
    BR0280I BRBACKUP time stamp: 2012-04-04 08.21.41
    BR0279E Return code from 'LANG=C cd /oracle/<SID>/sapbackup && /usr/sap/<SID>/SYS/exe/run/brtools -f detach LANG=C cpio -iuvB .tape.
    BR0359E Restore of /oracle/<SID>/sapbackup/.tape.hdr0 from /dev/rmt/0mn failed due to previous errors
    Have I created the device incorrectly, or does anyone have any ideas what could be the reason the write fails?
    Any help appreciated.
    Edited by: user11329299 on 04-Apr-2012 01:09

    Hi,
    Just to bring you up to speed, I have now fixed the issue.
    The resolution was all within the iniSID.sap file that the backup is using. I have changed a number of parameters within this file:
    1.     tape_copy_cmd = dd (was cpio)
    2.     rewind = "mt     -f $ rew; sleep 30" (was " mt -f $ rew")
    3.     rewind_offline = "mt -f $ offline; sleep 30" (was "mt -f $ offline")
    4.     tape_pos_cmd = "mt -f $ fsf $: sleep 30" (was "mt -f $ fsf $")
    5.     tape_size = 500G (was 18000M)
    After making those changes, the backup started from within DB13. I believe that the main culprit was the tape_copy_cmd, but the others were changed to allow the tape drive time to become online again after any query.

  • Add device cdrom non global zone

    Hi All ,
    how to add device cdrom to non global zone ?
    many thanks

    Hi David,
    Does your judgement that "you can definitely use a lofs mount to make /usr/local available inside the four zones" hold when /usr/local is an NFS mount in the global zone? James indicated that his /usr/local was an NFS mount, but I don't know if you targeted that specifically, or if you were just responding to the general case.
    I have not had any success in performing lofs mounts of NFS mounts and making them visible in a local zone. I don't have any problems making a lofs mount of a NFS mount work in the global zone, I just can't NFS mount anything in the global zone and make it accessible in a local zone.
    See http://www.opensolaris.org/jive/thread.jspa?threadID=2181&tstart=0 for a full description of what I tried and what happened. The short result is that when I try, accessing that directory in the local zone gives "I/O error" for file operations and "Not owner" for df.
    Thanks,
    Dab

  • What options do I have to patch the recommended patchset on Solaris 10 with a bunch of non-global zones?

    With the standard patching process(installcluster), it takes a looong time since each zone needs veridated. Any option that I can apply the patchset to the global zone only, then later upgrade the non-global zones?
    If possible, I'd like to use LU.

    You can use LU but it will depend of your system config. There are instructions in the README of the patchset to install it on an alternate boot environment (previously created using lucreate).
    If you plan to use LU, read the following docs first to avoid common issues:
    Solaris Live Upgrade Software Patch Requirements(Doc ID 1004881.1)
    List of currently unsupported Live Upgrade (LU) configurations (Doc ID 1396382.1)
    You can also use Parallel Patching feature to improve performance :
    https://blogs.oracle.com/patch/entry/zones_parallel_patching_feature_now
    Solaris 10 10/09: Zones Parallel Patching to ReducePatching Time (System Administration Guide: Oracle Solaris Containers…
    What you can't do is patch the global zone only and the non-global zones later (unless the zones are detached). It's a requirement that the global and non-global stay synchronize at all time (considering that they are sharing the same kernel).

  • Lucreate not working with ZFS and non-global zones

    I replied to this thread: Re: lucreate and non-global zones as to not duplicate content, but for some reason it was locked. So I'll post here... I'm experiencing the exact same issue on my system. Below is the lucreate and zfs list output.
    # lucreate -n patch20130408
    Creating Live Upgrade boot environment...
    Analyzing system configuration.
    No name for current boot environment.
    INFORMATION: The current boot environment is not named - assigning name <s10s_u10wos_17b>.
    Current boot environment is named <s10s_u10wos_17b>.
    Creating initial configuration for primary boot environment <s10s_u10wos_17b>.
    INFORMATION: No BEs are configured on this system.
    The device </dev/dsk/c1t0d0s0> is not a root device for any boot environment; cannot get BE ID.
    PBE configuration successful: PBE name <s10s_u10wos_17b> PBE Boot Device </dev/dsk/c1t0d0s0>.
    Updating boot environment description database on all BEs.
    Updating system configuration files.
    Creating configuration for boot environment <patch20130408>.
    Source boot environment is <s10s_u10wos_17b>.
    Creating file systems on boot environment <patch20130408>.
    Populating file systems on boot environment <patch20130408>.
    Temporarily mounting zones in PBE <s10s_u10wos_17b>.
    Analyzing zones.
    WARNING: Directory </zones/APP> zone <global> lies on a filesystem shared between BEs, remapping path to </zones/APP-patch20130408>.
    WARNING: Device <tank/zones/APP> is shared between BEs, remapping to <tank/zones/APP-patch20130408>.
    WARNING: Directory </zones/DB> zone <global> lies on a filesystem shared between BEs, remapping path to </zones/DB-patch20130408>.
    WARNING: Device <tank/zones/DB> is shared between BEs, remapping to <tank/zones/DB-patch20130408>.
    Duplicating ZFS datasets from PBE to ABE.
    Creating snapshot for <rpool/ROOT/s10s_u10wos_17b> on <rpool/ROOT/s10s_u10wos_17b@patch20130408>.
    Creating clone for <rpool/ROOT/s10s_u10wos_17b@patch20130408> on <rpool/ROOT/patch20130408>.
    Creating snapshot for <rpool/ROOT/s10s_u10wos_17b/var> on <rpool/ROOT/s10s_u10wos_17b/var@patch20130408>.
    Creating clone for <rpool/ROOT/s10s_u10wos_17b/var@patch20130408> on <rpool/ROOT/patch20130408/var>.
    Creating snapshot for <tank/zones/DB> on <tank/zones/DB@patch20130408>.
    Creating clone for <tank/zones/DB@patch20130408> on <tank/zones/DB-patch20130408>.
    Creating snapshot for <tank/zones/APP> on <tank/zones/APP@patch20130408>.
    Creating clone for <tank/zones/APP@patch20130408> on <tank/zones/APP-patch20130408>.
    Mounting ABE <patch20130408>.
    Generating file list.
    Finalizing ABE.
    Fixing zonepaths in ABE.
    Unmounting ABE <patch20130408>.
    Fixing properties on ZFS datasets in ABE.
    Reverting state of zones in PBE <s10s_u10wos_17b>.
    Making boot environment <patch20130408> bootable.
    Population of boot environment <patch20130408> successful.
    Creation of boot environment <patch20130408> successful.
    # zfs list
    NAME USED AVAIL REFER MOUNTPOINT
    rpool 16.6G 257G 106K /rpool
    rpool/ROOT 4.47G 257G 31K legacy
    rpool/ROOT/s10s_u10wos_17b 4.34G 257G 4.23G /
    rpool/ROOT/s10s_u10wos_17b@patch20130408 3.12M - 4.23G -
    rpool/ROOT/s10s_u10wos_17b/var 113M 257G 112M /var
    rpool/ROOT/s10s_u10wos_17b/var@patch20130408 864K - 110M -
    rpool/ROOT/patch20130408 134M 257G 4.22G /.alt.patch20130408
    rpool/ROOT/patch20130408/var 26.0M 257G 118M /.alt.patch20130408/var
    rpool/dump 1.55G 257G 1.50G -
    rpool/export 63K 257G 32K /export
    rpool/export/home 31K 257G 31K /export/home
    rpool/h 2.27G 257G 2.27G /h
    rpool/security1 28.4M 257G 28.4M /security1
    rpool/swap 8.25G 257G 8.00G -
    tank 12.9G 261G 31K /tank
    tank/swap 8.25G 261G 8.00G -
    tank/zones 4.69G 261G 36K /zones
    tank/zones/DB 1.30G 261G 1.30G /zones/DB
    tank/zones/DB@patch20130408 1.75M - 1.30G -
    tank/zones/DB-patch20130408 22.3M 261G 1.30G /.alt.patch20130408/zones/DB-patch20130408
    tank/zones/APP 3.34G 261G 3.34G /zones/APP
    tank/zones/APP@patch20130408 2.39M - 3.34G -
    tank/zones/APP-patch20130408 27.3M 261G 3.33G /.alt.patch20130408/zones/APP-patch20130408

    I replied to this thread: Re: lucreate and non-global zones as to not duplicate content, but for some reason it was locked. So I'll post here...The thread was locked because you were not replying to it.
    You were hijacking that other person's discussion from 2012 to ask your own new post.
    You have now properly asked your question and people can pay attention to you and not confuse you with that other person.

  • Problem to migrate a non-global zone to a different machine.

    Hi, recently, I had try to migrate a non-global zone to a different machine but it’s doesn’t work.
    1. First, this is the structure of my machine with my non-global zone:
    host1# uname -a
    SunOS testsolaris 5.11 snv_101b i86pc i386 i86pc
    host1# zfs list
    NAME USED AVAIL REFER MOUNTPOINT
    big-zone 1.71G 1.64G 20K /big-zone
    big-zone/export 1.71G 1.64G 22K /big-zone/export
    big-zone/export/big-zone 1.67G 1.64G 21K /big-zone/export/big-zon e
    big-zone/export/big-zone/ROOT 1.67G 1.64G 18K legacy
    big-zone/export/big-zone/ROOT/zbe 1.67G 1.64G 1.66G legacy
    big-zone/export/zonetest 41.8M 1.64G 21K /big-zone/export/zonetes t
    big-zone/export/zonetest/ROOT 41.8M 1.64G 18K legacy
    big-zone/export/zonetest/ROOT/zbe 41.8M 1.64G 1.66G /big-zone/export/zonetes t/root
    rpool 8.35G 7.28G 72K /rpool
    rpool/ROOT 6.86G 7.28G 18K legacy
    rpool/ROOT/opensolaris 6.86G 7.28G 6.73G /
    rpool/dump 575M 7.28G 575M -
    rpool/export 375M 7.28G 21K /export
    rpool/export/home 18K 7.28G 18K /export/home
    rpool/export/small-zone 375M 7.28G 21K /export/small-zone
    rpool/export/small-zone/ROOT 375M 7.28G 18K legacy
    rpool/export/small-zone/ROOT/zbe 375M 7.28G 375M legacy
    rpool/swap 575M 7.78G 56.8M -
    2. In second, I had detach my non-global zone “zonetest” whit this commands :
    host1# zoneadm –z zonetest halt
    host1# zoneadm –z zonetest detach
    3. In third, I had move my zonepath to my new host.
    host1# cd /big-zone/export
    host1# tar cf zonetest.tar zonetest
    host1# sftp jay@new-host
    host1# put zonetest.tar
    Uploading ….
    host1# quit
    4. Unpack my .tar file
    host2# cd /big-zone/export
    host2# tar xf zonetest.tar
    So, after this, I think that my zonepath is transfert to my new host.
    This is the structure of my new host :
    jay@alien:~$ uname -a
    SunOS alien 5.11 snv_101b i86pc i386 i86pc Solaris
    jay@alien:~$ zfs list
    NAME USED AVAIL REFER MOUNTPOINT
    rpool 18.3G 73.3G 72K /rpool
    rpool/ROOT 2.98G 73.3G 18K legacy
    rpool/ROOT/opensolaris 2.98G 73.3G 2.85G /
    rpool/dump 1023M 73.3G 1023M -
    rpool/export 13.3G 73.3G 19K /export
    rpool/export/home 13.3G 73.3G 19K /export/home
    rpool/export/home/jay 13.3G 73.3G 13.3G /export/home/jay
    rpool/swap 1023M 73.9G 321M -
    zdata 10.7G 80.8G 9.65G /zdata
    zdata/zones 1.08G 80.8G 18K /zdata/zones
    zdata/zones/zonetest 1.08G 80.8G 1.08G /big-zone/export/
    *I have a mountpoint to /big-zone/export
    5. I had try to configure my zone on my new host and I receive and error message:
    host2# zonecfg -z zonetest
    zonetest: No such zone configured
    Use 'create' to begin configuring a new zone.
    zonecfg:zonetest> create -a /big-zone/export/zonetest
    invalid path to detached zone
    zonecfg:zonetest>

    And my new big-zone (on the second host) show this in the /big-zone/export/zonetest folder :
    jay@alien:/zdata/zones# zfs list
    NAME USED AVAIL REFER MOUNTPOINT
    rpool 23.5G 68.0G 72K /rpool
    rpool/ROOT 6.31G 68.0G 18K legacy
    rpool/ROOT/opensolaris 6.31G 68.0G 6.18G /
    rpool/dump 1023M 68.0G 1023M -
    rpool/export 15.2G 68.0G 19K /export
    rpool/export/home 15.2G 68.0G 19K /export/home
    rpool/export/home/jay 15.2G 68.0G 15.2G /export/home/jay
    rpool/swap 1023M 68.6G 361M -
    zdata 11.6G 79.9G 10.7G /zdata
    zdata/zones 921M 79.9G 18K /zdata/zones
    zdata/zones/web 921M 79.9G 21K /zdata/zones/web
    zdata/zones/web/ROOT 921M 79.9G 18K legacy
    zdata/zones/web/ROOT/zbe 921M 79.9G 921M legacy
    zdata/zones/zonetest             54K  79.9G    18K  /big-zone/export/zonetest
    zdata/zones/zonetest/ROOT 36K 79.9G 18K legacy
    zdata/zones/zonetest/ROOT/zbe 18K 79.9G 18K legacy
    jay@alien:/zdata/zones/zonetest# pwd
    /zdata/zones/zonetest
    jay@alien:/zdata/zones/zonetest# ls -ls
    total 6
    3 drwxr-xr-x 2 root sys 2 Feb 8 2009 dev
    3 drwxr-xr-x 16 root root 19 Feb 8 2009 root
    jay@alien:/zdata/zones/zonetest# cd root
    jay@alien:/zdata/zones/zonetest/root# ls -ls
    total 52902
    1 lrwxrwxrwx 1 root root 9 Feb 1 20:29 bin -> ./usr/bin
    3 drwxr-xr-x 13 root sys 15 Feb 8 2009 dev
    11 drwxr-xr-x 55 root sys 168 Feb 8 2009 etc
    3 dr-xr-xr-x 2 root root 2 Jan 22 16:26 home
    15 drwxr-xr-x 9 root bin 241 Feb 4 2009 lib
    3 drwxr-xr-x 2 root sys 2 Jan 22 16:23 mnt
    3 dr-xr-xr-x 2 root root 2 Jan 22 16:26 net
    3 drwxr-xr-x 4 root sys 4 Jan 24 15:26 opt
    3 dr-xr-xr-x 2 root root 2 Jan 22 16:23 proc
    3 drwx------ 3 root root 7 Feb 6 2009 root
    5 drwxr-xr-x 2 root sys 47 Jan 22 16:24 sbin
    3 drwxr-xr-x 4 root root 4 Jan 22 16:23 system
    3 drwxrwxrwt 2 root sys 2 Feb 8 2009 tmp
    5 drwxr-xr-x 30 root sys 42 Feb 6 2009 usr
    3 drwxr-xr-x 32 root sys 32 Feb 6 2009 var
    52835 -rw-r--r-- 1 root root 42882560 Jan 22 16:35 webmin-1.441.pkg
    jay@alien:/zdata/zones/zonetest/root#
    I think my problem is there ...
    jay@alien:/big-zone/export/zonetest# pwd
    /big-zone/export/zonetest
    jay@alien:/big-zone/export/zonetest# ls -ls
    total 8
    2 ---------- 1 root root 114 Dec 31 1969 @LongLink
    3 drwxr-xr-x 2 root root 2 Feb 1 21:10 root
    3 drwx------ 4 root root 4 Feb 1 21:10 zonetest
    jay@alien:/big-zone/export/zonetest# cd zonetest/
    jay@alien:/big-zone/export/zonetest/zonetest# ls -ls
    total 6
    3 drwxr-xr-x 2 root sys 2 Feb 8 2009 dev
    3 drwxr-xr-x 4 root root 5 Feb 1 21:10 root
    jay@alien:/big-zone/export/zonetest/zonetest# cd root
    jay@alien:/big-zone/export/zonetest/zonetest/root# ls -ls
    total 7
    1 lrwxrwxrwx 1 root root 9 Feb 1 21:10 bin -> ./usr/bin
    3 drwxr-xr-x 4 root root 4 Jan 22 16:23 system
    3 drwxr-xr-x 23 root sys 28 Feb 1 21:11 usr
    I think I have a problem with my zfs mountpoint but I don't how to resolve this.
    Edited by: jaymachine on Feb 26, 2009 6:16 PM

  • Unable to add a device (e.g. /dev/cua0) to a non-global zone

    Hi,
    I've installed solaris 10u4 on a x86 machine with the latest patches, installed with the smpatch utility
    The history:
    I've installed solaris 10u3 without any patches, a quite minimum installation; I 've created a non-global zone, added a zfs dataset, added networking, add one serial device (/dev/cua0); installed hylafax from blastwave in the created zone using the attached modem on /dev/cua0, all was working fine, except some sendmail issues. Due to issues with samba, which I needed on this machine, I've tried to update the machine, after ending up in dependency hell, due the minimum installation, I gave up. I did a fresh install of solaris 10u4 instead also with latest patches applied with the smpatch utility, the I've created a new zone and want to add the device /dev/cua0 like in the s10u3 installation, but the device doesn't appear in the non-global zone, so I've installed hylafax in the global-zone temporary.
    The question, any ideas or workarrounds to bring the async device into a non-global zone again ?
    I'm not a newbie in nix like systems (several years with BSD and GNU/Linux), but for solaris I would classify myself as a newbie ;-)
    thanks in advance.

    Hmm. If that didn't work, then it's possible you're running into a different problem.
    But I checked again and this is the one I was thinking of. Toward the bottom, some patches are referenced. I suppose they won't hurt, but I'm worried you're seeing something related to the 'cua' device rather than the general problem of device creation.
    http://www.opensolaris.org/jive/thread.jspa?messageID=171187
    Darren

  • PHP in Solaris 10 and Non-Global Zones: Problem of performance?

    Hi friends
    We are feeling a poor performance with applications developed with PHP in Solaris 10, with non-global and global zones, while Intel platform (Xeon and Pentium), performance is very good. Difference between both platforms is about 200% aprox, one second in Intel to 9, 12 or 20 seconds in Solaris depending of model.
    Our tests were developed in:
    1. SF T2000 server Solaris 10 global zone
    2. SF T2000 server Solaris 10 non-global zone
    3. SF280R server Solaris 10 non-global zone
    4. V240 server with 1 GB memory, 1*US III-i 1.0 GHz and Solaris 9 (really this version for test and comparisons)
    5. V240 server with 8GB memory, 2*US III-i 1.5Ghz and Solaris 9 (really this version for test and comparisons too)
    Intel platforms were:
    1. Intel Pentium 4 2GHz 2GB memory, Linux Fedora and PHP 4.4.4
    2. Intel Xeon 2 core, 2.33GHz 2GB memory, Linux Fedora and PHP 4.4.3
    Versions of products are:
    1. Solaris 9 or Solaris 10
    2. PHP 4.4.7 downloaded from http://www.php.net/downloads.php
    3. Apache 2.0.59
    4. MySQL 4.1.15-log
    Our php compilation and installation were:
    ./configure --prefix=/usr/local/php-4.4.7 \
    --with-pear \
    --with-openssl=/usr/local/ssl \
    --with-gettext \
    --with-ldap=/usr/local \
    --with-iconv \
    --enable-ftp \
    --with-dom \
    --with-mime-magic \
    --enable-mbstring \
    --with-zlib \
    --enable-track-vars \
    --enable-sigchild \
    --disable-ctype \
    --disable-overload \
    --disable-tokenizer \
    --disable-posix \
    --with-gd \
    --with-apxs2=/usr/local/apache2.0.53/bin/apxs \
    --with-mysql  \
    --with-pgsql \
    --with-oci8=/oracle/product/9.2.0 \
    --with-oracle=/oracle/product/9.2.0  \
    --with-png-dir=/usr/local \
    --with-zlib-dir=/usr/local \
    --with-freetype-dir=/usr/local \
    --with-jpeg-dir=/usr/local
    make
    make install
    Questions:
    Is there any problem of PHP with SunFire T2000 servers or 64-bits platforms?
    Is there any flag of PHP would be use to compilarion PHP in 64-bits or multithread?
    I wait for any comments or suggestions about our problem with PHP compilation and performance in Solaris 10. Thanks a lot.
    Sergio.

    I presume you compiled php on the Sun server, was this done using gcc or the Sun One C compiler.
    If the latter then you can also use the flag: --enable-nonportable-atomics when you run configure                                                                                                                                                                                                                                                                                                                                                                                                   

  • Non-global zone installation problem

    I have created a non-global zone sqa45-zone as follows:
    zonecfg -z sqa45-zone
    zonecfg:sqa45-zone> info
    zonepath: /export/home/zones/sqa45-zone
    autoboot: false
    pool:
    inherit-pkg-dir:
    dir: /lib
    inherit-pkg-dir:
    dir: /platform
    inherit-pkg-dir:
    dir: /sbin
    inherit-pkg-dir:
    dir: /usr
    net:
    address: 10.4.105.105
    physical: bge0
    zonecfg:sqa45-zone> exit
    When I try to install the zone, this is the error message I see.
    sqa45# zoneadm -z sqa45-zone install
    could not verify net address=10.4.105.105 physical=bge0: No such device or address
    zoneadm: zone sqa45-zone failed to verify
    sqa45#
    I'm not sure how I can correct this. Any help is appreciated.
    Thanks.

    You havent specified a netmask.
    That might be causing it.
    You can specifiy it as address=10.4.105.105/24 or by a separate netmask entry.

  • Oracle 10 g non-global zones with asynchronous I/O

    Hi,
    I note that using direct I/O (by setting the forcedirectio while
    mounting the database file systems) and bypassing the file system
    cache may improve database performance significantly, but this should
    be done only for file systems in which database files and redo log
    files exist. If direct I/O is used and there is not enough database
    buffer cache, it may even decrease the performance by moving the
    problem from double buffering to a lack of database buffer cache. So,
    this performance tuning must be planned carefully, and the database
    buffer cache should be sized properly. The direct I/O option should
    not be used for other file systems used by other applications because
    they still need the UFS buffer cache.
    Now, I have Oracle database installed inside a non-global zone and I
    see a lot of Asynchronous I/O wait warnings in the Oracle Alert log
    file. Storage mount points with UFS filesystem contain the Oracle
    datafiles and redo log files. In addition, two Oracle datafiles of 10
    GB each reside on the local disks. The Oracle init.ora parameter to
    set asynchronous I/O for Oracle database files is
    FILESYSTEMIO_OPTIONS= SETALL.
    Although the above parameter was set during the database installation,
    the aiowait warnings don't seem to disappear.
    Can I use the "forcedirectio" option at the Operating System /etc/
    vfstab file for Oracle datafiles and redo log files?
    Or, should I just move the Oracle database files residing on the local
    disks to the external storage? Will this take care of aiowait warnings
    and if yes, how? The storage is a DAS.
    Regards
    Sandeep

    I presume you compiled php on the Sun server, was this done using gcc or the Sun One C compiler.
    If the latter then you can also use the flag: --enable-nonportable-atomics when you run configure                                                                                                                                                                                                                                                                                                                                                                                                   

  • Using a Fibre Channel HBA with a non-global zone.

    I am trying to let a non-global zone use a dual port HBA. Please note the goal is to use the HBA including the SAN devices, not just a device on the SAN. Does anyone know if and how this can be done?
    [root@global:/]# more /etc/release
                           Solaris 10 8/07 s10s_u4wos_12b SPARC ...
    [root@global:/]# zonecfg -z localzone info
    zonename: localzone
    zonepath: /zones/localzone
    brand: native
    autoboot: true
    bootargs:
    pool:
    limitpriv:
    scheduling-class:
    ip-type: shared
    net:
            address: x.x.x.x/24
            physical: qfe0
    device
            match: /dev/fc/fp[0-1]
    device
            match: /dev/cfg/c[1-2]
    device
            match: /dev/*dsk/c[1-2]*
    [root@global:/]# fcinfo hba-port
    HBA Port WWN: 210000e08b083b41
            OS Device Name: /dev/cfg/c1
            Manufacturer: QLogic Corp.
            Model: QLA2342
            Firmware Version: 3.3.24
            FCode/BIOS Version: No Fcode found
            Type: N-port
            State: online
            Supported Speeds: 1Gb 2Gb
            Current Speed: 2Gb
            Node WWN: 200000e08b083b41
    HBA Port WWN: 210100e08b283b41
            OS Device Name: /dev/cfg/c2
            Manufacturer: QLogic Corp.
            Model: QLA2342
            Firmware Version: 3.3.24
            FCode/BIOS Version: No Fcode found
            Type: N-port
            State: online
            Supported Speeds: 1Gb 2Gb
            Current Speed: 2Gb
            Node WWN: 200100e08b283b41
    [root@localzone:dev]# ls fc
    fp0  fp1
    [root@localzone:dev]# ls cfg
    c1  c2
    [root@localzone:dev]# ls dsk | grep s0
    c1t500601613021934Dd0s0
    c1t500601693021934Dd0s0
    c1t50060482D52D5608d0s0
    c1t50060482D52D5626d0s0
    c2t500601613021934Dd0s0
    c2t500601693021934Dd0s0
    c2t50060482D52D5608d0s0
    c2t50060482D52D5626d0s0
    [root@localzone:dev]# ls rdsk | grep s0
    c1t500601613021934Dd0s0
    c1t500601693021934Dd0s0
    c1t50060482D52D5608d0s0
    c1t50060482D52D5626d0s0
    c2t500601613021934Dd0s0
    c2t500601693021934Dd0s0
    c2t50060482D52D5608d0s0
    c2t50060482D52D5626d0s0
    [root@localzone:dev]# fcinfo hba-port
    No Adapters Found.

    You cannot present devices directly to the NGZ ( What a mouth/handful of words to say/type...sheesh! What's wrong with local zones, sun?)
    You can present filesystems and/or ZFS pools but not HBAs or other devices directly (AFAIK)

  • Sharing software package with non-global zone

    I've installed Solaris Studio 12.3 to a global zone on Solaris 11 following instructions from http://pkg-register.oracle.com which I want to make available to the non-global zones.  Do I need to install it independently for all zones, or can I export/share it?

    All files are normally installed in /opt/solarisstudio12.3. You can share it, but it is recommended to install it independently for all zones if you want different installed versions.

  • Dynamically adding a device to a non-global zone

    How can you add a new device to a non-global zone without having
    to reboot the zone ? Obviously using zonecfg and then rebooting the
    zone works but often rebooting the zone would be very user unfriendly.
    In particular we occasionally need to add a new disk mirror in the global
    zone and then let the non-global zone have access to it.

    So each zone has a /dev tree that's private in <zoneroot>/dev which gets lofs mounted to /dev in the zone. You can mknod a new device in here and it appears instantly in the zone. Use the same major and minor numbers that device has in the global zone. If it's a disk device, make sure to major both the block and character device.
    Of course, you want to edit the zone config to make it permanent (though I suspect it may actually be permanent if you mknod the device...)

  • Live upgrade - solaris 8/07 (U4) , with non-global zones and SC 3.2

    Dears,
    I need to use live upgrade for SC3.2 with non-global zones, solaris 10 U4 to Solaris 10 10/09 (latest release) and update the cluster to 3.2 U3.
    i dont know where to start, i've read lots of documents, but couldn't find one complete document to cover the whole process.
    i know that upgrade for solaris 10 with non-global zones is supported since my soalris 10 release, but i am not sure if its supported with SC.
    Appreciate your help

    Hi,
    I am not sure whether this document:
    http://wikis.sun.com/display/BluePrints/Maintaining+Solaris+with+Live+Upgrade+and+Update+On+Attach
    has been on the list of docs you found already.
    If you click on the download link, it won't work. But if you use the Tools icon on the upper right hand corner and click on attachements, you'll find the document. Its content is solely based on configurations with ZFS as root and zone root, but should have valuable information for other deployments as well.
    Regards
    Hartmut

  • Sun cluster 3.20, live upgrade with non-global zones

    I have a two node cluster with 4 HA-container resource groups holding 4 non-global zones running Sol 10 8/07 u4 which I would upgrade to sol10 u6 10/8. The root fileystem of the non-global zones is ZFS and on shared SAN disks so that can be failed over.
    For the LIve upgrade I need to convert the root ZFS to UFS which should be straight forward.
    The tricky stuff is going to be performing a live upgrade on non-global zones as their root fs is on the shared disk. I have a free internal disk on each of thenodes for ABE environments. But when I run the lucreate command is it going put the ABE of the zones on the internal disk as well or can i specifiy the location ABE for non-global zones. Ideally I want this to be shared disk
    Any assistance gratefully received

    Hi,
    I am not sure whether this document:
    http://wikis.sun.com/display/BluePrints/Maintaining+Solaris+with+Live+Upgrade+and+Update+On+Attach
    has been on the list of docs you found already.
    If you click on the download link, it won't work. But if you use the Tools icon on the upper right hand corner and click on attachements, you'll find the document. Its content is solely based on configurations with ZFS as root and zone root, but should have valuable information for other deployments as well.
    Regards
    Hartmut

Maybe you are looking for

  • Spry Drop Down menu issue in IE

    I created a menu bar with Spry and it works fine in Firefox but playing up in IE. The drop down menu starts displaying from the top of the page and not where the menu bar is. I have never come across this issue with Spry and cannot figure out whats w

  • HT2518 transfer from MacBook Pro to Hard Drive

    I used to store films from my Pc to an iomega HD multimedia, now I have this MacBook Pro and I can't transfer films to the same Hard Drive. Does it have a specific formate just to be used in a PC ?

  • Problem asking for username and password when deploy WebDynPro application

    Hi,     I try to deploy WebDynPro application to SAP JEE engine. When i deploy there is popup to put the user name and password. I have on idea that which username should be used as the username because when i'm installing the SAP CE7.2. They just as

  • How to make a setup file in java

    I have a softwave writting in java . I want to create a setup file. Please help me . thanks

  • Dynamic Columns in UWL  from the NWBC 3.0?

    Hi  People Is there anyone that know how to change the UWL so it uses Dynamic Columns for Worklist . I know that it is possible to do this in the portal by changing the XML's to the UWL. But at the moment i am trying to get this to work from the NWBC