File system: UFS2 end ZFS

I wanted to know if there is a way to mount a read-write file system ZFS and USF2.
Commercial solutions are also fine.
Thanks

At one point Apple was going to switch over from HFS+ to zfs but it didn't happen. There was a zfs project, so there might be some utilities specific to that. If not, I would use NFS, which doesn't give a rat's patootie about the file system. You have to have the same uid number on both for write access.

Similar Messages

  • Oc 11gR1 update 3: doesn't show ZFS file systems created on brownfield zone

    Subject is a pretty concise description here. I have several brownfield Solaris 10U10 containers running on M5000s, and I have delegated three zpool to each container for use by Oracle. Below is relevant output from zonecfg export for one of these containers. They were all built in the same manner, then placed under management by OC. (Wish I'd been able to build them as green field with Ops Center, but there just wasn't enough time to learn how to configure OpsCenter the way I needed to use it.)
    set name=Oracle-DB-Instance
    set type=string
    set value="Oracle e-Business Suite PREPROD"
    end
    add dataset
    set name=PREPRODredoPOOL
    end
    add dataset
    set name=PREPRODarchPOOL
    end
    add dataset
    set name=PREPRODdataPOOL
    end
    The problem is, none of the file systems built on these delegated pools in the container appear in the Ops Center File System Utilization charts. Does anyone have a suggestion for how to get OC to monitor the file systems in the zone?
    Here's the output from zfs list within the zone described by the zonecfg output above:
    [root@acdpreprod ~]# zfs list
    NAME USED AVAIL REFER MOUNTPOINT
    PREPRODarchPOOL 8.91G 49.7G 31K none
    PREPRODarchPOOL/d05 8.91G 49.7G 8.91G /d05
    PREPRODdataPOOL 807G 364G 31K none
    PREPRODdataPOOL/d02 13.4G 36.6G 13.4G /d02
    PREPRODdataPOOL/d03 782G 364G 782G /d03
    PREPRODdataPOOL/d06 11.4G 88.6G 11.4G /d06
    PREPRODredoPOOL 7.82G 3.93G 31K none
    PREPRODredoPOOL/d04 7.82G 3.93G 7.82G /d04
    None of the file systems in the delegated datasets appear in Ops Center for this zone. Are there any suggestions for how I correct this?

    Do you mean adopt the zone? That requires the zone be halted and it also says something about copying all file systems to the pool created for the zone. Of the 12 zones I have (four on each of three M5000s), seven of them are already in "production" status, and four of those seven now support 7x24 world-wide operations. A do-over is not an option here.

  • Change ZFS root dataset name for root file system

    Hi all
    A quick one.
    I accepted the default ZFS root dataset name for the root file system during Solaris 10 installation.
    Can I change it to another name afterward without reinstalling the OS? For example,
    zfs rename rpool/ROOT/s10s_u6wos_07b rpool/ROOT/`hostname`
    zfs rename rpool/ROOT/s10s_u6wos_07b/var rpool/ROOT/`hostname`/var
    Thank you.

    Renaming the root pool is not recommended.

  • How so I protect my root file system? - x86 solaris 10 - zfs data pools

    Hello all:
    I'm new to ZFS and am trying to understand it better before I start building a new file server. I'm looking for a low cost file server for smaller projects I support and would like to use the ZFS capabilities. If I install Solaris 10 on a x86 platform and add a bunch of drives to it to create a zpool (raidz), how do I protect my root filesystem? The files in the ZFS file system are well protected, but what about my operating system files down in the root ufs filesystem? If the root filesystem gets corrupted, do I lose the zfs filesystem too? or can I independantly rebuild the root filesystem and just remount the zfs filesystem? Should I install solaris 10 on a mirrored set of drives? Can the root filesystem be zfs too? I'd like to be able to use a fairly simple PC to do this, perhaps one that doesn't have built in raid. I'm not looking for 10 terabytes of storage, maybe just four 500gb sata disks connected into a raidz zpool.
    thanks,

    patrickez wrote:
    If I install Solaris 10 on a x86 platform and add a bunch of drives to it to create a zpool (raidz), how do I protect my root filesystem?Solaris 10 doesn't yet support ZFS for a root filesystem, but it is working in some OpenSolaris distributions.
    You could use Sun Volume Manager to create a mirror for your root filesystem.
    The files in the ZFS file system are well protected, but what about my operating system files down in the root ufs filesystem? If the root filesystem gets corrupted, do I lose the zfs filesystem too?No. They're separate filesystems.
    or can I independantly rebuild the root filesystem and just remount the zfs filesystem? Yes. (Actually, you can import the ZFS pool you created).
    Should I install solaris 10 on a mirrored set of drives?If you have one, that would work as well.
    Can the root filesystem be zfs too?Not currently in Solaris 10. The initial root support in OpenSolaris will require the root pool be only a single disk or mirrors. No striping, no raidz.
    Darren

  • Unbootable Solaris 10 x86 installed on ZFS root file system

    Hi all,
    I have unbootable Solaris 10 x86 installed on ZFS root file system. on an IDE HDD
    The bios keep showing the msg
    DISK BOOT FAILURE , PLEASE INSERT SYSTEM BOOT DISK
    please note :
    1- the HDD is connected properly and recognized by the system
    2- GRUB don't show any messages
    is there any guide to recover the system , or detail procedure to boot system again
    Thanks,,,

    It's not clear if this is a recently installed system that is refusing to boot OR if the system was working fine and crashed.
    If it's the former, I would suggest you check the BIOS settings to make sure it's booting from the right hard disk. In any case, the Solaris 10 installation should have writting the GRUB stage1 and stage2 blocks to the beginning of the disk.
    If the system crashed and is refusing to boot, you can try to boot from a Solaris 10 installation DVD. Choose the single user shell option and see if it can find your system. You should be able to use format/devfsadm/etc to do the actual troubleshooting. If your disk is still responding, try a `zpool import` to see if there is any data that ZFS can recognize (it usually has many backup uberblocks and disk labels scattered around the disk).

  • Sol10 u8 installed on a ZFS Root File System have different swap needs?

    Does Sol10 u8 installed on a ZFS Root File System have different swap needs/processes?
    Information:
    I've installed Solaris 10 (10/09 s10s_u8wos_08a SPARC, Assembled 16 September 2009) on a half dozen servers and every one of them no longer mount swap at boot.
    The install program commented out the old swap entry and created this one:
    # grep swap /etc/vfstab
    swap - /tmp tmpfs - yes -
    Everything works like a champ. I didn't discover the issue until I tried to install some patches and the install failed. It didn't fail because of lack of swap - it refused to run because it found "No swap devices configured".
    Here are the symptoms:
    # swap -s
    total: 183216k bytes allocated + 23832k reserved = 207048k used, 13600032k available
    # swap -l
    No swap devices configured
    # mount | grep swap
    /etc/svc/volatile on swap read/write/setuid/devices/xattr/dev=5ac0001 on Mon Apr 19 08:06:45 2010
    /tmp on swap read/write/setuid/devices/xattr/dev=5ac0002 on Mon Apr 19 08:07:40 2010
    /var/run on swap read/write/setuid/devices/xattr/dev=5ac0003 on Mon Apr 19 08:07:40 2010
    #

    Hi Nitabills,
    I assume that you create a zfs entry for swap with the commande zfs create -V $size
    did you launch the command :
    swap -a /dev/zvol/dsdk/$ZPOOL/swap
    Try this entry below in the vfstab :
    /dev/zvol/dsdk/$ZPOOL/swap - - swap - no -

  • Convert ZFS root file system to UFS with data.

    Hi, I would need to covert my ZFS root file systems to UFS and boot from the other disk as a slice (/dev/dsk/c1t0d0s0)
    I am ok to split the hard disk from root pool mirror. any ideas on how this can be acheived?
    Please sugget. Thanks,

    from the same document that was quoted above in the Limitations section:
    Limitations
    Version 2.0 of the Oracle VM Server for SPARC P2V Tool has the following limitations:
    Only UFS file systems are supported.
    Only plain disks (/dev/dsk/c0t0d0s0), Solaris Volume Manager metadevices (/dev/md/dsk/dNNN), and VxVM encapsulated boot disks are supported on the source system.
    During the P2V process, each guest domain can have only a single virtual switch and virtual disk server. You can add more virtual switches and virtual disk servers to the domain after the P2V conversion.
    Support for VxVM volumes is limited to the following volumes on an encapsulated boot disk: rootvol, swapvol, usr, var, opt, and home. The original slices for these volumes must still be present on the boot disk. The P2V tool supports Veritas Volume Manager 5.x on the Solaris 10 OS. However, you can also use the P2V tool to convert Solaris 8 and Solaris 9 operating systems that use VxVM.
    You cannot convert Solaris 10 systems that are configured with zones.

  • ZFS file system mount in solaris 11

    Create a ZFS file system for the package repository in the root pool:
    # zfs create rpool/export/repoSolaris11
    # zfs list
    The atime property controls whether the access time for files is updated when the files are read.
    Turning this property off avoids producing write traffic when reading files.
    # zfs set atime=off rpool/export/repoSolaris11
    Create the required pkg repository infrastructure so that you can copy the repository
    # pkgrepo create /export/repoSolaris11
    # cat sol-11-1111-repo-full.iso-a sol-11-1111-repo-full.iso-b > \
    sol-11-1111-repo-full.iso
    # mount -F hsfs /export/repoSolaris11/sol-11-1111-repo-full.iso /mnt
    # ls /mnt
    # df -k /mnt
    Using the tar command as shown in the following example can be a faster way to move the
    repository from the mounted file system to the repository ZFS file system.
    # cd /mnt/repo; tar cf - . | (cd /export/repoSolaris11; tar xfp -)
    # cd /export/repoSolaris11
    # ls /export/repoSolaris11
       pkg5.repository README
       publisher sol-11-1111-repo-full.iso
    # df -k /export/repoSolaris11
    # umount /mnt
    # pkgrepo -s /export/repoSolaris11 refresh
    =============================================
    # zfs create -o mountpoint=/export/repoSolaris11 rpool/repoSolaris11
    ==============================================I am trying to reconfigure the package repository with above steps. when reached the below step
    # zfs create -o mountpoint=/export/repoSolaris11 rpool/repoSolaris11
    created the mount point but not mounted giving the error message
    cannot mount ,directory not empty When restarted the box, threw service adm screen with error message
    not able to mount all pointsPlease advise and Thanks in advance.

    Hi.
    Don't mix content of directory as mountpoint and what you see after FS was mounted.
    On othet ZFS - mount point also clear. You see contetn of ZFS file system.
    For check you can unmount any other ZFS and see that mountpoint also clear.
    Regards.

  • Making backups of zones created in Zeta File System (ZFS)

    Hello,
    We have several zones installed in /export/zones. /export/zones is using ZFS as its file system.
    We would like to make backups (preferrably to tape), but get the following messages when we try:
    tcas:root> ufsdump 0uf /dev/rmt/0cn /export/zones/idmtest
    DUMP: `/export/zones/idmtest' is not on a locally mounted filesystem
    DUMP: The ENTIRE dump is aborted.
    What I am thinking of is to somehow take a snapshot of the zone and save it to /export/home (which is UFS), then running ufsdump on that.
    Is this a viable solution or are there better ways of backing up? I'm sure other people have the same type of setup.
    Best!
    Frank

    I can't believe that Sun would put a file system on
    public release without having a viable tape backup
    strategy. The choice was to delay or release it without one. 'zfs send' is a low-level mechanism that could be part of a real backup system (rather like 'tar' forms part of NetBackup).
    It sounds like though stable, ZFS is far from being
    being ready for production.Production is different things to different people. There were some commercial sites for whom the features were so compelling, they were using it in production before it was part of any Solaris release.
    Obviously you'll want to (try to) understand all the drawbacks before you use any software or system.
    I will log a tar to Sun and see what the response is.
    I will post it here if I get one that works for us.one what?
    Darren

  • Best practices for ZFS file systems when using live upgrade?

    I would like feedback on how to layout the ZFS file system to deal with files that are constantly changing during the Live Upgrade process. For the rest of this post, lets assume I am building a very active FreeRadius server with log files that are constantly updating and must be preserved in any boot environment during the LU process.
    Here is the ZFS layout I have come up with (swap, home, etc omitted):
    NAME                                USED  AVAIL  REFER  MOUNTPOINT
    rpool                              11.0G  52.0G    94K  /rpool
    rpool/ROOT                         4.80G  52.0G    18K  legacy
    rpool/ROOT/boot1                   4.80G  52.0G  4.28G  /
    rpool/ROOT/boot1/zones-root         534M  52.0G    20K  /zones-root
    rpool/ROOT/boot1/zones-root/zone1   534M  52.0G   534M  /zones-root/zone1
    rpool/zone-data                      37K  52.0G    19K  /zones-data
    rpool/zone-data/zone1-runtime        18K  52.0G    18K  /zones-data/zone1-runtimeThere are 2 key components here:
    1) The ROOT file system - This stores the / file systems of the local and global zones.
    2) The zone-data file system - This stores the data that will be changing within the local zones.
    Here is the configuration for the zone itself:
    <zone name="zone1" zonepath="/zones-root/zone1" autoboot="true" bootargs="-m verbose">
      <inherited-pkg-dir directory="/lib"/>
      <inherited-pkg-dir directory="/platform"/>
      <inherited-pkg-dir directory="/sbin"/>
      <inherited-pkg-dir directory="/usr"/>
      <filesystem special="/zones-data/zone1-runtime" directory="/runtime" type="lofs"/>
      <network address="192.168.0.1" physical="e1000g0"/>
    </zone>The key components here are:
    1) The local zone / is shared in the same file system as global zone /
    2) The /runtime file system in the local zone is stored outside of the global rpool/ROOT file system in order to maintain data that changes across the live upgrade boot environments.
    The system (local and global zone) will operate like this:
    The global zone is used to manage zones only.
    Application software that has constantly changing data will be installed in the /runtime directory within the local zone. For example, FreeRadius will be installed in: /runtime/freeradius
    During a live upgrade the / file system in both the local and global zones will get updated, while /runtime is mounted untouched in whatever boot environment that is loaded.
    Does this make sense? Is there a better way to accomplish what I am looking for? This this setup going to cause any problems?
    What I would really like is to not have to worry about any of this and just install the application software where ever the software supplier sets it defaults to. It would be great if this system somehow magically knows to leave my changing data alone across boot environments.
    Thanks in advance for your feedback!
    --Jason                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                           

    Hello "jemurray".
    Have you read this document? (page 198)
    http://docs.sun.com/app/docs/doc/820-7013?l=en
    Then the solution is:
    01.- Create an alternate boot enviroment
    a.- In a new rpool
    b.- In the same rpool
    02.- Upgrade this new enviroment
    03.- Then I've seen that you have the "radious-zone" in a sparse zone (it's that right??) so, when you update the alternate boot enviroment you will (at the same time) upgrading the "radious-zone".
    This maybe sound easy but you should be carefull, please try this in a development enviroment
    Good luck

  • Root disk craches how to retrive the data from ZFS file systems.

    Hi Friends,
    The solaris 10 OS (root disk) is crached.i have configered execpt root disk to all disks for ZFS file systems.We using to application.Now any possble to retrive the data.
    Server model - V880
    Pls help me.
    Advance thanks.

    If the OS wasn't on ZFS, then just rebuild the server, hook up the drives and run 'zpool import'. It should find the pool on the disks and offer it up to be imported.
    Darren

  • I am trying to install End Note for Mac I am getting the message 'no mountable file systems'?

    I am trying to install End Note for Mac I am getting the message 'no mountable file systems'?

    Hi
    This is the link it is endnote for mac from my university library site it is a referencing tool. http://moodle.uws.ac.uk/mod/page/view.php?id=132753

  • ZFS File   system space issue

    Hi All,
    Kindly help how to resolve ZFS File system space issue. I have deleted nearly 20 GB but File system size remains the same. Kindly advice on it.
    Thanks,
    Kumar

    The three reasons that I'm aware of that causes deleting files not to return space are 1) the file is linked to multiple names 2) deleting files which are still open by a process and 3) deleting files which are backed up by a snapshot. While it is possible you deleted 20GB in multiply linked and/or open files, I'd guess snapshots to be most likely case.
    For multiple "hard" links, you can see the link count (before you delete the file) in the "ls -l" command (the second field). If it is greater than one, deleting the file won't free up space. You have to delete all file names linked to the file to free up the space.
    For open files, you can use the pfiles command to see what files a process has open. The file space won't be recovered until all processes with a file open close it. Killing the process will do the job. If you like to use a big hammer, a reboot kills everything.
    For snapshots: Use the "zfs -t snapshot" command and look at the snapshots. The space used by the snapshot indicates how much space is held by the snapshot. Deleting the snapshot will free up space unless the space is still being held by another snapshot. To free space held by a file, you have to delete all snapshots which contain that file.
    Hopefully, I got all of this right.

  • Zfs destroy DOES NOT CHECK NFS mount file-systems

    I asked this question twitter once and the answer was a good one, but I did some checking today and was surprised!!
    # zfs destroy mypool/home/andrew
    The above command will destroy this file-system no questioned asked but if the file-system is mounted you will get back the Device busy and if you have snapshot then they will be protected as well
    server# zfs destroy mypool/home/andrew
    cannot unmount 'tank/home/andrew
    server# zfs destroy dpool/staff/margaret
    cannot destroy 'dpool/staff/margaret': filesystem has children
    use '-r' to destroy the following datasets:
    dpool/staff/margaret@Wed18
    dpool/staff/margaret@Wed22
    BUT?
    server# zfs destroy dpool/staff/margaret@Wed18
    server# zfs destroy dpool/staff/margaret@Wed22
    NFSclient# cd /home/margaret
    NFSlient# ls -l
    drwx------+ 2 margaret staff 2 Aug 29 17:06 Mail
    lrwxrwxrwx 1 margaret staff 4 Aug 29 17:06 mail -> Mail
    drwx--x--x+ 2 margaret staff 2 Aug 29 17:06 public_www
    server# zfs destroy dpool/staff/margaret
    server#
    GONE!!!
    I will file a bug report to see what Oracle say!
    Comments?
    I think there should be a hold/protect of file-systems
    # zfs hold dpool/staff/margaret
    Andrew

    The CR is already filed:
    6947584 zfs destroy should be recoverable or prevented
    The zfs.1m man page, which covers the mounted case and the ZFS admin guide are pretty clear
    about the current zfs destroy behavior.
    http://docs.oracle.com/cd/E23824_01/html/821-1448/gamnq.html#gammq
    Caution - No confirmation prompt appears with the destroy subcommand. Use it with extreme caution.
    zfs destroy [-rRf] filesystem|volume
    Destroys the given dataset. By default, the command
    unshares any file systems that are currently shared,
    unmounts any file systems that are currently mounted,
    and refuses to destroy a dataset that has active depen-
    dents (children or clones).
    I'm sorry that you were surprised.
    Accidents happen too, like destroying the wrong file system, so always have good backups.
    Thanks, Cindy

  • ZFS file system options

    Dear Friends,
    Which is the command to create an zfs file system similar to the following command.
    newfs -i 200000 -c 256 -C 8 -m 1 /dev/rdsk/c1t1d0s0
    Thanks,
    Sal.

    Dear Darren & Friends,
    I am trying to install lotus domino in T5240 server, where the mail files will be stored on zfs file system. I would like to tune the ZFS file system for donino before going to the production. Unfortunately sun doesn't have any documnetation for the ZFS file system tuning for domino, but many for ufs. In ufs tuning sun is suggesting like this
    "Modern disk subsystems can perform larger data transfers than their predecessors, and Solaris can take
    advantage of this to read more data than Domino asks for in hopes of having the next piece of a file already in
    memory as soon as Domino asks for it. Unfortunately, Domino doesn't always use the &ldquo;anticipated&rdquo; data but
    instead next asks for data from an entirely different region of the file, so the effort spent reading the extra data
    may be wasted. In extreme circumstances and with modern disk systems Solaris can be fooled into reading
    fifty or sixty times as much data as is actually needed, wasting more than 98% of the I/O effort.
    To prevent this pathological behavior, build or tune the file systems that hold NSF databases so that Solaris
    won't try to read more than about 64KB at a time from them. The instructions that follow show how to do this for
    the default Unix File System (UFS); if you are using an alternative file system, consult its documentation.
    If you are in a position to build or rebuild the file systems, we suggest using the command
    newfs -i 200000 -c 256 -C 8 -m 1 /dev/rdsk/...
    The important option is -C 8, which limits the amount of read-ahead to no more than eight pages of 8KB each.
    The other options are less important, and may even cause newfs to issue warning messages for some disk
    sizes; these warnings can be ignored since newfs will adjust the values to suit the actual disk."
    Can I build an zfs file system atleast with the option similaiar to -C 8
    Thanks,
    Sal.

Maybe you are looking for

  • Firefox will not open my previous session

    Firefox 4 will not open any previous session once I shut down. I can have say 5 tabs open, close firefox, and reopen it and the session was saved. If I have those same 5 tabs open and shut down, then turn on my computer the next day, I can click on f

  • 32'' LCD TV model 32FT2U firmware corrupted, help please!!

    Hi All!!: Just recently my 32'' LCD TV model 32FT2U got corrupted firmware during an ac power blink, firmware updates are not posted at the support site, please, where can i get the original firmware, can anyone help me? Thanks in advance!!

  • HT1657 Can I watch a rented movie on the road(without wifi)?

    I was interested in watching killing them softly( a movie) on the way to vacation with my family. Since I already downloaded it because I rented it, can I start watching it for the first time without wifi on the road.

  • Report that reflects Supervisor Console

    All, Look, CUIC sucks it.  There, I said it.  That being said, I'm looking to create a report that reflects that data that is shown in the Supervisor Console.  This is what everyone wants.  Inbound/outbound call details for each agent and logout/not

  • Displaying date field in menu using HTML DB

    How do I display a date field in HTML DB as drop down. I have the below HTML code which needs to be represented in HTML DB. The display on the page should be like Enter Start Date : "Month" "Day of the Month" "Year". The above values in quotes should