SunCluster, MPXIO, Clariion and ZFS?
Hi,
we have a 2 node cluster (SunCluster 3.2). Our Storage is a EMC Clariion CX700. We have created some zpools and integrated them into the suncluster.
We cannot use PowerPath 5.1 and 5.2 for this because sun cluster and zfs is not supported in this environment. So, we want to use mpxio. Our question is, if there is a SP-Failover at the clariion, does mpxio support this and everything works fine without any problems?
Thanks!
Greets
Björn
Hi,
What you need todo is the following.
edit this file /kernel/drv/scsi_vhci.conf
follow the directions of this link
http://www.filibeto.org/sun/lib/nonsun/emc/SolarisHostConectivity.pdf?bcsi_scan_1BD4CB6F2E356E40=0&bcsi_scan_filename=SolarisHostConectivity.pdf
regards
Filip
Similar Messages
-
Hi,
Any one can share what is max file size can be created in Solaris 10 UFS and ZFS ?
What will be max size file compression using tar,gz ?
Regards
Sivafrom 'man ufs':
A sparse file can have a logical size of one terabyte.
However, the actual amount of data that can be stored
in a file is approximately one percent less than one
terabyte because of file system overhead.
As for ZFS, well, its a 128bit filesystem, and the maximum size of a file or directory is 2 ^64^ bytes, which i think is somewhere around 8 exabyte (i.e 8192 petabyte), even though my calculator gave up on calculating it.
http://www.sun.com/software/solaris/ds/zfs.jsp
.7/M.
Edited by: abrante on Feb 28, 2011 7:31 AM
fixed layout and 2 ^64^ -
Hi All,
I'm using clustered zones and ZFS and I get these messages below.
Is this something that I need to be worried about ?
Have I missed something when I created the resource which actually
is configured "by the book"?
Will HAStoragePlus work as expected?
Can I somehow verify that the zpool is monitored?
Apr 4 15:38:07 dceuxa2 SC[SUNW.HAStoragePlus:4,dceux08a-rg,dceux08a-hasp,hastorageplus_postnet_stop]: [ID 815306 daemon.warning] Extension properties GlobalDevicePaths and FilesystemMountPoints are both empty.
/Regards
UlfThanks for Your quick replies.
The HASP resource was created with -x Zpools="orapool1,orapool2"
and all other properties is at their defaults.
part of clrs show -v...
Resource: dceux08a-hasp
Type: SUNW.HAStoragePlus:4
Type_version: 4
Group: dceux08a-rg
R_description:
Resource_project_name: default
Enabled{dceuxa1:dceux08a}: True
Enabled{dceuxa2:dceux08a}: True
Monitored{dceuxa1:dceux08a}: True
Monitored{dceuxa2:dceux08a}: True
FilesystemMountPoints: <NULL>
GlobalDevicePaths: <NULL>
Zpools: orazpool1 orazpool2
(Solaris10u3/Sparc SC3.2, EIS 27-Feb)
/BR
Ulf -
We have ZfD running on one server for approx. 600 users (Sybase db on
NetWare 6.5).
We use it for; WS registration, WS Inventory, Application Mgmt, NAL
database, Imaging)
I have a mixture of Microsoft Windows and Novell NetWare servers.
Approximately:
30 Microsoft Windows servers (2000 and 2003)
10 Novell NetWare servers (NW 5.1 SP7 and NW 6.5 SP3)
Q1: Is it feasable to have the ZfS backend running on the same server that
hosts the ZfD backend ?
We are trying to find a way to monitor all server for disk usage. Ideally
we want to get a view/report of all servers (regardless of Novell or
Microsoft) to see where each disk is at with regards to available space and
also see historical trends for disk usage.
Q2: Can ZfS do this for us? We are licensed to use it but so far we've
only implemented the ZfD 6.5.2 and are quite please with the results.
Q3: Also, since we are licensed to use the latest ZfD and ZfS, any reason
to implement ZfS 7 instead of ZfS 6.5? We know that ZfD 7 is pretty much
the same as ZfD 6.5.2 so we've decided to hold back on this upgrade. If we
move forward with ZfS, I'm guessing that sticking with same version being
used with ZfD is a good idea?
Thanks for any answers!
MarcMarc Charbonneau,
>Q1: Is it feasable to have the ZfS backend running on the same server that
>hosts the ZfD backend ?
>
>We are trying to find a way to monitor all server for disk usage. Ideally
>we want to get a view/report of all servers (regardless of Novell or
>Microsoft) to see where each disk is at with regards to available space and
>also see historical trends for disk usage.
Yes, it's very workable with both ZFD and ZFS on the same box. ZFS can
monitor any of these features. It uses SNMP to do this on both netware and
windows.
>
>Q2: Can ZfS do this for us? We are licensed to use it but so far we've
>only implemented the ZfD 6.5.2 and are quite please with the results.
>
Glad to hear ZFD is working for you.
>Q3: Also, since we are licensed to use the latest ZfD and ZfS, any reason
>to implement ZfS 7 instead of ZfS 6.5? We know that ZfD 7 is pretty much
>the same as ZfD 6.5.2 so we've decided to hold back on this upgrade. If we
>move forward with ZfS, I'm guessing that sticking with same version being
>used with ZfD is a good idea?
Yes, although ZFS7 subscribers can run on XP, but I don't think 6.5 can.
In a way, zfd and zfs are very separate and the patches do not have to
match, but if you can keep it the same, than do. :)
Hope that helps.
Jared
Systems Analyst at Data Technique, INC.
jjennings at data technique dot com
Posting with XanaNews 1.17.6.6 in WineHQ
Check out Novell WIKI
http://wiki.novell.com/index.php/IManager -
We have ZfD running on one server for approx. 600 users (Sybase db on
NetWare 6.5).
We use it for; WS registration, WS Inventory, Application Mgmt, NAL
database, Imaging)
I have a mixture of Microsoft Windows and Novell NetWare servers.
Approximately:
30 Microsoft Windows servers (2000 and 2003)
10 Novell NetWare servers (NW 5.1 SP7 and NW 6.5 SP3)
Q1: Is it feasable to have the ZfS backend running on the same server that
hosts the ZfD backend ?
We are trying to find a way to monitor all server for disk usage. Ideally
we want to get a view/report of all servers (regardless of Novell or
Microsoft) to see where each disk is at with regards to available space and
also see historical trends for disk usage.
Q2: Can ZfS do this for us? We are licensed to use it but so far we've
only implemented the ZfD 6.5.2 and are quite please with the results.
Q3: Also, since we are licensed to use the latest ZfD and ZfS, any reason
to implement ZfS 7 instead of ZfS 6.5? We know that ZfD 7 is pretty much
the same as ZfD 6.5.2 so we've decided to hold back on this upgrade. If we
move forward with ZfS, I'm guessing that sticking with same version being
used with ZfD is a good idea?
Thanks for any answers!
MarcMarc,
It appears that in the past few days you have not received a response to your
posting. That concerns us, and has triggered this automated reply.
Has your problem been resolved? If not, you might try one of the following options:
- Do a search of our knowledgebase at http://support.novell.com/search/kb_index.jsp
- Check all of the other support tools and options available at
http://support.novell.com.
- You could also try posting your message again. Make sure it is posted in the
correct newsgroup. (http://support.novell.com/forums)
Be sure to read the forum FAQ about what to expect in the way of responses:
http://support.novell.com/forums/faq_general.html
If this is a reply to a duplicate posting, please ignore and accept our apologies
and rest assured we will issue a stern reprimand to our posting bot.
Good luck!
Your Novell Product Support Forums Team
http://support.novell.com/forums/ -
After updating kernel and ZFS modules, system cannot boot
Starting Import ZFS pools by cache file...
[ 4.966034] VERIFY3(0 == zap_lookup(ddt->ddt_os, ddt->ddt_spa->spa_ddt_stat_object, name, sizeof (uint64_t), sizeof (ddt_histogram_t) / sizeof (uint64_t), &hht->ddt_histogram[type][class])) failed (0 == 6)
[ 4.966100] PANIC at ddt.c:124:ddt_object_load()
[*** ] A start job is running for Import ZFS pools by cache (Xmin Ys / no limit)
And then occasionally I see
[ 240.576219] Tainted: P O 3.19.2-1-ARCH #1
Anyone else experiencing the same?Thanks!
I did the same and it worked... kind of. The first three reboots it failed (but did not stop the system from booting) producing:
zpool[426]: cannot import 'data': one or more devices is currently unavailable
systemd[1]: zfs-import-cache.service: main process exited, code=exited, status=1/FAILURE
The second boot also resulted in a kernel panic, but as far as I can tell unrelated to zfs.
After reboot one and three imported the pool manually.
From the fourth reboot on loading from cache file always succeeded. However, I it takes faily long (~8 seconds) and even shows
[*** ] A start job is running for Import ZFS pools by cache (Xmin Ys / no limit)
briefly. Altough I might only notice that because the recent updates sped up oder parts of the boot process. Did you observe a slowdown during boot time, too, kinghajj?
Last edited by robm (2015-03-22 01:21:05) -
Solaris 10 upgrade and zfs pool import
Hello folks,
I am currently running "Solaris 10 5/08 s10x_u5wos_10 X86" on a Sun Thumper box where two drives are mirrored UFS boot volume and the rest is used in ZFS pools. I would like to upgrade my system to "10/08 s10x_u6wos_07b X86" to be able to use ZFS for the boot volume. I've seen documentation that describes how to break the mirror, create new BE and so on. This system is only being used as iSCSI target for windows systems so there is really nothing on the box that i need other then my zfs pools. Could i simply pop the DVD in and perform a clean install and select my current UFS drives as my install location, basically telling Solaris to wipe them clean and create an rpool out of them. Once the installation is complete, would i be able to import my existing zfs pools ?
Thank you very muchSure. As long as you don't write over any of the disks in your ZFS pool you should be fine.
Darren -
Solaris 10 JET install and ZFS
Hi - so following on from Solaris Volume Manager or Hardware RAID? - I'm trying to get my client templates switched to ZFS but it's failing with:
sudo ./make_client -f build1.zfs
Gathering network information..
Client: xxx.14.80.196 (xxx.14.80.0/255.255.252.0)
Server: xxx.14.80.199 (xxx.14.80.0/255.255.252.0, SunOS)
Solaris: client_prevalidate
Clean up /etc/ethers
Solaris: client_build
Creating sysidcfg
WARNING: no base_config_sysidcfg_timeserver specified using JumpStart server
Creating profile
Adding base_config specifics to client configuration
Adding zones specifics to client configuration
ZONES: Using JumpStart server @ xxx.14.80.199 for zones
Adding sbd specifics to client configuration
SBD: Setting Secure By Default to limited_net
Adding jass specifics to client configuration
Solaris: Configuring JumpStart boot for build1.zfs
Solaris: Configure bootparams build
Starting SMF services for JumpStart
Adding Ethernet number for build1 to /etc/ethers
cleaning up preexisting install client "build1"
removing build1 from bootparams
removing /tftpboot/inetboot.SUN4V.Solaris_10-1
svcprop: Pattern 'network/tftp/udp6:default/:properties/restarter/state' doesn't match any entities
enabling network/tftp/udp6 service
svcadm: Pattern 'network/tftp/udp6' doesn't match any instances
updating /etc/bootparams
copying boot file to /tftpboot/inetboot.SUN4V.Solaris_10-1
Force bootparams terminal type
-Restart bootparamd
Running '/opt/SUNWjet/bin/check_client build1.zfs'
Client: xxx.14.80.196 (xxx.14.80.0/255.255.252.0)
Server: xxx.14.80.199 (xxx.14.80.0/255.255.252.0, SunOS)
Checking product base_config/solaris
Checking product custom
Checking product zones
Product sbd does not support 'check_client'
Checking product jass
Checking product zfs
WARNING: ZFS: ZFS module selected, but not configured to to anything.
Check of client build1.zfs
-> Passed....
So what is "WARNING: ZFS: ZFS module selected, but not configured to to anything." referring to? I've amended my template and commented out all references to UFS so I now have this:
base_config_profile_zfs_disk="slot0.s0 slot1.s0"
base_config_profile_zfs_pool="rpool"
base_config_profile_zfs_be="BE1"
base_config_profile_zfs_size="auto"
base_config_profile_zfs_swap="65536"
base_config_profile_zfs_dump="auto"
base_config_profile_zfs_compress=""
base_config_profile_zfs_var="65536"
I see there is a zfs.conf file in /opt/SUNWjet/Products/zfs/zfs.conf do I need to edit that as well?
Thanks - J.Hi Julian,
You MUST create /var as part of the installation in base_config, as stuff gets put there really early during the install.
The ZFS module allows you to create additional filesystems/volumes in the rpool, but does not let you modify the properties of existing datasets/volumes.
So,
you still need
base_config_profile_zfs_var="yes" if you want a /var dataset.
/export and /export/home are created by default as part of the installation. You can't modify that as part of the install.
For your zones dataset, seems to be fine and as expected, however, the zfs_rpool_filesys needs to list ALL the filesystems you want to create. It should read zfs_rpool_filesys="logs zones". This makes JET look for variables of the form zfs_rpool_filesys_logs and zfs_rpool_filesys_zones. (The last variable is always picked up, in your case the zones entry. Remember, the template is a simple name=value set of variables. If you repeat the "name" part, it simply overwrites the value.)
So you really want:
zfs_rpool_filesys="logs zones"
zfs_rpool_filesys_logs="mountpoint=/logs quota=32g"
zfs_rpool_filesys_zones="mountpoint=/zones quota=200g reservation=200g"
(incidentally, you don't need to put zfs_pools="rpool" as JET assumes this automatically.)
So, if you want to alter the properties of /var and /export, the syntax you used would work, if the module was set up to allow you to do that. (It does not currently do it, but I may update it in the future to allow it).
(Send me a direct e-mail and I can send you an updated script which should then work as expected, check my profile and you should be able to guess my e-mail address)
Alternatively, I'd suggest writing a simple script and stick it into the /opt/SUNWjet/Clients/<clientname> directory with the following lines in them:
varexportquotas:
#!/bin/sh
zfs set quota=24g rpool/export
zfs set quota=24g rpool/ROOT/10/var
and then running it in custom_scripts_1="varexportquotas"
(Or you could simply type the above commands the first time you log in after the build. :-) )
Mike
Edited by: mramcha on Jul 23, 2012 1:39 PM
Edited by: mramcha on Jul 23, 2012 1:45 PM -
A question about where the ARC cache resides in a Sun ZFS 7320 Storage Appliance? Does it run in the cache of the storage head or the RAM of the node?
Thanks for the reply. I see you are pointing to the physical 'read' hardware in the storage head or readzilla. I believe this is where L2ARC storage is maintained. My question is about the Adaptive Replacement Cache (ARC). I am confused about where this and the ghost lists are maintained. References in the various blogs talk about main memory/system memory, which memory is this - the memory in the server node or the memory in the storage head - say the ZFS 7320 Storage as a standalone device or the ZFS 7320 Storage Appliance embedded in an Exalogic.
-
Replace a mirrored disk with SVM meta and ZFS
hello everybody,
i've a mirrored disk that have some metadata (svm) configured (/ - /usr -/var and swap ) and a slice with zfs filesystem .
I need to replace the disk .
Someone could help me ?
Edited by: Nanduzzo1971 on Jul 18, 2008 4:46 AMIt's quite easy, just check the videos on the link below.
http://www.powerbookmedic.com/manual.php?id=4 -
SunMC Agent 3.6.1 and ZFS
Hello,
I was wondering if a SunMC agent is able to recognize a ZFS filesystem? I've tried it on one of our test servers there is no category under Kernel Reader-Filesystem Usage for ZFS...only ufs and vxfsI was wondering if a SunMC agent is able to recognize
a ZFS filesystem? I've tried it on one of our test
servers there is no category under Kernel
Reader-Filesystem Usage for ZFS...only ufs and vxfsNot quite yet. In fact a SunMC Server will refuse to even install on a ZFS partition without some minor changes to its setup utils. But the next release should be fully ZFS aware and compatible.
Regards,
[email protected]
http://www.HalcyonInc.com -
Hi - I have a JET (jumpstart) server that I've used many times before to install various Solaris SPARC servers with - from V240's to T4-1's. However when I try to install a brand new T4-2 I keep seeing this on screen and the install reverts to a manual install:
svc:/system/filesystem/local:default: WARNING: /usr/sbin/zfs mount -a failed: one or more file systems failed to mount
There's been a previous post about this but I can't see the MOS doc that is mentioned in the last post.
The server came pre-installed with Sol11 and I can see the disks:
AVAILABLE DISK SELECTIONS:
0. c0t5000CCA016C3311Cd0 <HITACHI-H109030SESUN300G-A31A cyl 46873 alt 2 hd 20 sec 625> solaris
/scsi_vhci/disk@g5000cca016c3311c
/dev/chassis//SYS/SASBP/HDD0/disk
1. c0t5000CCA016C33AB4d0 <HITACHI-H109030SESUN300G-A31A cyl 46873 alt 2 hd 20 sec 625> solaris
/scsi_vhci/disk@g5000cca016c33ab4
/dev/chassis//SYS/SASBP/HDD1/disk
If I drop to the ok prompt there is no hardware RAID configured and raidctl also shows nothing:
root@solarist4-2:~# raidctl
root@solarist4-2:~#
The final post I've found on this forum for someone with this same problem was "If you have an access to MOS, please check doc ID 1008139.1"
Any help would be appreciated.
Thanks - J.Hi Julian,
I'm not convinced that your problem is the same one that is described in this discussion:
Re: Problem installing Solaris 10 1/13, disks no found
Do you see the missing volume message (Volume 130 is missing) as described in this thread?
A google search shows that there are issues with for a T4 Solaris 10 install due to a network driver problem and also if the system is using
a virtual CD or device through a LDOM.
What happens when you boot your T4 from the installation media or server into single-user mode? You say that you can see the disks, but can you create a ZFS storage pool on one of these disks manually:
# zpool create test c0t5000CCA016C3311Cd0s0
# zpool destroy test
For a T4 and a Solaris 10 install, the disk will need an SMI (VTOC) label, but I would expect a different error message if that was a problem.
Thanks, Cindy -
ISCSI and ZFS Thin Provisioning Sparse Volumes - constraints?
Hello,
I am running an iSCSI target using COMSTAR.
I activated Time Slider (Snapshot feature) for all pools.
Now I want to set up an iSCSI target using thin provisioning, storing the data in a file system rather than a file.
Is there any official documentation about thin p.?
All I found was
http://www.cuddletech.com/blog/pivot/entry.php?id=729
http://www.c0t0d0s0.org/archives/4222-Less-known-Solaris-Features-iSCSI-Part-4-Alternative-backing-stores.html
Are there any problems to be expected about the snapshots?
How would I set up a 100 GByte iSCSI target with mentioned thin p.?
Thanks
n00bTo create a thin provisioned volume:
zfs create -V <SIZE> -s path/to/volume
Where <SIZE> is the capacity of the volume and path/to/volume is the ZFS path and volume name.
To create a COMSTAR target:
stmfadm create-lu /dev/zvol/rdsk/path/to/volume
You'll get a LU ID, which you can then use to create a view, optionally with target and host groups to limit access.
-Nick -
Hello All,
Is there any specific shell command on Solaris 10, to find out if the file has NFSv4 or ZFS style ACLs set on them ??
-- I see there is a system call available for this purpose, but did not come across any specific command just to tell the presence of the above style ACLs.
-- "ls -v" displays the ACLs itself, but its too verbose
Also, are there any Perl modules available for checking the above style ACLs
Edited by: manjunathbs on Dec 12, 2008 6:03 PMYou will need to have concurrent ACLs applied to an interface.
The access lists are address familiy specific in their syntax and features so they cannot be mixed. An indicative example is shown below.
interface Ethernet1/1
ip access-group test-v4 in
ipv6 traffic-filter test-v6 in
ip access-list extended test-v4
permit ip any host 1.1.1.1
deny ip any any
ipv6 access-list test-v6
permit ipv6 any host 2001:DB8::1
deny ipv6 any any -
Ldmp2v and ZFS root source system question
hi
reading the ldmp2v doc, it seems to implay that p2v only support source system with UFS root
this is fine for s8 and s9 system.
what about the new s10 with zfs root?
thxChk the links
Transfer global settings - Multiple source systems
Re: Difference between Transfer Global Setting & Transfer Exchange rates
Regards,
B
Maybe you are looking for
-
Hi everyone, I just completed an upgrade to Lion and while I love it I've suddenly found myself short 1 display. I use 2x24" with my MacBook Pro - 1 via mini Display Port [the Thunderbolt port] and another via an eVGA UV+ [both displays have a resolu
-
will apple next update to help fix the location following data problems will the ipod touch 2g be included ?
-
CRM 5.0 - IC WebClient - Due Date - Working Hours for Service Ticket
Hi Gurus We're using a Service Ticket Transaction which has a Date Profile which provides a 48 Hour window for which the user must complete the Service Ticket Resolution. The issue is I want to tie this up to the Business working Hours. I just cannot
-
I have mac osx 10.6.8 but i cannot update but win i upgriding osx mountain lion tha mac tell me (cheke network ???) I HAVE mac book pro 2011
-
On mapping validation send e-mail on errors
I would like to be able to send an e-mail if the mapping comes up with an error. I have a field (Supplier ID) and if its blank during the message mapping I want it to send an e-mail. I have all the objects ready to send emails e.g. Mail adaptor etc,