NSS Pools/Volumes Deactivate - VMware

Greetings,
I've been trying to virtualize two of our NetWare 6.5 servers. My thought
was to install two new virtual servers and migrate the data from the
physical server to the virtual server. However, I've been running into
problems with the NSS volumes on the virtual servers deactivating
themselves ... not exactly the sort of thing you really want happening to
a production box.
When the pools deactivate, the errors displayed are similar to the
following:
10-05-2007 3:36:54am: COMN-3.25-1092 [nmID=A0025]
NSS-3.00-5001: Pool SERVERNAME/IFOLDER is being deactivated.
An I/O error (20204(zio.c[2260])) at block 23602279(file block
77251)(ZID
3) has compromised pool integrity.
10-05-2007 3:36:56am: SERVER-5.70.1534 [nmID=B0013]
Device "[V358-A2-D2:0] VMware Virtual disk f/w:1.0" deactivated by
driver
due to device failure.
FWIW, the VMware hosts are running ESX 3.0.2 and are connected to a NetApp
3050C filer via fiber channel. There are about 60 other Linux and Windows
virtual machines in the mix, all connected the same way that are not
exhibiting these sorts of issues.
I contacted VMware support regarding the issue and was told to use
VMDK-based disks rather than mapped raw LUNs. When that server
experienced the same issue the tech suggested I use locally attached
storage (!?). Other than that, the only other suggestion he gave to me
was to wait for a new LSIMPTNW.HAM to correct the problem I was
experiencing.
Obviously, many of you are running NetWare 6.5 on VMware...what's the key
for success?
Thanks,
Rob

Rob Walters,
>
> I've been trying to virtualize two of our NetWare 6.5 servers. My
> thought was to install two new virtual servers and migrate the data from
> the physical server to the virtual server. However, I've been running
> into problems with the NSS volumes on the virtual servers deactivating
> themselves ... not exactly the sort of thing you really want happening
> to a production box.
>
> When the pools deactivate, the errors displayed are similar to the
> following:
>
> 10-05-2007 3:36:54am: COMN-3.25-1092 [nmID=A0025]
> NSS-3.00-5001: Pool SERVERNAME/IFOLDER is being deactivated.
> An I/O error (20204(zio.c[2260])) at block 23602279(file block
> 77251)(ZID
> 3) has compromised pool integrity.
>
> 10-05-2007 3:36:56am: SERVER-5.70.1534 [nmID=B0013]
> Device "[V358-A2-D2:0] VMware Virtual disk f/w:1.0" deactivated by
> driver
> due to device failure.
>
> FWIW, the VMware hosts are running ESX 3.0.2 and are connected to a
> NetApp 3050C filer via fiber channel. There are about 60 other Linux
> and Windows virtual machines in the mix, all connected the same way that
> are not exhibiting these sorts of issues.
>
> I contacted VMware support regarding the issue and was told to use
> VMDK-based disks rather than mapped raw LUNs. When that server
> experienced the same issue the tech suggested I use locally attached
> storage (!?). Other than that, the only other suggestion he gave to me
> was to wait for a new LSIMPTNW.HAM to correct the problem I was
> experiencing.
>
Rob, I'm having the same problem on one of my NW6.5 servers that's on
VMWare. Last time, I think it showed the same error message you report.
It happened again just this morning, but this time it says:
10-11-2007 8:50:44 am: COMN-3.25-1196 [nmID=A0026]
NSS-3.00-5002: Pool MEDDIR/VOL2 is being deactivated.
An error (20012(nameTree.c[45])) at block 20227(file block -20227)
(ZID6) has compromised volume integrity.
Several other NW6.5 servers on the same VMware host have never given me
any problem. Too weird. If you find out what the culprit is, please
let me know.
Thanks,
Doug

Similar Messages

  • OES11 SP2 - Linux File System and NSS Pools & Volumes

    Planning to install our first OES11 SP2 server into an existing tree - the
    idea is to run this new OES11 server virtualized on VMware ESXi 5.5
    The existing tree has two physical NW6.5SP8 servers running eDirectory
    Version 8.7.3.10b (NDS Version 10554.64). One of the existing Netware
    servers is used for DHCP/DNS, File Sharing from three NSS volumes and
    Groupwise 7.0.4 whilst the second server is used for FTP services and
    eDirectory redundancy. Ultimately the plan is to have two virtualized OES11
    SP2 server with one for file sharing and the other for making the move from
    GW7 to GW2012. And we're planning to stick with NSS for file sharing on the
    new OES11 SP2 server.
    I've come across a couple of posts for earlier versions of OES which
    recommended not to put the Linux Native OS File System and NSS storage
    pools/volumes on the same hard drive. Apparently the advice was a result of
    needing to allow EVMS to manage the drive which could be problemmatic.
    I've read the OES11 documentation which says that "The Enterprise Volume
    Management System (EVMS) has been deprecated in SLES 11, and is also
    deprecated in OES 11. Novell Linux Volume Manager (NLVM) replaces EVMS for
    managing NetWare partitions under Novell Storage Services (NSS) pools."
    So I'm wondering if there is still a compelling requirement to keep the
    Linux Native File System and NSS pools/volumes on separate hard drives or
    can they both now safely co-exist on the same drive without causing
    headaches or gotchas for the future?
    Neil

    Hi Willem,
    Many thanks for the further reply.
    So we can just use the VMWare setup to "split" the one physical drive into
    two virtual drives (one for the OS and the second for the pools).
    And I've seen posts in other forums about the need for a decent battery
    backed cache module for the P410i controller so I'll make sure we get one
    (probably 512Mb module + battery).
    Can I ask what is the advantage of configuring each VM's virtual disk to run
    on it's own virtual SCSI adapter (by setting disk1 to scsi 0:0, disk2 to
    scsi 1:0, and so on)?
    Cheers,
    Neil
    >>> On 9/5/14 at 12:56, in message
    <[email protected]>,
    magic31<[email protected]> wrote:
    > HI Niel,
    >
    > xyzl;2318555 Wrote: >
    >> The new installation will run on a Proliant ML350 G6 with P410i>
    > controller
    >> so we can use the raid capability to create two different logical drives>
    > as
    >> suggested.
    >
    > As you will be using ESXi 5.5 as host OS, it's not needed to split
    > thehost server storage into two logical drives... unless that's what
    > youwant in perspective for "general performance" or redundancy reasons.
    > Italso depends on the options that P410i controller has.
    >
    > On a side note, I'm not too familiar with the P410i controller... domake
    > sure you have a decent battery backed cache module installed, asthat will
    > greatly help with the disk performance bit.
    > If the controller can handle it and the controller can handle it, go
    > forraid 10 or raid 50. That might be too big a space penalty but will
    > helpwith disk performance.
    >
    > Once you have your VMware server up and running, you can configure
    > thetwo VM's with each two or more drives attached (on for the OS,
    > thesecond or others for your pools).
    > I usually create a virtual disk per pool+volume set (e.g. DATAPOOL
    > &DATAVOLUME on one vm virtual disk, USERPOOL & USER volume on an other
    > vmvirtual disk).
    > With VMware you can than also configure each VM's virtual disk to run
    > onit's own virtual SCSI adapter (bij setting disk1 to scsi 0:0, disk2
    > toscsi 1:0, and so on).
    >
    >
    > xyzl;2318555 Wrote: > Do you have any suggestions for the disk space that
    > should be reserved> or
    >> used for the Linux Native OS File System (/boot, /swap and LVM)?
    >>
    >
    > Here's one thread that might be of interest (there are more
    > throughoutthe SLES/OES
    >
    forums):https://forums.novell.com/showthread...rtitioning-%28
    > moving-from-NW%29
    >
    > I still contently follow the method I choose for back in 2008, justwith
    > a little bigger sizing which now is:
    >
    > On a virtual disk sized 39GB:
    >
    > primary partition 1: 500 MB /boot , fs type ext2
    > primary partition 2: 12GB / (root), fs type ext3
    > primary partition 3: 3 GB swap , type swap
    >
    > primary partition 4: LVM VG-SYSTEM (LVM partition type 8E), takes up
    > therest of the disk **
    > LVM volume (lv_var): 12 GB /var , fs type ext3
    > LVM volume (lv_usr-install): 7GB /usr/install, fs type ext3
    > * there's still a little space left in the LVM VG, in case var needs
    > toquickly be enlarged
    >
    > One thing that's different in here vs what I used to do: I replaced
    > the/tmp mountpoint with /usr/install
    >
    > In /usr/install, I place all relevant install files/IOS's
    > andinstallation specifics (text files) for the server in question. Keeps
    > itall in one neat place imo.
    >
    > Cheers,
    > Willem-- Knowledge Partner (voluntary sysop)
    > ---
    > If you find a post helpful and are logged into the web interface,
    > please show your appreciation and click on the star below it.
    >
    Thanks!---------------------------------------------------------------------
    ---magi
    > c31's Profile: https://forums.novell.com/member.php?userid=2303View this
    > thread: https://forums.novell.com/showthread.php?t=476852

  • Searching for a word or phrase in a file on NSS Pool/Volume

    I have encountered a strange issue in that if I search for a word or phrase in a file at any level of the volume, from the root, down to sub folders, I get different results from 2 or more PC's.
    Workstations are all Win XP SP2, Novell Client 4.91 SP4 for Windows
    Server NW 6.5 SP6 with all post SP's installed.
    I have ran a nss /PoolRebuild=POOLNAME and a NSS /VisibilityRebuild=VOLNAME
    There is only one volume on this pool.
    And I still encounter the same results.
    As one example:
    On my Admin workstation When I search the top level Folder of Closed Client folders for the word chronology, I get 2 files returned named chronfile1 and chonfile2, when I search from any other workstation (I have tried 4 different workstations at this point) at the same folder level, for the same word Chronology, I get multiple files with the Word in the File, but I do not see files chronfile1 or chronfile2. Likewise, the file names returned form the search results on the other workstations, are not returned in the search list ran on the admin (Mine) workstation. And yes, I am logged in as admin on the other workstations when running the search.
    So far, this issue is only appearing on 1 volume.
    Thoughts?

    Originally Posted by ataubman
    If you log in to the second PC as Admin, does the symptom move? IOW does it follow the user or the workstation? If the user you may have some odd file system rights set, although your specific results you report are puzzling.
    I'm assuming you are using the Windows desktop search tool?
    I ran a NSS /PoolRebuild last night which fixed an issue the workstations not being able to see the 2 files in question.
    Previous night I had just performed an NSS /PoolVerify.
    Part of the other anomoly, the other workstations seeing more files than the admin workstation (I feel so Stupid, after 25 years you think I'd be able to spot the obvious) the reason the other workstations were returning files that the Admin workstation was not - Admin workstation doesn't have Word Perfect installed.
    Somebody, give me a Gibbs head slap, now!

  • Netware NSS-Pools under VMWare ESX

    Just for my planning: When placing the Netware-Server on one
    ESX-Datastore is it possible/advisable to place an
    additional Netware-NSS--Pool for this server on a second
    ESX-Datastore in the same ESX-Host (ESX 4.1)? How does
    Netware recognize the additional disk-space?
    Sincerely
    Karl

    Thanks a lot, Massimo.
    Sincerely
    Karl
    >>> Massimo Rosen<[email protected]> 26.01.2012 12:08
    >>>
    >On 26.01.2012 11:55, [email protected] wrote:
    >> Just for my planning: When placing the Netware-Server on
    >one
    >> ESX-Datastore is it possible/advisable to place an
    >> additional Netware-NSS--Pool for this server on a second
    >> ESX-Datastore in the same ESX-Host (ESX 4.1)?
    >
    >Of course. And Of course, that works with *every* OS.
    >Because the OS has
    >*absolutely no idea*. All it sees is two drives.
    >
    >
    >> How does
    >> Netware recognize the additional disk-space?
    >
    >Exactly in the same way as Netware recognizes more than
    one
    >physical
    >harddrive. It's just a drive to it.
    >
    >CU,
    >--
    >Massimo Rosen
    >Novell Knowledge Partner
    >No emails please!
    >http://www.cfc-it.de

  • Migrate NSS pools to new device?

    Hi everyone:
    This may sound strange, but it is what it is : We're trying to migrate a series of cluster enabled pools from one SAN to another SAN (the fiber connects have already been made and tested...so that's no part of the equation at this time).
    Problem is, we've been asked to do this with ZERO disruption to the users. And we're talking about 30 tb of data here encompassing a cornucopia of volumes and pools servicing a large spectrum of users across a rather massive eDirectory structure (many login scripts to modify among other things).
    SO assuming I can activate pools and mount volumes to a server from both SANs (which I can), is there any possible way to force the NSS Pool to activate on another device if the device has been mirrored from one SAN to the other?
    Right now what happens is that the secondar lun (which has a copy of the data on the primary lun) shows up as a new device, and there doesn't seem to be any way to simply mount the pool that's been mirrored to it.
    Barring that, is there any way to expand the pool to encompass both devices, but then shrink it back down so that it only encompasses the new device?
    Any help would be greatly appreciated.
    Thanks!

    Originally Posted by jgray
    To put it into a nut shell. You mirror the two sans together. Once the mirror is do, you break the mirror and now the data is on the new san. If the sans are from the same vendor, then they should have a tool to tranfer the data from one san to the other.
    More on the mirroring process, in maybe a cassaba shell:
    Mirrors for Moving Partitions - CoolSolutionsWiki
    /dps

  • NSS pool errors during every NSS background check

    We are running a NW 6.5 SP6 server with NSS pools SYS and POOL1.
    POOL1 contains two volumes: VOL1 and MAIL. Both can grow to max. pool size.
    During every NSSbackground check we receive the messages below.
    I checked the novell support site, but can't find anything usefull on this.
    Anyone got an idea what they mean and how to get rit of them?
    Kind regards,
    Mario van Rijt
    systems administrator
    System console
    4-07-2008 16:59:10 : COMN-3.25-178 [nmID=A0020]
    NSS-2.70-5009: Pool MFP02/POOL1 had an error
    (20012(beastTree.c[510])) at block 23579655(file block -23579655)(ZID
    1).
    4-07-2008 16:59:10 : COMN-3.25-180 [nmID=A0022]
    NSS-2.70-5008: Volume MFP02/VOL1 had an error
    (20012(beastTree.c[510])) at block 23579655(file block -23579655)(ZID
    1).
    Logger screen
    4 Jul 2008 16:59:10 NSS<COMN>-3.25-xxxx: comnPool.c[2629]
    Pool POOL1: System data error 20012(beastTree.c[510]). Block
    23579655(file
    block -23579655)(ZID 1)
    4 Jul 2008 16:59:10 NSS<COMN>-3.25-xxxx: comnVol.c[8852]
    Volume VOL1: System data error 20012(beastTree.c[510]). Block
    23579655(fil
    e block -23579655)(ZID 1)

    Originally Posted by Mario vanRijt
    (20012(beastTree.c[510])) at block 23579655(file block -23579655)(ZID
    1).
    Logger screen
    4 Jul 2008 16:59:10 NSS<COMN>-3.25-xxxx: comnPool.c[2629]
    Pool POOL1: System data error 20012(beastTree.c[510]). Block
    23579655(file
    block -23579655)(ZID 1)
    This one zeems to be
    "20012 zERR_MEDIA_CORRUPTED: The media is corrupted."
    This may indicate that it is time to do a pool rebuild, although I don't know enough about what line 510 of beastTree.c is looking for to say much about what the actual problem is.
    /dps
    4 Jul 2008 16:59:10 NSS<COMN>-3.25-xxxx: comnVol.c[8852]
    Volume VOL1: System data error 20012(beastTree.c[510]). Block
    23579655(fil
    e block -23579655)(ZID 1)[/QUOTE]

  • Recovering NSS pool after server crash

    Server crashed the other day - a SLES 11 sp1 XEN host server system with 3 VMs, 2 OES servers (SLES 10 sp3) NSS pools/partitions and 1 pure SLES 11 server. After the crash (due either to power reset or raid controller failure) I had to run fsck to clear up multiple, multiple error conditions on the SLES partitions in order for the servers to reboot. Once booted up, edir and of course NSS would not load. eDir has been recovered and is now running but one of these servers (the one with most all of the data on NSS) is unstable and still has some boot problems. Would the following be a viable option at this point:
    1. Take this server out of the tree (if possible)
    2. Install fresh, new OES server in a logical drive
    3. Join the existing tree
    4. Attach the NSS partition to this server
    5. Bring up NSS
    Any reason why this would not work to get to a stable server that can mount the existing NSS pool/partition? Thanks.
    Don

    That will work, assuming the new server does not have a volume/pool by the same name already.

  • Large NSS Pool

    I have a NSS Pool spanning 4 SAN volumes of 2TB a piece for a total pool of 8tb. No problems, until I need more space--but that's another issue. I wanted see if anyone has recommendations for NSS settings for such a big pool. I am using the defaults. Why I ask is that we are implementing a new Backup solution (Commvault) and our servers that have smaller pools, not spanning partitions get much better throughput.---identical hardware, HBA, RAM, etc.---just not 8tb of SAN volume. More of a ONE to ONE ratio---POOL to VOLUME to SAN VOLUME to Partition.
    Could it be that the POOL spanning and it's size causing slowness?

    Yes, the size and spanning would make this a little slower. If it were me, I would never do it like this. Just think if something happened to that pool and you had to rebuild it. A rule of thumb for a rebuild it 45 min to an hour for every 100G of uged disk space. As that pools starts to fill up, it will take days to repair it. I would not want the down to to do that. Even if you didn't repair it with rebuild, you decided that you were going to delete it an recreate it and then restore the data. That too will take a long time. To get back to the question, yes, it would slow it down.
    jgray

  • NSS Pool Deactivating - take 2

    I'm having a similar situation to the other thread by this name.
    Our server had a failed RAID drive and replaced it. Since then we had issues with some data files. When we tried to back them up or read them the NSS pool would dismount and we'd have to restart the server. I ran "nss /poolrebuild /purge" yesterday on all the pools and the problem persists.
    We're running NW65 and have N65NSS8C installed.
    What's our next step? I'm not quite ready to throw in the towel, back up what we can, and rebuild the server. I gotta believe there's SOMEthing else we can do.

    Originally Posted by zeffan
    I'm having a similar situation to the other thread by this name.
    Our server had a failed RAID drive and replaced it. Since then we had issues with some data files. When we tried to back them up or read them the NSS pool would dismount and we'd have to restart the server. I ran "nss /poolrebuild /purge" yesterday on all the pools and the problem persists.
    We're running NW65 and have N65NSS8C installed.
    What's our next step? I'm not quite ready to throw in the towel, back up what we can, and rebuild the server. I gotta believe there's SOMEthing else we can do.
    What make of server / array? What RAID level? Software or hardware RAID? If you replace a RAID drive and you lose data, or even have an interruption of service, you didn't have RAID. In many instances, if there is an array rebuild failure, the array will tell you this and then indicate that the bad data ( mirror mismatches ) will still exist until its overwritten. HP does this. The problem is that NSS cannot know which blocks are bad or not, and if it reads a bad block, the read fails and the NSS thinks the disk channel has failed ( which it has ).
    Ultimately if the RAID failed to do its job, every block is suspect. Restore from last good backup.
    More details are required to fully answer your question. But assuming you had an array failure that caused a loss of data, you should recreate the array and restore.
    -- Bob

  • Lost NSS Pool After Patching

    Just ran the latest patches on my OES11SP1 server and one of the NSS pools has disappeared. The lost pool was built on top of a linux RAID5 array and worked with no problems until last night. After patching and restarting the server, the pool was no longer listed via nssmu and the raid is no longer listed under devices - a check of the linux array showed that it is up and stable. I tried to rebuild the pool via nsscon, but it said that it does not exist - any ideas on how can I get the pool up and running again?

    Originally Posted by magic31
    I'm assuming this is an OES install, installed directly on the hardware (no virtualisation layer)?
    You mention the raid is no longer listed... but are the disk devices themselves listed? If not in nssmu, are you seeing them listed in the output of "fdisk -l"?
    Along that line... What kind of disk controller(s) do you have in the server? Have you installed any separate drivers for controllers (not included in the SLES install) that might have not loaded with a kernel update?
    If you have the option to open an SR alongside getting pointers here, that would be wise imo.
    -Willem
    This was an install of OES11SP1 directly to the hardware.
    The disk themselves show up in nssmu and mdadm sees the array with no errors.
    The disk controller is on the motherboard (Adaptec chipset) but did not need any special drivers or config during the install.
    What I have seen in the docs is that with OES11SP1, it frowned upon to run NSS on a Linux-software array like I was using. I tried to mount the array using using the mnt command and a file type of "nsspool" but got an error; nsscon says to validate and rebuild the pool, but then comes back and won't let me put it in maintenance mode. I am wondering if I need to delete the old Pool object from NDS, but that won't address the lack of NSS seeing the device...
    -L

  • Mounting NSS Cluster Volumes on other Linux Nodes

    Hi,
    have anyone experience how to mount OES NSS Cluster Volumes on other
    Linux Servers (SLES/RedHat/Centos)?
    Traditional i would say i have to use NFS, but did this run fine on OES
    Cluster Volumes?
    How must i do this here, what are the recommend mount options?
    Documentation say do it "native to Linux"???
    hmm, how?
    How should i write the /etc/exports file on OES Cluster Nodes(should
    these look same on all Cluster Nodes)?
    What happen on Resource Failover, should nfsd restarted?
    Have anyone running this?
    Regards
    Thomas

    Since you're using NSS in a clustered setup, I am assuming you're using just NCP?
    I'm not sure if the ncpmount or the NCL (Novell Client for Linux) work on anything other than the Suse platforms (you mentioned RHEL and CentOS).
    Otherwise you MAY be able to also run CIFS to "publish" the NSS volume and then you should be able to use native SMB connection methods to connect from the other linux OS's.
    I would probably avoid NFS only because it has no concept of file locking (or at least doesn't seem to care) and may happily overwrite a file that's in use by someone with NCP access.

  • Setting up NICs and Volumes on VMware

    Virtual NW 6.5.8/BM 3.8.5 on VMware 4.1. The only TID I found was for
    setting up BM on VMware 2.5.
    1. In the VM Configuration tab, the VM Network shows vmnic0, which is
    used to access the VM machine through vSphere, and vmnic1, which is
    Down. Both adapters are in or associated with vSwitch0.
    In the BM virtual machine, I added the second adapter, but I'm not sure
    if that is correct or whether two additional virtual adapters need to be
    created/used.
    2. When installing BM, I created two cache volumes. However, they are
    listed in iManager as one traditional partition. Is this a problem?
    Even if not, the free space is not considered a pool (the only pool
    listed is SYS), but rather as two free space partitions, which I cannot
    seem to be able to use for two volumes I want to create.
    Any help with either issue will be greatly appreciated.

    On 8/26/2010 11:14 AM, Jack Cunha wrote:
    > Virtual NW 6.5.8/BM 3.8.5 on VMware 4.1. The only TID I found was for
    > setting up BM on VMware 2.5.
    >
    > 1. In the VM Configuration tab, the VM Network shows vmnic0, which is
    > used to access the VM machine through vSphere, and vmnic1, which is
    > Down. Both adapters are in or associated with vSwitch0.
    >
    > In the BM virtual machine, I added the second adapter, but I'm not sure
    > if that is correct or whether two additional virtual adapters need to be
    > created/used.
    >
    > 2. When installing BM, I created two cache volumes. However, they are
    > listed in iManager as one traditional partition. Is this a problem? Even
    > if not, the free space is not considered a pool (the only pool listed is
    > SYS), but rather as two free space partitions, which I cannot seem to be
    > able to use for two volumes I want to create.
    >
    > Any help with either issue will be greatly appreciated.
    To be clear, the two cache "volumes" are not listed as Volumes in
    iManager, only as a single partition.

  • Clustering CIFS for *existing* NSS pools: Howto?

    Hi.
    I found several documents stating this one:
    "Once the CIFS protocol is enabled for the shared pool the load and unload script
    files will get automatically modified to handle the CIFS service."
    e.g. http://tinyurl.com/OES-2SP2CIFS and the docu file_cifs_lx.pdf
    Currently CIFS is installed and running on both nodes of this cluster.
    I can access the CIFS shares, login of users through the CifsProxyUsers is OK
    addressing the ClusterNode, that currently has the NSS volumes mounted.
    What I cannot figure out is how to create a appropriate vitual server object, and
    what lines need to be added to the corresponding cluster ressource of the NSS
    volumes I need to share using CIFS.
    Who could point me towards more detailed documentation, or, provide some example
    load / unload scripts from his/her environment?
    TIA, regards, Rudi.

    Hi.
    I still fail to "cluster" the CIFS functionality.
    BTW: Accessing \\node02-w\ works fine, this is shared, accessible through *ANY* ip
    currently active on that node. Propably that's the issue, because the ports are
    occupied, even if I add another, new secondary.IP? But if rcnovell-cifs is not
    running, the load script line doesn't work either??
    Did I get correctly,
    - that rcnovell-cifs has to *run* on all designated CIFS cluster nodes?
    - that the line ...
    novcifs --add --vserver=.cn=NW06.o=MAIN.t=TREE. --ip-addr=10.27.1.51
    ... will add the resource to the virtual server NW06, so that
    it will service it using CIFS?
    Any further suggestions appreciated, regards, Rudi.
    more details: This is what /var/log/cifs/cifs.log presents:
    rcnovell-cifs running, then executing the line above: (each line shows up 2x, why?)
    CRITICAL: BROWSER: Bind is failed for virual server ip: System Error No= 98
    CRITICAL: NBNS: Bind is failed for virual server ip =10.27.1.51: System Error No = 98
    CRITICAL: NBNS: Socket creation and binding for virtual server failed
    nmap for 10.27.1.51
    137/udp open|filtered netbios-ns
    138/udp open|filtered netbios-dgm
    139/tcp open netbios-ssn
    adding another secondary ipaddress first doesn't make it work either:
    .. /opt/novell/ncs/lib/ncsfuncs
    add_secondary_ipaddress 10.27.99.51
    novcifs --remove --vserver=.cn=NW06.o=MAIN.t=TREE. --ip-addr=10.27.99.51
    cifs.log:
    CRITICAL: ENTRY: Not Able To Conntect To Server 10.27.99.51
    CRITICAL: ENTRY: Not Able To Conntect To Server 10.27.99.51
    ERROR: CLI: CIFS: Server fdn .cn=NW06.o=Main.t=Tree. and server_name has already
    been added
    for that additional sec.IP ports look the same:
    nmap: 10.27.99.51
    137/udp open|filtered netbios-ns
    138/udp open|filtered netbios-dgm
    139/tcp open netbios-ssn

  • NSS pool rebuild

    Hi,
    I have a corrupted pool that need to repair, (vol = 520GB)
    i started it through nss /poolrebuild ->choose the pool appear on list
    it appears not doing anything, I have restarted the server after 20hours later, started all over --> same screen and exact value display i.e no data moment etc
    it seem to me the tool had been stopped due to somesort of log that block it as it write to c:\ drive, could it be to do with .RLF file ?
    could anyone shed somelight on this?

    Originally Posted by mrosen
    Hi.
    Any different result when you directly call nss /poolrebuild=poolname?
    Anything in the logger screen or console then?
    And yes, an issue with C:\ could play a role here, is it full possibly?
    Las but not least, you could try to bring up the server without autoexec
    (server -na from dos), and try then, from the physical console. That way
    you can be reasonably sure no other software may be interfering.
    On 24.11.2013 12:26, rdy1 wrote:
    >
    > Hi,
    > I have a corrupted pool that need to repair, (vol = 520GB)
    > i started it through nss /poolrebuild ->choose the pool appear on list
    > it appears not doing anything, I have restarted the server after 20hours
    > later, started all over --> same screen and exact value display i.e no
    > data moment etc
    > it seem to me the tool had been stopped due to somesort of log that
    > block it as it write to c:\ drive, could it be to do with .RLF file ?
    > could anyone shed somelight on this?
    >
    >
    Massimo Rosen
    Novell Knowledge Partner
    No emails please!
    Untitled Document
    Thanks Massimo,
    I had tried pretty much the same you suggested, the nss /poolrebuild is not even running !!!
    it was so bad the corruption to pool was so severe
    but however, there is a way to mount back that volume even-though your pool is so corrupted.
    we managed to get our data back, thank god for that.

  • Change ownership of NSS pool/vols

    I have two servers connected to an iSCSI SAN - both servers are NW 6.5 SP8. I need to retire server 1 and replace with server 2 and I would like to permanently mount server 1 NSS volumes on server 2 and change the ownership of the eDir objects to reflect server 2. How do I go about doing this?
    Thanks,
    bp

    On 27.05.2011 15:06, BPainter wrote:
    >
    > Yes, one of the volumes does contain user directories. I assume I would
    > need to manually change the paths?
    Well, there are two options:
    1. You can manually (e.g via LDAP) modify the existing volume object to
    point to the new server. However, this isn't perfect, as some activities
    tend to undo this change.
    2. You change the home directory setting of all your users after moving
    the volume. This of course can be done manually, but there are also
    tools that can help you with that. Search for "homes" on coolsolutions
    for instance.
    But generally, yes, moving volumes between servers can be done just easily.
    CU,
    Massimo Rosen
    Novell Product Support Forum Sysop
    No emails please!
    http://www.cfc-it.de

Maybe you are looking for

  • Any way to add a URL variable to EVERY SINGLE url in the site, all at once?

    Hi, I'm managing a web application, and we're looking at changing from using cookies, to passing session variables in every HTTP request to maintain session information. So basically, I have a variable that needs to be appended to every single URL in

  • Multiple (independent) Image Rotation

    Usually I'm good at finding information to solve problems by searching the forums. but honestly I've never seen this problem solved or ever heard any mention of it. . . . here goes. I'm writing a space shooter that uses images as the hero and enemy s

  • Why does Safari redirect to Skype

    i just loaded Mountain Lion. Web sites that require adobe flash player no longer work properly. i am trying to access the adobe site to install the latest version of flash player and safari keeps redirecting me to skype.com.  I read somewhere that i

  • Recovery partition or not?

    My Lenovo T60 has a recovery partition located at the end of its 80GB drive, named SERVICEV001 (a primary FAT 32 partition). I recently purchased a second T60 notebook which is almost identical (the only major difference is that it has a CDRW optical

  • VSV corruption

    Hi all, I think I have gotten myself into a predicament and I am hopeful someone can come to my rescue..:-) Before my PC crashed I created several Virtual PC's and associated VSV files for their saved state. In those states I had a number of applicat