WD mycloud fehler 004 : Data volume failed to mount

Hallo, Seit einiger zeit habe ich nurnoch eine rote LED und kein zugriff auf meine WD Mycloud 2TB ins Mycloud daschboard komme ich noch sporadisch. zeigt aber fehler 004 Haupsächlich hatte ich die mycloud als speicher für fotos genutzt, die natürlich nirgendwo anderst gespeichert sindKomme ich irgendwie wieder an meine Fotos ?? DankeMatthias  

Hast Du mal über das Dashboard "nur System" auf Werkseinstellungen zurück gesetzt? Wenn es dann icht geht, kannst Du den Support kontaktieren. Die werden Dir aber vermutlich nicht dabei helfen Deine Daten wiederzubekommen, sondern bestenfalls ein neues/repariertes Gerät ohne Deine Daten senden. Ansonsten hatte hier jemand das gleiche Problem lösen können:http://community.wd.com/t5/WD-My-Cloud/Data-volume-failed-to-mount/td-p/631497Versuchen kannst Du es: This is "my" solution, may not work for everyone. But it would seem that datablocks would corrupt when power is abruptly cut to the drive. SSH must be enabled for this to work. 1. Log onto the drive, root and welc0me are the default username/password combo.2. Make sure your paritions are intact. # parted -l will tell you this.3. Type # mke2fs -n /dev/sda4. This will get you the filesystem, but more importantly at the end give you the superblock backup locations by block location.4. Pick one, then enter this command. e2fsck -b blocknumber /dev/sda4 It will look for bad blocks and ask for your confirmation to continue/ignore then write over any corrupted ones. Yes to ignore, yes to rewrite. It will do up to five passes of diagnostics, prompting if it finds anything out of the ordinary such as file countes and errant inodes. I answered yes to everything. When it is complete it will either report "Killed" or a summary. Type # reboot and let the drive rebuild its library. 30-ish minutes later (depending on how much data you had on it) you should have access to your files.

Similar Messages

  • A volume failed to mount/Operation cannot be completed

    We bought a MacBook on the weekend. Our Apple concierge insisted that we can keep our music, videos and photos on our Windows XP desktop and "share" the files on the Mac--in other words, we can pull down what we need to edit or save on the Mac as need be.
    We've set up the shared directories on the Windows machine, renamed the Workgroup to match the Mac and disabled the firewall. We've "connected to server" using Finder and can see the two shared directories we've set up.
    But when I click on either one of them, I get an error message: "A volume failed to mount. The volume "Mac" could not be mounted."
    After some time of being on the network, I see a Shared directory in Finder with my PC. If I click on the icon, the title bar says: "Connected as: Guest". The window displays the same two directories as above but when I click on one or the other I get an error message: "The operation cannot be completed because the original item for "Mac" cannot be found."
    I've called Apple technical support and the agent's response was: I don't think you can do that. Just copy your files to the Mac using DVDs.
    Any ideas? We don't want to have to transfer a lifetime of photos, video and music to our laptop--we want to use the XP desktop as our hub.

    I'm having the same trouble. From an Intel Mac Mini running 10.5.3 I've been able to access smb shares fine on an old PC running XP pro. These are shared volumes on external drives on the PC. I can still access them fine from an old PowerBook G4 running 10.5.3, but while I can see computers on the network, if I try to connect to a share on the computer I get the message "the operation cannot be completed because the original item for <share name> cannot be found". If I try to mount it via a shortcut, I get the message "A volume failed to mount. The volume <share name> could not be mounted.
    In the console, I'm seeing things like:
    6/2/08 10:45:01 PM com.apple.coreservicesd[58] NOTE: Using non-mach-based version of client -> server communication, via direct function calls.
    but I don't know if that has anything to do with this or not.
    Has anyone heard of a fix for this?
    Thanks,
    Mark

  • "Volume Fails To Mount" Error

    I have Leopard 10.5.8 and am trying to connect to a Windows Vista share via SMB. The two systems can communicate and authenticate without issue. After using Finder's Go To Server and connecting via SMB, I am prompted for a username and password. Once given, the Vista machine authenticates and presents me with a list of shared folders to chose from. At this point, when I select a folder, Leopard immediately returns with the "Volume Fails To Mount" error and I am returned to the list of shares after closing the dialog box. Thinking it might be an issue with Finder, I found and downloaded an application, AutomountMaker. Its results were the same. I can connect from the Vista machine to Leopard with no issues whatsoever, so I know the machines are working together. I have been unable to find any explanation for this error message nor how to correct the issue. I would appreciate it if anyone can offer some ideas.
    As a bit of additional information, I have another machine that is running Windows 7 and I was able to connect to a share on its drive from Leopard without a problem.

    This page may help http://www.ifelix.co.uk/tech/3022.html

  • Time Capsule "Data" Volume won't Mount

    I've been using a Time Capsule (4th Gen/2TB) for both my rMBP Time Machine backups and as a NAS device. After upgrading to Mountain Lion, I've experienced some problems. For the past few weeks, my Time Machine backups have been frequently interrupted with an error message stating that my Backup Disk is non-journaled and non-extended, and to please choose an alternate disk.  Although unsettling, after pressing return - and upon connecting the power cord (as I opted only to backup when using the power cord), Time Machine has always propmtly resumed.  This has probably occurred daily.
    Today however, I've been unable to actually mount the Data volume, although my network is functioning properly, and I can access the Internet.  The Time Capsule status indicator is green, its icon appears in the Finder sidebar, and Airport Utility recognizes it and displays its correct settings.  Upon trying to connect (via Finder), however, an error message states that the server (in this case, the Time Capsule, I guess) can't be located, and suggets I check my network settings and try again.  Needless to say, the settings appear fine.
    An admittedly brief search within this forum yielded no discussions concerning this specific problem, but I'm hoping the community's more knowledgable members will be able to at least provide some helpful insights, if not a solution to this problem.
    The inability to back up my rMBP, access my spasebundle, or manage my externally stored files is very disconcerting.  Any solutions, or insights regarding this issue will be gratefully received. 
    Michael Henke

    There is clearly a bug in Lion that was exacerbated in Mountain Lion, which Apple is yet to fess up to. This loss of connection happens and there is no fix I can give you other than reboot to get the network running again.
    I would strongly recommend against using the TC as a NAS.. it is designed as a backup target for Time Machine.
    If you want a NAS then buy a true NAS, synology or Qnap, being the top products in that field. They have raid and automatic backup with units where you can access hard disks and can replace them. TC is a sealed unit without any way to back itself up. The design was never as a NAS.
    Since your TC is an older one, can you try running 7.5.2 firmware. None of the TC made in the last year or more can go back that far.. they are all stuck on 7.6, the earlier Gen4 did have 7.5.2 which I think has better stability..
    Now exactly what issues come up with ML I am not sure, but I would be interested to hear your experience.
    Please do a backup particularly of your itunes before you start fiddling.
    Is the TC main router in the network or is it bridged? I have a few suggestions for each type of setup, which may keep it running longer.. but nothing is absolute fix.
    Note also the TC seems to be the main problem device.. but the OS does still have issues.. some changes were made with AFP security at Lion, which have not worked very well.

  • Drives Fail to Mount

    After upgrade, my other hard drives are failing to mount, although they are present and Disk Utility sees them fine. Permissions are fine. Cannot even boot from a CD (anything on the IDE Bus).
    This happens on two PowerMac G4's (one was cloned from the other) so, I know it is not the hardware.

    I realise this might be related to my problem?
    http://discussions.apple.com/thread.jspa?threadID=1249242&tstart=0
    A non-working /Volumes fails to mount anything. Booting from the DVD works however (running on an Intel mini).

  • Home Directory and different volume failed generates log in failure

    I'm running OS X server 10.5.2. The set up is still used for test and learning purposes but I'm nearly finishing and knowing what I can and want to do.
    However, there remains a large issue unresolved.
    I have several volumes available on me server
    1. the startup volume
    2. a data volume
    The users home folders are generated on either the start up or the data volume.
    Login is successful for all users when I log in directly on the server.
    When I login using my test client (MBP running 10.5.2) the accounts with their home located at the data volume are receiving the following messages:
    "You are unable to log in to the user account '<account>' at this time. Logging in to the account failed because an error occurred"
    Anyone who can help me with resolving this issue?
    Anyone who knows where the error is logged?

    I just added the group name and granted that with read and write access to the volume.
    That seemed to be the trick.
    I'll hope this solves some issues for others, I spend months on this one ...

  • RAID 0 volume failed - Intel RST, W530 (wonder if it is hardware issue now)

    Hi,
    My laptop is about two years old. I got one of two intel RAID 0 volumes failed several months after I got this laptop. This problem has been on my laptop for a long time.
    The functionality of my laptop started to become poor after this volume failed.In the beginning, windows explorer stopped working and restarted every time I opened the laptop. The windows repair tool said windows could not be repaired... So I reinstalled the OS and reset it to factory setting. However, the crashes persisted and gradually became more frequent. Many other error reports showed up. After searching for solutions, I finally did a clean windows install and used the BIOS Intel RST to repair the corrupted volume (I deleted old volumes created new ones).
    Unfortunately, now, the same volume fail error appears again and the functionality of my laptop is impacted. Sometime I cannot save my excel or pdf file because program stop working, or cannot open a pdf file. I have to restart the OS.
    The detail error message of Intel RST is: "your volume appears inaccessible. Try fixing any problems reported on the array disks. If this fails, review the Troubleshooting section under Help for more information. In all cases, back up any accessible data before taking action".
    I'm now wondering if it is the hardware issue. If so, I'm thinking about buying an extended warranty because my orignal warranty has expired. However, I'm not sure if the warranty can cover this... 
    In this community, I find there are two types of extended warranty, one will start from the expiring date of the original one and the other will start from the date of purchase. I suppose I need to find the second one.
    I'm also wondering the quality of the warranty. Will it take a long time to repair if I send it? or will it take a long time for them to ship?
    Any idea, solution and experience in warranty / repairment would be appreciated.
    Best Regards,
    Haoyi
    BTW, here is also my experience with Lenovo if you would like to have a look at:
    I got my W530 two months after purchases... I found it was a common problem in Lenovo... 
    When my laptop was still in the original warranty, I contacted the lenovo customer service for repairment of the failed volume. Those people told me it was errors coming from my software. They said the repairment was not covered in the original warranty and I needed to buy a new type of warranty (related to software repairment). I had no confidence in the quality of lenovo's customer serivce... So I did not buy it.

    Hi, Colonel,
    Thank you for reply. I tried HD Tune, but it cannot show any information about my harddisk. It only shows Intel Raid 0 Volume, the firemware version is 1.0 and the capacity is 931.5GB. There is no other information, such as serial number, standard, supported, and buffer. I cannot choose supported features. HD Tune also cannot display the temperature of my drives. What should I do?

  • Xsan Metadata volume failed

    Hello,
    I am in a very bad possition and I need help.
    We have set up Xsan system but the metadata volume was on RAID 0 instaed of RAID 1 - do not ask why.
    Now the drive on the RAID 0 failed and the metadata volume is gone.
    Please tell me if there is a way to save the information on the data volume which is RAID 6 and haven't been touched but i do not have access to it.
    If I connect the RAID system to a computer directly will I see the data and be able to back it up?
    Please please help

    No. You need metadata to know where files' extents live on disk, what their names are, folder hierarchy, etc. Try to get the metadata back. A good data recovery outfit might be able to get data off the failed hard drive and onto a new drive to get the RAID-0 back online.

  • Data Volume - Calculation performance

    Hi,
    We are experience degrading calculation performance as data volume increases.
    We are implementing BPC 7.5 SP05 NW (on BW 7.0. EHP1).
    An allocation script that ran in 2 minutes when the database contained only 800,000 records, took over 1 hour after the database was populated with a full year of data.
    All logics have been written to calculate on a reduced defined scope but it does not seem to improve the execution time. When I check the formula log, the scope is respected.
    The application is not that large either: 12 dimensions, the largest containing 300 members and 3 hierarchical levels.
    We optimize the database but to no avail.
    What can be done to optimize performance? Are there any technical settings in BPC or BW that can be fine-tuned?
    Thanks,
    Regis

    Hi Ethan,
    Take a look at one of the allocation script: http://pastebin.com/TA16xCd3
    We are testing RUNLOGIC but we are facing some problems in two situations:
    - passing the DM package variable to the RUNLOGIC script
    - using a passed variable in the called script
    The DM prompts for 3 selections: ENTITY, TIME and CATEGORY.
    The RUNLOGIC script:
    *SELECT(%DIVISIONS%,"[ID]",DIVISION,"[LEVEL]='DIV' AND [STORECOMMON]<>'Y'")
    *SELECT(%BRANCHES%,"[ID]",BRANCH,"[BRANCHTYPE]='STORE'")
    *START_BADI RUNLOGIC
         QUERY=OFF
         WRITE=ON
         LOGIC=ALLOC_DIV_ACTUAL_S.LGF
         DIMENSION ENTITY=C1000
         DIMENSION TIME=FY10.MAY
         DIMENSION CATEGORY=ACTUAL
         DIMENSION DIVISION=%DIVISIONS%
         DIMENSION DIVISION=%BRANCHES%
         CHANGED=ENTITY,TIME,CATEGORY,DIVISION,BRANCH
         DEBUG=ON
    *END_BADI
    In ALLOC_DIV_ACTUAL_S.LGF, we are using a %DIVISION_SET% variable. At the time of validating, we get a message "Member "" does not exist".
    When we run the package, it fails with the same error message:
    An exception with the type CX_UJK_VALIDATION_EXCEPTION occurred, but was neither handled locally, nor declared in a RAISING clause
    Member "" not exist
    Thanks
    Regis

  • OES2sp2 miggui Failed to mount volume

    Trying to migrate NW65SP8 (DS8.8.4) to OES2sp2.
    miggui is running on the OES2sp2 server which is patched up to date using rug.
    Source and target servers are in the came container in the DS.
    I can define the source and target servers and get to the point where I want to configure the migration of volumes, but then I get a failure. The log (minus the timestamps for clarity) reads:
    ERROR - FILESYSTEM:volmount.rb:Command to mount source: ruby /opt/novell/migration/sbin/volmount.rb -s 138.37.100.118 -a "cn=joshua,ou=sys,o=qmw" -c cp437 -f "/var/opt/novell/migration/cc18/fs/mnt/source" -m -t NW65
    INFO - FILESYSTEM:volmount.rb:*****************Command output start**********************************
    INFO - FILESYSTEM:volmount.rb:
    INFO - FILESYSTEM:volmount.rb:Information: ncpmount using code page 437
    INFO - FILESYSTEM:volmount.rb:Information: ncpshell command executed as: LC_ALL=en_US.UTF-8 /opt/novell/ncpserv/sbin/ncpshell --volumes --ip=138.37.100.118 --u="joshua.sys.qmw"
    INFO - FILESYSTEM:volmount.rb:Information: Mounting Volume = _ADMIN
    INFO - FILESYSTEM:volmount.rb:Information: Executing command: /opt/novell/ncl/bin/nwmap -s 138.37.100.118 -d /var/opt/novell/migration/cc18/fs/mnt/source/_ADMIN -v _ADMIN
    INFO - FILESYSTEM:volmount.rb:Fatal: Failed to mount volume _ADMIN for server 138.37.100.118
    INFO - FILESYSTEM:volmount.rb:Fatal: SystemCallError, Unknown error 1008 - Failed to mount volume _ADMIN for server 138.37.100.118 .
    I tried executing what appears to be the offending command by hand:
    ruby /opt/novell/migration/sbin/volmount.rb -s 138.37.100.118 -a "cn=joshua,ou=sys,o=qmw" -c cp437 -f "K" -m -t NW65 -p password --debug
    which produces the output:
    Information: ncpmount using code page 437
    Information: ncpshell command executed as: LC_ALL=en_US.UTF-8 /opt/novell/ncpserv/sbin/ncpshell --volumes --ip=138.37.100.118 --u="joshua.sys.qmw"
    Information: Mounting Volume = _ADMIN
    Information: Executing command: /opt/novell/ncl/bin/nwmap -s 138.37.100.118 -d K/_ADMIN -v _ADMIN
    Fatal: Failed to mount volume _ADMIN for server 138.37.100.118
    Fatal: SystemCallError, Unknown error 1008 - Failed to mount volume _ADMIN for server 138.37.100.118 .
    Information: File K/_ADMIN/Novell/Cluster/PoolConfig.xml does not exist. No cluster resources attached
    Information: Mounting Volume = SYS
    Information: Executing command: /opt/novell/ncl/bin/nwmap -s 138.37.100.118 -d K/SYS -v SYS
    Fatal: Failed to mount volume SYS for server 138.37.100.118
    Information: unmounting all mounted volumes
    Information: Executing command:/opt/novell/ncl/bin/nwlogout -f -s QMWCC18
    Cannot perform logout: Cannot connect to server:[QMWCC18]. Error:NWCCOpenConnByName:
    Information: Command Output:
    Information: Executing command: rm K/*
    rm: cannot remove `K/*': No such file or directory
    Fatal: SystemCallError, Unknown error 1008 - Failed to mount volume SYS for server 138.37.100.118 .
    SLP shows both servers, they are on the same network as each other, and the firewall is turned off.
    Does anyone have any idea what may be causing the mount failure or what error 1008 might be?
    Tim

    cgaa183 wrote:
    > Thankyou all for helpful suggestions, I'll go through them.
    >
    > 1. novfsd
    > /etc/rc.d/novfsd status
    >
    > running
    >
    > /etc/rc.d/novfsd restart
    > Stopping Novell novfs daemon...
    >
    > done
    > Starting Novell novfs daemon...
    >
    > No Config File Found - Using Defaults
    > novfsd: Novell Client for Linux Daemon
    > Copyright 1992-2005, by Novell, Inc. All rights reserved.
    > Version 3.0.1-503
    >
    > done
    >
    > Nothing changes, so I don't think that was the problem in this case.
    >
    did you restart the server after installing OES2 SP2?
    > 2. try using migfiles
    > /opt/novell/migration/sbin/migfiles -s 138.37.100.118 -v APPS -V APPS
    >
    > The result:
    > Error:
    > Error: nbackup: Unable to retrieve the Target Service Name list from
    > 138.37.100.118
    > Error:
    > Fatal: nbackup command failed to execute: nbackup: Connection denied
    >
    migfiles is not able to connect to the TSA on the source server. Either
    TSAFS is not loaded on the source server or not able to locate the TSA
    using SLP.
    > This might be informative to someone who knows a little more than me. I
    > wonder if I can call it with options to get more information. I have
    > just re-checked I can attach and mount volumes with the same username
    > from another system.
    >
    > 3. ruby version
    >
    > rpm -qa |grep ruby
    > ruby-1.8.4-17.20
    > rubygems-0.9.2-4.4
    > rubygem-needle-1.3.0-1.5
    > rubygem-net-sftp-1.1.0-1.5
    > ruby-devel-1.8.4-17.20
    > rubygem-net-ssh-1.0.9-1.5
    >
    > Doesn't appear to be an afflicted version and migfiles --help tells
    > me all about the options available.
    >
    >
    > As #2 looked interesting I thought I'd look at it a bit more. I turned
    > up TID 7001767. The reason migfiles failed for me was that SMDR and
    > TSAFS weren't loaded, at least I don't get the error and file migration
    > appears to start now they are loaded, though it does appear to have
    > ceased copying files rather prematurely....
    >
    > Going back to volmount.rb I now realise its using ncpfs, and not a
    > lightweight Novell client. So I tried mounting a volume by hand:
    >
    > qmwcc28:~ # ncpmount -S 138.37.100.118 -U joshua.sys.qmw K
    > Logging into 138.37.100.118 as JOSHUA.SYS.QMW
    > Password:
    > ncpmount: Server not found (0x8847) when trying to find 138.37.100.118
    >
    > A bit of a giveaway, but why doesn't it work?
    > It seems I need to use -A DNSname -S servername and then it works.
    > The next important bit seems to be
    > /opt/novell/ncpserv/sbin/ncpshell --volumes --ip=138.37.100.118
    > --u="joshua.sys.qmw"
    > which executed by hand lists volumes correctly with the output:
    > Please enter your password ?
    > [0] SYS
    > [1] _ADMIN
    > [2] APPS
    > 3 NCP Volumes Mounted
    >
    > "ncpshell" appears to be from Novell's client for Linux so I don't
    > understand why we'd be trying to use that if we're using ncpfs, and we
    > already know which volumes are mounted by looking in the folder in which
    > we mounted the server using ncpfs. AFAICS is used to invoke NLMs on
    > NetWare remotely using OES so its not testing anything we don't already
    > know.
    >
    > This takes me inevitably to "nwmap". "nwmap" is also from Novell's
    > client for Linux so maybe the ncpfs stuff is unnecessary.
    > /opt/novell/ncl/bin/nwmap -s 138.37.100.118 -d sys -v SYS
    > produces:
    > map: server not Found:138.37.100.118 - drive sys not mapped
    >
    ncpmount uses udp as the default. Add the option -o tcp to the ncpmount
    command then mount should work.
    > nwmap doesn't ask for a username. Maybe I'm wrong, but as far as the
    > Novell client goes I don't think it can have attached or logged into the
    > source server (ncpfs having a different connection table and ncpshell
    > asking the remote server to return the answer). I can't actually see
    > where /volmount.rb is calling nwmap at the moment but the results I get
    > my calling it at the command prompt with the same options given in the
    > log are the same.
    >
    if there is an existing connection to the same tree, nwmap does not ask
    for user name. Use the command "nwconnections" to check the existing
    connections. Use nwlogout to logout the connection. check
    /var/opt/novell/nclmnt/root/ for any stale entries.
    > I've tried logging in using nwlogin, but that fails too saying:
    > Cannot perform login: The system could not log you into the network.
    > Make sure that the name and connection information are correct, then
    > type your password again.
    >
    > ncl_scanner -T does list NDS trees but I suspect its only querying an
    > SLP server and nothing more useful. ncl_scanner -S produces:
    > INFORMATION FOR SERVER [QMWCC18] :
    > Server Name : [QMWCC18]
    > Bindery : [FALSE]
    > eDirectory Context : []
    > should it show a context?
    >
    > Looking at the files of the Novell client on the system, it looks a
    > rather cut down set with no config files. Even having introduced
    > protocol.conf the situation is not improved, but I'm now sure the
    > problem lies in this area. Possibly a full client installation is
    > required, or maybe there is something else wrong which is preventing the
    > client from working correctly. namcd is looking suspect.
    >
    >
    You do not need all files for Novel Client. If you want
    You can logout all connections using the command "nwlogout -a" and try
    the nwmap command again.
    "/opt/novell/ncl/bin/nwmap -s 138.37.100.118 -d
    /var/opt/novell/migration/cc18/fs/mnt/source/SYS -v SYS"
    Looks like the novell client is failed to resolve the IP address. You
    can do the following to configure different name resolution methods in
    the following way:
    Create a file: /etc/opt/novell/ncl/protocol.conf with data.
    Name_Resolution_Providers=NCP,SLP,DNS
    then restart rcnovfsd deamon using the command "rcnovfsd restart"
    do you see any NCP errors in the network packet trace?
    regards
    Praveen

  • Submit Button error - Data Connection failing

    I am the form designer and am currently testing a new form I have created, however there is an error with my submit button which Saves the form to SharePoint, I receive the following error message:
    InfoPath cannot submit the form.
    An error occurred while the form was being submitted.
    The form cannot be submitted to the following location: http://servername/siteA/siteB/siteC/FormLibrary/FormName.xml
    The file: http://servername/siteA/siteB/siteC/FormLibrary/FormName.xml" is checked out for editing by DOMAIN\MyUserName.
    The operation completed successfully.
    The form is not checked out to me, or open by anybody else, also check out functionality is not forced on this SharePoint library.
    On searching forums I have read about using friendly server names and I am using the server's name in the URL, i.e.
    http://servername instead of the alias
    http://intranet - this had solved the problem in other forms, but not this one.
    Looking closely at the rules on my submit button, the following actions occur:
    1st rule checks the selection of certain fields, if the conditions are right it will set the values of 3 different fields.
    The 2nd rule saves to SharePoint using a data connection (please note this data connection works fine on different submit buttons in different views). The second rule also closes the form.
    To test where the error occurs I added a series of actions to display messages, and the problem is definitely on the data connection rule.
    The data connection is to save to a Sharepoint library, using a field called "filename" as the file name. The data connection allows overwrites. The file name is set the first time a file is saved (it has a condition which checks a field which
    gets set upon any save event - if this field is set to 0 filename gets set and this field gets set to 1 - because this field is now equal to 1, the filename field will nwever get set again).
    I have 2 views on the form, one named phase 1, the other named phase 2. When the custom submit button is pressed on the phase 1 view, my data connection works, but on the save 2 view, the exact same data connection fails with the error message above.
    I am using Windows 7, InfoPath 2010, SharePoint 2010, InfoPath filler form).
    In Windows XP, for a different user, the error messaeg is a little different, and the form does actually save:
    InfoPath cannot submit the form
    An error occurred while the form was being submitted
    The form cannot be submitted to the following location: http://servername/siteA/siteB/siteC/FormLibrary/FormName.xml
    The file is in use by another application or user.
    As I said the form does actually save its changes but still throws up this error message which will be very confusing to end users.
    I've exhausted all avenues that I know of to find the cause of the error messages and I'm at a loss, please help!
    Thanks,
    Ben

    bgvball14. I found that the error was down to the friendly name / server name issue. I have made sure that everything uses the friendly name e.g.
    http://intranet instead of the servername e.g.
    http://servername. I made sure the data connection and all options in the publishing wizard are set to the friendly name. There were still some forms that didn't work though - this was because when trying to fix this problem I had switched several times
    between using the friendly /server names and some forms were looking for the template with the server name (these didn't work), other forms looked for the friendly name (these did work).
    What I did was open my form library in windows explorer view and then drag each form into Notepad so I could see the source code. On the top two lines I could see the reference to the template and in some cases the template weas referenced using the server
    aname, which I changed to the friendly name and this worked. I did have to check every form in the library. Basically, everything should reference the friendly name (e.g.
    http://intranet.
    Ben

  • External Data Refresh Failed. We cannot locate a server to load the workbook Data Model. ThisWorkBookDataModel

    Hi All,
    I have been trying to fix this for days now. I have tried solutions in many articles but to no avail. So while the error message is something you may have seen may times, I just can't find a solution in my case.
    This is the error:
    And in text just in case the image isn't viewable:
    "External Data Refresh Failed. We cannot locate a server to load the workbook Data Model. We were unable to refresh one or more data connections in this workbook. The following connections failed to refresh: ThisWorkBookDataModel."
    What is worse is I have checked the ULS (SharePoint Trace Logs), the Event Viewer Logs and the OWA Logs and I cannot find a specific error that would help pin point the problem.
    Excel Workbook
    So what am I doing? I have an Excel 2013 workbook and I create an "SQL Server" connection to the AdventureWorksDW database, add a pivot table and a pivot chart, test in in Excel and all works fine.
    I save the Excel workbook to SharePoint 2013 and then select "Data" then "Refresh All Connections" and then I get the error in the picture above.
    Even more puzzling is I have another Excel workbook that also has pivot tables and pivot charts in the AdventureWorksDW2012Multidimensional cube database in "SQL Analysis Services" and this works fine. Hmmm.
    My Environment
    My environment is Windows 2008 R2 Server, SharePoint 2013 with the April Service Pack1 and a separate server with OWA2013 SP1. It has an SQL Server 2008 R2 database which has been upgraded to SQL Server 2012.
    Data Model Settings
    In Excel Services this is set to my server name which is "server-name". As I do not have instances all I can enter is the server name. As this works everywhere else including the workbook outside of SharePoint I do not think this is the problem.
    But I could be wrong.
    Unattended Account
    I have set this up for the PowerPivot Services App and Excel Services App.
    ODC Connections in Excel
    I have tried all 3 authentication modes, Windows, Secure Store ID and "None" which is the unattended account. I have not tried the other connection types, should I?
    Not in WOPI
    I am not in WOPI mode.
    AD Accounts
    I have added permissions in the SharePoint Services and SQL Server, and as they work in Excel outside of SharePoint, I do not think it is a permissions issue. I could be wrong of course, but the problem is in one of SharePoint, OWA, AD,
    SQL Server, Excel, and Windows Server.
    Isolate the Error
    Below is a list of errors I think are relevant but they do not tell me much. The SharePoint logs are not really giving me an error that tells me what to do and where to do it, or even why it cannot refresh, (perhaps not noticeable by the untrained eye).
    Problem with SQL Server Not Analysis Services
    So my cube database in analysis services works fine in SharePoint/OWA but not the databases in sql server. This is my best clue but I have no idea what it means. Why would it work with an Analysis Services connection but not an "SQL Server" connection?
    It Works Outside of SharePoint
    If I run the excel worksheet outside of SharePoint all works fine. When inside OWA this is where the refresh error occurs.
    Errors from Event Viewer on SharePoint Server using ULS Viewer
    "Failed to create an external connection or execute a query. Provider message: There are no servers available or actively being initialized., ConnectionName: , Workbook:"
    "Refresh failed for 'ThisWorkbookDataModel' in the workbook 'http://server...'. [Session: 1.V22.26itT0lx8piNFeqtuGVhN214.5.en-US5.en-US36.98c0e158-9113-46e9-850e-edda81d9ed1c1.A1.N User: 0#.w|ad\testuser1]"
    And an error in the ULS under the "Data Model" category:
    "--> Check Deployment Mode (server-name): Fail (Expected: SharePoint, Actual: Multidimensional)."
    This last error, as it turned out, defined the problem concisely, although I was yet to work out what it meant in some detail.

    I finally solved this myself (or should I say with the help of several key articles).
    The refresh did not work because the database was not in "SharePoint Mode". Yes, SQL Server has modes, 3 of them in fact.
    If you installed SharePoint to the default SQL instance which would be called <servername> then you cannot use this default instance for Excel 2013 workbooks in OWA 2013 because the refresh only works if the database is in SharePoint mode.
    So what are these 3 modes? The Deployment Mode property in the msmdsrv.ini file has them as:
    0 = Multidimensional mode (the default whenever you install SQL Server normally)
    1 = PowerPivot for SharePoint mode
    2 = Tabular mode mode
    How do you know what mode you are in? That's easy, open SQL Studio Manager and connect to all your SQL database engine instances (ignore Analysis Services or SSRS as they are not database engines). If you only have the default instance then that is almost
    definately in Multidimensional mode which is the default and what SharePoint installs its databases to.
    You must have an instance called <servername>\POWERPIVOT. This instance is the "sharepoint mode" needed, and the default instance name when you install an SQL instance in this mode.
    If you don't see <servername>\POWERPIVOT in SQL server then you are not in "sharepoint mode". It is more accurate to say, you do not have an instance that is in sharepoint mode. This is because you cannot simply switch modes on an SQL server.
    You have to install a new instance in the required mode, thats the only way.
    That's easy enough. Load up the SQL Server setup CD and run setup. Install a brand new instance and select "SQL Server PowerPivot for SharePoint" when you get there in the wizard.
    Now you will have the default instance that stores all the SharePoint databases and that is in mode 0, and a new instance called <servername>\POWERPIVOT that is in mode 1. The "<servername>\POWERPIVOT" instance connection is what you
    will use for Excel 2013 when rendering in OWA 2013.
    You also need to ensure OWA 2013 is not in WOPI mode for Excel worksheets. See the last link below for more information about WOPI.
    Next you should go to the Excel Service App in CA and click Data Model Settings and add the <servername>\POWERPIVOT instance.
    Then you have to either turn off the firewall on the SQL server machine, or create an inbound rule on the Windows firewall to open the TCP port for the <servername>\POWERPIVOT instance:
    1. Start Task Manager and then click Services to get the PID of the MSOLAP$InstanceName.
    2. Run netstat –ao –p TCP from the command line to view the TCP port information for that PID.
    Finally, you can now create Excel 2013 workbooks that run in OWA without refresh errors, as long as you are connecting to the <servername>\POWERPIVOT instance. Hooray.
    REFERENCES
    Look for the string "There are no servers available or actively being initialized" in this article:
    http://blogs.msdn.com/b/analysisservices/archive/2012/08/02/verifying-the-excel-services-configuration-for-powerpivot-in-sharepoint-2013.aspx
    Determine the server mode:
    http://msdn.microsoft.com/en-au/library/gg471594(v=sql.110).aspx
    Install the SharePoint PowerPivot instance (aka SharePoint mode)
    http://msdn.microsoft.com/en-au/library/eec38696-5e26-46fa-bc83-aa776f470ce8(v=sql.110)
    Open the port for the new SQL instance:
    http://msdn.microsoft.com/en-us/library/ms174937(v=sql.110).aspx
    Turn Off WOPI for Excel OWA
    http://blogs.technet.com/b/excel_services__powerpivot_for_sharepoint_support_blog/archive/2013/01/31/powerpivot-for-sharepoint-browser-refresh-fails-data-refresh-not-supported-in-office-web-apps.aspx

  • Excel 2007 = Refresh All = Error Message "Initialization Data Source Failed"

    Excel 2007 => Data Tab - Refresh All => Error Message "Initialization Data Source Failed
    2nd Error message:
    The following data ranged failed to refresh:  ExternalData1 - Continue to refresh all
    How do I resolve this? Trying to connect to Oracle DB environment 12C

    What will happen if you rebuild the PivotTable/Table ?
    Make sure the external database is available and the Oracle Provider for OLE DB is installed correctly.
    Wind Zhang
    TechNet Community Support

  • Data load failed while loading data from one DSO to another DSO..

    Hi,
    On SID generation data load failed while loading data  from Source DSO to Target DSO.
    Following are the error which is occuuring--
    Value "External Ref # 2421-0625511EXP  " (HEX 450078007400650072006E0061006C0020005200650066
    Error when assigning SID: Action VAL_SID_CONVERT, InfoObject 0BBP
    So, i'm  not getting  WHY in one DSO i.e Source  it got successful but in another DSO i.e. Target its got failed??
    While analyzing all i check that SIDs Generation upon Activation is ckecked in source DSO but not in Target DSO..so it is reason its got failed??
    Please explain..
    Thanks,
    Sneha

    Hi,
    I hope your data flow has been designed in such a way where the 1st DSO as a staging Device and all transformation rules and routine are maintained in between 1st to 2nd dso and sid generation upon activation maintained in 2nd DSO.  By doing so you will be getting your data 1st DSO same as your source system data since you are not doing any transformation rules and routine etc.. which helps to avoid data load failure.  
    Please analyze the following
    Have you loaded masterdata before transaction data ... if no please do it first
    go to the property of first dso and check whether there maintained sid generation up on activation (it may not be maintained I guess)
    Goto the property of 2nd Dso and check whether there maintained sid generation up on activation (It may be maintained I hope)
    this may be the reason.
    Also check whether there is any special char involvement in your transaction data (even lower case letter)
    Regards
    BVR

  • Project Server 2013 - External Data Refresh Failed

    Hi, 
    I have built  new Project Server 2013 in a 3 tier farm, with OWA, configured PWA site everything working fine except, Reports in BI Center, where i am getting - External Data Refresh Failed error
    http://Sp2013/PWA/ProjectBICenter/Sample%20Reports/Forms/AllItems.aspx
    i have stopped using Excel Web also to use Excel Services, get below error
    External Data Refresh Failed
    An error occurred while accessing application id ProjectServerApplication from Secure Store Service. The following connections failed to refresh:
    Project Server - Issue Data
    Project Server - Risk Data
    Learn
    more about data refresh
    Steps Tried:
    1.deleted Excel Services - recreated
    2.Deleted Secure store - recreated
    Built  test box without OWA still have the same issue
    Any suggestions/ Solutions greatly appreciated.
    Thanks 
    Aleem
    Alim

    Hi,
    If you are still facing the same issue follow the below steps:
    1) Access PWA from your server
    2) Download any built in report from Reports section (in your case
    Project Server - Issue Data Project
    Server - Risk Data) to your server.
    3) Open report in excel, go to Data->Connections in Workbook Connection dialog open "Project Server - Issue Data" go to properties.
    4) Here select tab 'Definition' under connection file remove tick for 'Always use connection file'
    5) Tweak your connection string & tick save password option 
    6) Excel Services: Authentication Settings (suit yourself)
    Now, save & publish on PWA, hope it works :)

Maybe you are looking for