Check-out originals from Content Server to File server?

Hello Experts,
We have urgent requirement to check-out huge number of originals from the Content server and send to an application server (ftp server).
The input data that we can use to fetch the originals are  PHOIO nunber or LOIO or Original File name.
I tested below function module but they does not seem to meet our requirements.
BAPI_DOCUMENT_CHECKOUTVIEW2  - Originals are checked-out to C:\Temp\
CV120_KPRO_CHECKOUT_TO_SERVER
CV120_KPRO_CHECKOUT_TO_TABLE
SDOK_PHIO_LOAD_CONTENT - here the content is written to table in binary mode.
CV120_DOC_GET_FILE
Also i tested program - DMS_KPRO_READ. In this case, manual action is required to select the path in front end where the originals needs to be checked-out.
Please help if you have suitable solutions.
Thanks!
sandeep

Hello Sandeep,
You can use "DMS_KPRO_READ" report to achive your purpose. In this report you will get PHIO ID & LOIO ID. with these ID's you will get originals path to download the files uploaded on content server to application server or FTP server.
Use F_BROWSE function to select all files in foreground/background to downlaod it on desktop. I have implement in my project & it is working fine till date.
Thanks
Pravin

Similar Messages

  • DMS - need to check out originals using BSP

    Hi,
    I am trying to check out DMS originals through a BSP application. For this purpose, I use the function module '‘BAPI_DOCUMENT_CHECKOUTVIEW2’'. When this FM is used in a SAP GUI report program, the originals are checked out into the TEMP directory. But when the same FM is used in a BSP application, the FM fails with the exception 'Document Program not started via RFC. No return possible'.
    Alternatively, I tried using the method 'checkoutview' (in the class 'cl_incl_doc_files'). The method executes without any error, but it does not check out the originals. The original is not available in the 'TEMP' directory. Will this method check out originals to the 'TEMP' directory?? If so, is there any specific thing that needs to be set? Have anyone tried using this?
    My requirement is that the originals need to be downloaded to the local machine.
    Any help is highly appreciated.
    Regards,
    Vidhya.

    Dear Pranav,
        I checked these settings. They are already maintained in SM59. Still it is not working out. I tested the connection also and it is fine. What else could be the reason??
    I am also trying an alternate method. I am first downloading the contents of the file (in binary format) into an internal table using the function module 'SDOK_PHIO_LOAD_CONTENT'. Then, by calling the function module 'GUI_DOWNLOAD', I try to download it as a PDF file. Again, when I try this in a report program, it is successful .When the function 'GUI_DOWNLOAD' (or the method 'cl_gui_frontend_services=>gui_download') is called from a BSP application, it throws the exception 'CONTROL_FLUSH_ERROR'. Is it because through BSP application we cannot download to the local machine?? Is there any other way to achieve this?? I tried all possible ways. In no method, I get full solution.
    This is urgent. Highly appreciate your help.
    Regards,
    Vidhya.

  • How to Check out  standard business content in BI Machine???

    Hi Guyz
    Please let me know
    How to Check out  standard business content in BI Machine???
    Regards
    Jawed

    Hi Jawed,
    Hope this helps - http://help.sap.com/saphelp_nw04/helpdata/en/80/1a66d5e07211d2acb80000e829fbfe/frameset.htm
    http://help.sap.com/saphelp_nw04/helpdata/en/c1/ea683cc5e8ca68e10000000a114084/frameset.htm
    Can you please tell us what do you mean by Checking?
    Bye...
    Message was edited by:
            Naga Timmaraju
    Message was edited by:
            Naga Timmaraju

  • Error in Kpro server when checking in originals from NW Portal

    Hi Experts,
    I am not able to check in originals in my NetWeaver Portal.
    When I try to save I get the follwoing error message:
    - Error in Kpro server
    But if I try in to check in the same original in the PLM system through SAP GUI it works just fine.
    Any thoughts about what I can do?
    Thanks in advance,
    Mikael

    Hi Mikahel,
    I have tried changing the whitelist and now I am able to check-in document to a content server storage category but not able to display it..
    Could you please tell waht additional setting you have done with data carrier?
    We have maintained PC and Default front end..
    Regards,
    Aby

  • Source path too long :-Unable to delete mutiple number of folder and files from windows 2008 R2 file server

    Hi Team,
    we have a file server on windows server 2008 R2, I have copied some data from one server to another server using robocopy . I have checked access permission it seems okay no issues with access permission. but when i am trying to delete these folder i am
    getting below error message
    Error message:- 
    The source file name are larger than is supported by the file system.try moving to a location which has a shorter path name or try renaming to shortcut names before attempting this operations
    Regards,
    Triyambak 
    Regards, Triyambak

    I have tried everything , but nothing help , getting same error 
    could any one help me regarding this.
    Regards, Triyambak
    I have not looked at the other thread mentioned, but usually when this ends up happening, There are several ways to go about it.    One is to map a drive to the UNC path deeper into the folder structure.   
    So instead of C:\long\path\that\we\dont\want\files\in
    You map a drive.. lets say M:\   to     \\computer\c$\long\path\that\we\dont\want
    Now, when you open Explorer to M:\ you've elimintated the length of the path down to M:\files\in which is completely usable.
    The other way, is to shorten the names of everything.    For example...
    Folder structure like this:
    Root
    -----Folder1
    -----Folder2
    --------------Folder2A
    -----Folder3
    --------------Folder3A
    Rename all the folders to shorten up the path.     Rename Folder1 to '1' Rename Folder2 to '2', Rename folder3 to '3.    Then try to delete.  If it's still too long,  go one folder deeper.   Rename Folder2A to '1'  and
    Rename Folder3A to '1' and so on.     Basically keep renaming everything to a single digit character and eventually you'll shrink the path down to where you can shift-delete everything remaining.
    Hope that helps.    
    Brian / ChevyNovaLN

  • Item Check out/in in content areas published as portlet

    I have a content area I have published as a portlet. I have noticed that if I add an item (a file) and enable item check out, the icon to check out the file doesn't appear in the portlet. You have to open the content area in the content area navigation page to get the check out/in icon (and a few others.) Is there a way to do this inside the portlet?
    Thanks
    null

    This is the intended behaviour, to simplify the portlet interface. You'll notice, too, that you can't edit a checked-out item in the portlet.
    This will all change in the 3.1 release of the product.
    Regards,
    Jerry

  • Authorization: Can I check-out processes from Solution in Read mode?

    Hi NG.
    I am building a new set of user roles, but seems to get stuck with this one.
    With T-code: SOLMAN_DIRECTORY I would like to issue read access to users, but still give them the possibility to check-out processes to the maintenance project.
    So far - I have managed to get the read access right, but the check-out button is greyed out (Inactive).
    My wish is to keep the users from changing the contents and set-up of the Solution, but still be able to go look for a process and request the check-out when found.
    Any help would be greatly appreciated!
    Best Regards
    Per Hjorth Christiansen

    Hi Per Hjorth,
    when you activate the Check-out/Check-in functionality the complete process structure of the solution will only be available in display mode. So you don't need additional authority checks to pretend changes in the solution.
    In general, the authority object AI_SOL_DIR is used in Solution Directory. The activities "02 Change" and  "03 Display" are available for the solution structure. In addition, an extended maintenance activity is necessary to be able to activate/deactivate the check-out/check-in functionality for a solution this is "36 Extended Maintenance".
    Best regards,
    Michael

  • Using File Adapter to transfer files from directories on unix file server.

    I have this scenario. I have a PI server on hpus45 and I want to use it to transfer files from a source directory to a destination directory on a unix file server. The unix file server is on a different box eg hpus30.
    What do I configure in SLD and Integration Directory? All the examples I have seen is transfering files from directories on the PI server but none on a different box.'
    Is it possible? I need help asap. I need to determine what the best approach is by today.

    Hi Agasthuri Doss,
    I'm using PI 7.1. I'm not seeing your instructions in ID.
    The objects in PI 71. are in this order
    Unassigned Objects
    Party
    communication component
    communication channel
    receiver determination
    interface determination
    sender agreement
    receiver agreement
    direct connection
    integrated configurations
    configuration scenario
    value mapping group.

  • Using Product Structure for checking out originals of child DIR's

    I have a requirement of creating a package of documents.
    Some of my DIR's are linked to service notification, some are linked to Purchse Order, & some are linked to Sales order.....
    Each service notification, PO, SO is linked to one master DIR, so that they can be seen under Product Structure.
    Is there any method to extract(Check Out) all the originals of all these DIR's at one place using Product structure.
    Or can we atleast create a link between Master DIR and all child DIR's.
    Please giude me.
    Points for sure.

    Hi,
    as far as I know there is no way to check out a product structure, but you can check out a document structure.
    If you DIR has a bill of documents, you can use the BAPI CheckOutView2. The parameter GetStructure determines weather the structure is read or not.
    - GetStructure = 0 Document structure is not read
    - GetStructure = 1 Document structure is exploded one level
    - GetStructure = 2 Document structure is exploded completely
    Regards Iring

  • Migrating from SAP DMS to File Server

    Hi,
    Our client is currently using the SAP DMS to store documents for the EHS module. This has brought with it lot of handicaps.. primarily the huge size of the database. There is an initiative now to move from SAP DMS to a file server where the files would be physically stored whereas the linking would be made via SAP EHS module.
    This effectively means that the things would remain as they were.. only that the specifications in EHS would now be pointing to a new location for its documents.
    My question is whether this is a techinallly feasible solution and second how should the migration be done. keeping in  mind that the linkages should be kept intact.
    Suggestions are solicited... points are guaranteed.
    Thanks!
    Aviroop

    Hi Aviroop,
    Got any idea about how File server works with DMS ?
    I am in urgent need for details about the file server functionality and configuration.
    Can I use a sharepoint site (internet site0 as a file server?
    Pls send me some documents regarding this to my mail id [email protected]
    Thanks in advance and warm regards
    Purnendu

  • Cannot save files to Windows server 2003 file server from MAC

    Hi, 
    We have few MAC users and others mostly windows based users. Problem now is MAC users cannot save any files to the file server. I had the issues to save in the portable hard drive then i format the HDD to fat 32 and now i can save, But the file server cannot
    be done like that. 
    What i can do to save the files to file server? please advice me
    thanks 
    saththiyan

    I cannot attach another NAS as CEO need to save files to all the folders which is there in the file server already, 
    You can share the folder and map a drive across the network to access the shared folder on the file server. Shared folders formatted NTFS do not affect network access. You would do that by using the SMB. Here's how:
    How to Map Network Drive in Mac OS X
    http://www.technipages.com/how-to-map-network-drive-in-mac-os-x.html
    What OS version is the server? Is it a DC? You may need to reduce security on the server to allow Macs to access the server. Here's more info:
    How to Disable SMB 2.0 on Windows Vista/2008/2008 R2
    http://www.petri.co.il/how-to-disable-smb-2-on-windows-vista-or-server-2008.htm 
    "To disable SMB ... Otherwise, disable SMB signing on Windows Server 2003 domain controllers.
    http://support.microsoft.com/kb/325379
    How to enable Windows 98/ME/NT and MAC clients to logon to Windows 2003 based Domains
    http://support.microsoft.com/default.aspx/kb/555038
    Ace Fekay
    MVP, MCT, MCITP/EA, MCTS Windows 2008/R2 & Exchange 2007, Exchange 2010 EA, MCSE & MCSA 2003/2000, MCSA Messaging 2003
    Microsoft Certified Trainer
    Microsoft MVP - Directory Services
    Technical Blogs & Videos: http://www.delawarecountycomputerconsulting.com/
    This post is provided AS-IS with no warranties or guarantees and confers no rights.

  • Check-out button is not disable after the check out request from the same user

    HI all,
            I want to use check out/ in procedure in solution manager along with email notification. For that, I have maintained required settings and maintenance project for the intended solution. I want to implement a team leader- manger hiearchy for any changes to the solution. When a team member requests for the check out for any process scenario/ step, the status goes to check out requested. But, the problem is, check-out button is not disable and same user can do the check out which I want to restrict. The check out can be done only by the project manger. I guess, this is related to authorization roles for both team member and manger. But I am not able to find out. Please help me to have the authorization roles for this.
    Thanks,
    Rutvik Upadhyay.

    Why not putting a af:fileDownloadActionListener inside the command button. In the bean you need an method like
        public void exportExcel(FacesContext aFacesContext, OutputStream aOutputStream)
    } You can use the output stream and when you finished just flush it. Don't close the stream and don't complete the response.
    Timo

  • Server 2012 File Server Cluster Shadow Copies Disappear Some Time After Failover

    Hello,
    I've seen similar questions posted on here before however I have yet to find a solution that worked for us so I'm adding my process in hopes someone can point out where I went wrong.
    The problem: After failover, shadow copies are only available for a short time on the secondary server.  Before the task to create new shadow copies happens the shadow copies are deleted.  Failing back shows them missing on the primary server as
    well when this happens.
    We have a 2 node (hereafter server1 and server2) cluster with a quorum disk.  There are 8 disk resources which are mapped to the cluster via iScsi.  4 of these disks are setup as storage and the other 4 are currently set up as shadow copy volumes
    for their respective storage volume.
    Previously we weren't using separate shadow copy volumes and seeing the same issue described in the topic title.  I followed two other topics on here that seemed close and then setup the separate shadow copy volumes however it has yet to alleviate the
    issue.  These are the two other topics :
    Topic 1: https://social.technet.microsoft.com/Forums/windowsserver/en-US/ba0d2568-53ac-4523-a49e-4e453d14627f/failover-cluster-server-file-server-role-is-clustered-shadow-copies-do-not-seem-to-travel-to?forum=winserverClustering
    Topic 2: https://social.technet.microsoft.com/Forums/windowsserver/en-US/c884c31b-a50e-4c9d-96f3-119e347a61e8/shadow-copies-missing-after-failover-on-2008-r2-cluster
    After reading both of those topics I did the following:
    1) Add the 4 new volumes to the cluster for shadow copies
    2) Made each storage volume dependent on it's shadow copy volume in FCM
    3) Went to the currently active node directly and opened up "My Computer", I then went to the properties of each storage volume and set up shadow copies to go to the respective shadow copy volume drive letter with correct size for spacing, etc.
    4) I then went back to FCM and right clicked on the corresponding storage volume and choose "Configure Shadow Copy" and set the schedule for 12:00 noon and 5:00 PM.
    5) I noticed that on the nodes the task was created and that the task would failover between the nodes and appeared correct.
    6) Everything appears to failover correctly, all volumes come up, drive letters are same, shadow copy storage settings are the same, and 4 scheduled tasks for shadow copy appear on the current node after failover.
    Thinking everything was setup according to best practice I did some testing by changing file contents throughout the day making sure that previous versions were created as scheduled on server1.  I then rebooted Server1 to simulate failure.  Server2
    picked up the role within about 10 seconds and files were avaiable.  I checked and I could still see previous versions for the files after failover that were created on server1.  Unfortunately that didn't last as the next day before noon I was going
    to make more changes to files to ensure that not only could we see the shadow copies that were created when Server1 owned the file server role but also that the copies created on Server2 would be seen on failback.  I was disappointed to discover that
    the shadow copies were all gone and failing back didn't produce them either.
    Does anyone have any insight into this issue?  I must be missing a switch somewhere or perhaps this isn't even possible with our cluster type based on this: http://technet.microsoft.com/en-us/library/cc779378%28v=ws.10%29.aspx
    Now here's an interesting part, shadow copies on 1 of our 4 volumes have been retained from both nodes through the testing, but I can't figure out what makes it different though I do suspect that perhaps the "Disk#s" in computer management / disk
    management perhaps need to be the same between servers?  For example, on server 1 the disk #s for cluster volume 1 might be "Disk4" but on server 2 the same volume might be called "Disk7", however I think that operations like this
    and shadow copy are based on the disk GUID and perhaps this shouldn't matter.
    Edit, checked on the disk numbers, I see no correlation between what I'm seeing in shadow copy and what is happening to the numbers.  All other items, quotas, etc fail and work correctly despite these diffs:
    Disk Numbers on Server 1:
    Format: "shadow/storerelation volume = Disk Number"
    aHome storage1 =   16 
    aShared storage2 = 09
    sHome storage3 =   01
    sShared storage4 = 04
    aHome shadow1 =   10
    aShared shadow2 = 11
    sHome shadow3 =   02
    sShared shadow4 = 05
    Disk numbers on Server 2:
    aHome storage1 = 16 (SAME)
    aShared storage2 = 04 (DIFF)
    sHome storage3 = 05 (DIFF)
    sShared storage4 = 08 (DIFF)
    aHome shadow1 = 10 (SAME)
    aShared shadow2 = 11 (SAME)
    sHome shadow3 = 06 (DIFF)
    sShared shadow4 = 09 (DIFF)
    Thanks in advance for your assistance/guidance on this matter!

    Hello Alex,
    Thank you for your reply.  I will go through your questions in order as best I can, though I'm not the backup expert here.
    1) "Did you see any event ID when the VSS fail?
    please offer us more information about your environment, such as what type backup you are using the soft ware based or hard ware VSS device."
    I saw a number of events on inspection.  Interestingly enough, the event ID 60 issues did not occur on the drive where shadow copies did remain after the two reboots.  I'm putting my event notes in a code block to try to preserve formatting/readability.
     I've written down events from both server 1 and 2 in this code block, documenting the first reboot causing the role to move to server 2 and then the second reboot going back to server 1:
    JANUARY 2
    9:34:20 PM - Server 1 - Event ID: 1074 - INFO - Source: User 32 - Standard reboot request from explorer.exe (Initiated by me)
    9:34:21 PM - Server 1 - Event ID: 7036 - INFO - Source: Service Control Manager - "The Volume Shadow Copy service entered the running state."
    9:34:21 PM - Server 1 - Event ID: 60 - ERROR - Source: volsnap - "The description for Event ID 60 from source volsnap cannot be found. Either the component that raises this event is not installed on your local computer or the installation is corrupted. You can install or repair the component on the local computer.
    If the event originated on another computer, the display information had to be saved with the event.
    The following information was included with the event:
    \Device\HarddiskVolumeShadowCopy49
    F:
    T:
    The locale specific resource for the desired message is not present"
    9:34:21 PM - Server 1 - Event ID 60 - ERROR - Source: volsnap - "The description for Event ID 60 from source volsnap cannot be found. Either the component that raises this event is not installed on your local computer or the installation is corrupted. You can install or repair the component on the local computer.
    If the event originated on another computer, the display information had to be saved with the event.
    The following information was included with the event:
    \Device\HarddiskVolumeShadowCopy1
    H:
    V:
    The locale specific resource for the desired message is not present"
    ***The above event repeats with only the number changing, drive letters stay same, citing VolumeShadowCopy# numbers 6, 13, 18, 22, 27, 32, 38, 41, 45, 51,
    9:34:21 PM - Server 1 - Event ID: 60 - ERROR - Source: volsnap - "The description for Event ID 60 from source volsnap cannot be found. Either the component that raises this event is not installed on your local computer or the installation is corrupted. You can install or repair the component on the local computer.
    If the event originated on another computer, the display information had to be saved with the event.
    The following information was included with the event:
    \Device\HarddiskVolumeShadowCopy4
    E:
    S:
    The locale specific resource for the desired message is not present"
    ***The above event repeats with only the number changing, drive letters stay same, citing VolumeShadowCopy# numbers 5, 10, 19, 21, 25, 29, 37, 40, 46, 48, 48
    9:34:28 PM - Server 1 - Event ID: 7036 - INFO - Source: Service Control Manager - "The NetBackup Legacy Network Service service entered the stopped state."
    9:34:28 PM - Server 1 - Event ID: 7036 - INFO - Source: Service Control Manager - "The Volume Shadow Copy service entered the stopped state.""
    9:34:29 PM - Server 1 - Event ID: 7036 - INFO - Source: Service Control Manager - "The NetBackup Client Service service entered the stopped state."
    9:34:30 PM - Server 1 - Event ID: 7036 - INFO - Source: Service Control Manager - "The NetBackup Discovery Framework service entered the stopped state."
    10:44:07 PM - Server 2 - Event ID: 7036 - INFO - Source: Service Control Manager - "The Volume Shadow Copy service entered the running state."
    10:44:08 PM - Server 2 - Event ID: 7036 - INFO - Source: Service Control Manager - "The Microsoft Software Shadow Copy Provider service entered the running state."
    10:45:01 PM - Server 2 - Event ID: 48 - ERROR - Source: bxois - "Target failed to respond in time to a NOP request."
    10:45:01 PM - Server 2 - Event ID: 20 - ERROR - Source: bxois - "Connection to the target was lost. The initiator will attempt to retry the connection."
    10:45:01 PM - Server 2 - Event ID: 153 - WARN - Source: disk - "The IO operation at logical block address 0x146d2c580 for Disk 7 was retried."
    10:45:03 PM - Server 2 - Event ID: 34 - INFO - Source: bxois - "A connection to the target was lost, but Initiator successfully reconnected to the target. Dump data contains the target name."
    JANUARY 3
    At around 2:30 I reboot Server 2, seeing that shadow copy was missing after previous failure. Here are the relevant events from the flip back to server 1.
    2:30:34 PM - Server 2 - Event ID: 60 - ERROR - Source: volsnap - "The description for Event ID 60 from source volsnap cannot be found. Either the component that raises this event is not installed on your local computer or the installation is corrupted. You can install or repair the component on the local computer.
    If the event originated on another computer, the display information had to be saved with the event.
    The following information was included with the event:
    \Device\HarddiskVolumeShadowCopy24
    F:
    T:
    The locale specific resource for the desired message is not present"
    2:30:34 PM - Server 2 - Event ID: 60 - ERROR - Source: volsnap - "The description for Event ID 60 from source volsnap cannot be found. Either the component that raises this event is not installed on your local computer or the installation is corrupted. You can install or repair the component on the local computer.
    If the event originated on another computer, the display information had to be saved with the event.
    The following information was included with the event:
    \Device\HarddiskVolumeShadowCopy23
    E:
    S:
    The locale specific resource for the desired message is not present"
    We are using Symantec NetBackup.  The client agent is installed on both server1 and 2.  We're backing them up based on the complete drive letter for each storage volume (this makes recovery easier).  I believe this is what you would call "software
    based VSS".  We don't have the infrastructure/setup to do hardware based snapshots.  The drives reside on a compellent san mapped to the cluster via iScsi.
    2) "Confirm the following registry is exist:
    - HKEY_LOCAL_MACHINE\System\CurrentControlSet\Services\VSS\Settings"
    The key is there, however the DWORD value is not, would that mean that the
    default value is being used at this point?

  • DMS Server with File server

    Hi,
    I have installed DMS, Content server and using KPRO.
    I need that when I check in my files, It may go to file server instead of Content server or R3 database.
    What all configurations need to made? Can you please guie me step by step, or if you have any document kindly inform me.
    Regards,
    Manish Sharma

    Hi Christoph,
    Thanks for replying, But I am really not able to understand.
    When I try "check in" documents in vault, it works perfectly fine.
    When I try to "Check In" Docs in R3 database, it wors perfectly fine.
    But when I check kpro, I get errors while creating repository  in CSADMIN, i get Microsoft ODBC Drivers error. Moreover where I have to define path of the fileserver where my files will be stored. Can you please provide me some procedure where I can get step by step information for configuration.
    Kindly help me. Witing for your reply.
    Regards,
    Manish Sharma

  • Hyper-V 2012 High Availability using Windows Server 2012 File Server Storage

    Hi Guys,
    Need your expertise regarding hyper-v high availability. We setup 2 hyper-v 2012 host in our infra for our domain consolidation project. Unfortunately, we don't have hardware storage that is said to be a requirement that we can use to create a failover cluster
    for hyper-v host to implement HA. Here's the setup:
    Host1
    HP Proliant L380 G7
    Windows Server 2012 Std
    Hyper-V role, Failover Cluster Manager and File and Storage Services installed
    Host2
    Dell PowerEdge 2950
    Windows Server 2012 Std
    Hyper-V role, Failover Cluster Manager and File and Storage Services installed
    Storage
    Dell PowerEdge 6800
    Windows Server 2012 Std
    File and Storage Services installed
    I'm able to configure the new feature about Shared Nothing Live Migration - i'm able to move VM's back and fort between my hosts without storage. But this is planned and proactive approach. My concern is to have my hyper-v host to become highly available in
    the event of system failure. If my host1 dies, the VMs should go/move to host2 and vice versa. In setting this up, i believe i need to enable the failover clustering between my hyper-v hosts which i already did but upon validation, it says "No disks
    were found on which to perform cluster validation tests." Is it possible to cluster it using just a regular windows file server? I've read about SMB 3.0 and i've configured it as well i'm able to save VMs on my file server, but i don't think that my hyper-v
    hosts are already highly available.
    Any feedback and suggestions or recommendation is highly appreciated. Thanks in advance!

    Hi Guys,
    Need your expertise regarding hyper-v high availability. We setup 2 hyper-v 2012 host in our infra for our domain consolidation project. Unfortunately, we don't have hardware storage that is said to be a requirement that we can use to create a failover cluster
    for hyper-v host to implement HA. Here's the setup:
    Host1
    HP Proliant L380 G7
    Windows Server 2012 Std
    Hyper-V role, Failover Cluster Manager and File and Storage Services installed
    Host2
    Dell PowerEdge 2950
    Windows Server 2012 Std
    Hyper-V role, Failover Cluster Manager and File and Storage Services installed
    Storage
    Dell PowerEdge 6800
    Windows Server 2012 Std
    File and Storage Services installed
    I'm able to configure the new feature about Shared Nothing Live Migration - i'm able to move VM's back and fort between my hosts without storage. But this is planned and proactive approach. My concern is to have my hyper-v host to become highly available in
    the event of system failure. If my host1 dies, the VMs should go/move to host2 and vice versa. In setting this up, i believe i need to enable the failover clustering between my hyper-v hosts which i already did but upon validation, it says "No disks were
    found on which to perform cluster validation tests." Is it possible to cluster it using just a regular windows file server? I've read about SMB 3.0 and i've configured it as well i'm able to save VMs on my file server, but i don't think that my hyper-v hosts
    are already highly available.
    Any feedback and suggestions or recommendation is highly appreciated. Thanks in advance!
    Your shared storage is a single point of failure with this scenario so I would not consider the whole setup as a production configuration... Also setup is both slow (all I/O is travelling down the wire to storage server, running VMs from DAS is ages faster)
    and expensive (third server + extra Windows license). I would think twice about what you do and either deploy a built-in VM replication technologies (Hyper-V Replica) and apps built-in clustering features that does not require shared storage (SQL Server and
    Database Mirroring for example, BTW what workload do you run?) or use some third-party software creating fault tolerant shared storage from DAS or investing into physical shared storage hardware (HA one of course). 
    Hi VR38DETT,
    Thanks for responding. The hosts will cater a domain controller (on each host), Web filtering software (Websense), Anti-Virus (McAfee ePO), WSUS and an Auditserver as of the moment. Is the Hyper-V Replica somewhat give "high availability" to VMs or Hyper-V
    hosts? Also, is the cluster required in order to implement it? Haven't tried that but worth a try.

Maybe you are looking for

  • Policies for URL's with question mark (?) in Access Manager 7.1.2

    I have an environment with Access Manager 7.1 patch 2 and OpenSSO 3.0 Web Policy Agent (Apache). Apparently, with the 3.0 agents the wildcard * no longer matches the query string separator (?) in a URL accordining to Appendix C in the manual (http://

  • How to implement sheduling services in Ep

    Hi all, I would like to know how to implement scheduling services in ep in the sense .. running jobs in portal . Any pointers to this would be of great help. Thanks& Regards, Uma,

  • Version mgmt in PR

    I am using 3 level rel strategy for PR; I want to develop a report for the release dates along with its codes. Does version management stores the rel dates for a PR. VS

  • Glance Screen for Lumia 625

    Hello. I just bought this monster phone, the Lumia 625. I'm kinda wondering why it does not have the glance screen on it? Is it available for installation?

  • Buying a Power Mac G5 DP 1.8GHz Dual with a PPC 7450 Processor???

    I just purchased but have not received as of yet a Power Mac G5 DP 1.8GHz Dual PCI-X 1 GB 160 GB. In the Details it is stated: "1.8 GHz DP PPC 7450 (G5) x2 RISC Processor." A friend of mine mentioned that the PPC 7450 is a G4 processor but it might h