Correct idea to scale out testing environment and test service pack2 installation

Hi
I have a sharepoint 2010 farm it has one sharepoint server, one database server
In one server  below services are running
Central administration service
SharePoint Server Search 
User Profile Service 
Microsoft SharePoint Foundation Web Application
so i want to scale out this form  to
1 application server
1 web front end server
1 Search server (index server)
1 databse server
here how i scale out to this form
1)here how i move  sharepoint  server search service to new  index server and
2) here how i move  Microsoft SharePoint Foundation Web Application to new webfront end server
and in this single server  some web appllications are running also how i move these to new wf server
i want to do like this  because i want to test service pack 2 installation, now  sharepoint version is : service pack1
my actual production environment has
2 application servers
2 webfront end servers
2 index servers
1 databae server
so this correct idea to scale out testing environment and test service pack2 installation
adil

Hi Adil,
The link below describes how to scale SharePoint Web Front-End with only web applications and the search query server  out of one SharePoint server with all roles running.
http://sharepointsolutions.com/sharepoint-help/blog/2011/02/how-to-scale-out-a-sharepoint-2010-farm-from-two-tier-to-three-tier-by-adding-a-dedicated-application-server/
Now you have two SharePoint server with:
Tier 1 – SharePoint Server dedicated as a Web Front-End (WFE) with only the web application(s) and the search query service running on it
Tier 2 – SharePoint Server dedicated as an Application Server with all of the other service applications running on it, but no web applications or query service
Tier 3 – SQL Server for the databases
Then you would scale out WFE server with web applications from tier1. Now please install a new SharePoint server and join it to the existing farm and deploy it as Web Front Server. Enable the relevant services on Web Front servers per the topology picture
below, and stop the services running on the old server.
http://technet.microsoft.com/en-us/library/cc263044(v=office.14).aspx
Regards
Rebecca Tu
TechNet Community Support

Similar Messages

  • How to find out the IPS and DS Service Pack and patch levels in windows server.

    Hi,
    I have installed the Information Platform Services 4.0 and Data services 4.0 in windows QA system, now we need to install the same version of IPS & DS  in windows PRD system for this how could we find out the exact version of IPS & DS  ( Service pack and patch levels ) and where we can find the same.
    Many Thanks,
    Sudheer

    Hi Sudheer,
    On a Windows Operating Systems you can check the “Programs and Features” in control panel.You can see a list of both the Name and Version number of SAP BusinessObjects that is installed.
    option 2:
    Open CMC go to Settings. In the properties, you will find the Product Version.
    Please check below SAP notes for more information.
    http://service.sap.com/sap/support/notes/1602088
    Regards,
    Pavan

  • Problems with G6 - PT API and Content Service Upload Installation

    Hi all,
    What port number is PT API using in G6? Examples in Installation Guide are using 11905for PT API, is that the right port or just a dummy port? Same question for Content Service Upload, is port 11910dummy?
    I installed all Plumtree Foundation Components in one box as a Test. Followed every single instruction in the installation guide and the installation went successfully. Installation logs have no errors at all. Run diagnostic.bat test and it went successfully. All Plumtree services started successfully.
    But the installation didn't create a "ptapi" virtual directory in IIS, hence the verifying installation (following instruction in installation guide page 4-25) step failed. PTSpy captured an error message (please see my reply coz the collab message can't exceed 16K chars).
    Verifying Content Upload Service also failed (Tried to upload a simple document to a KD folder, error occured and upload was unsuccessful). PTSpy captured some error messages (please see my reply coz the collab message can't exceed 16K chars).
    Can someone give me a hint ? Has anyone experienced these problems too???
    Regards,
    JS

    Content Service Upload Error
    406 portal.saptg6.Administrator 2-9-2006 16:12:10.296 Error Portal_Browsing .NET #37 com.plumtree.portalpages.browsing.directory.DirRepostControl DirRepost::PerformAction The following error occured when trying to submit a card: Error in function PTDataSource.ImportDocument (vDocumentLocationBagAsXML == '<?xml version="1.0" encoding="ucs-2"?><PTBAG V="1.1" xml:space="preserve"><I N="PTC_PBAGFORMAT">2002</I><S N="PTC_UNIQUE">http://www.copyright.gov/legislation/dmca.pdf</S><I N="PTC_DTM_SECT">1001</I><S N="PTC_DOC_ID">application/pdf</S></PTBAG>', lDocumentTypeID == 103, pCard == com.plumtree.server.impl.directory.PTCard, bSummarize == true, pProvider == ): Error in function PTDataSource.ImportDocument (vDocumentLocationBagAsXML == <?xml version="1.0" encoding="ucs-2"?><PTBAG V="1.1" xml:space="preserve"><I N="PTC_PBAGFORMAT">2002</I><S N="PTC_UNIQUE">http://www.copyright.gov/legislation/dmca.pdf</S><I N="PTC_DTM_SECT">1001</I><S N="PTC_DOC_ID">application/pdf</S></PTBAG>, lDocumentTypeID == 103, pCard == com.plumtree.server.impl.directory.PTCard, bSummarize == true, pProvider == ): Error in function PTFileAccessorBase.AttachToDocument (pBagDocumentLocation == <?xml version="1.0" encoding="ucs-2"?><PTBAG V="1.1" xml:space="preserve"><I N="PTC_PBAGFORMAT">4000</I><S N="PTC_FILE">C:\Program Files\plumtree\ptportal\6.0\tmp\tmp\{73B926FC-52D7-4675-B0EA-BBFC610A8253}~~~1139465529921.pdf</S><S N="PTC_WEB_URL">http://www.copyright.gov/legislation/dmca.pdf</S><S N="PTC_WEB_CT">Windows-1252</S></PTBAG>, pSession == com.plumtree.server.impl.core.PTSession): Year must be a positive numbercom.plumtree.openfoundation.util.XPIllegalArgumentException: Error in function PTDataSource.ImportDocument (vDocumentLocationBagAsXML == '<?xml version="1.0" encoding="ucs-2"?><PTBAG V="1.1" xml:space="preserve"><I N="PTC_PBAGFORMAT">2002</I><S N="PTC_UNIQUE">http://www.copyright.gov/legislation/dmca.pdf</S><I N="PTC_DTM_SECT">1001</I><S N="PTC_DOC_ID">application/pdf</S></PTBAG>', lDocumentTypeID == 103, pCard == com.plumtree.server.impl.directory.PTCard, bSummarize == true, pProvider == ): Error in function PTDataSource.ImportDocument (vDocumentLocationBagAsXML == <?xml version="1.0" encoding="ucs-2"?><PTBAG V="1.1" xml:space="preserve"><I N="PTC_PBAGFORMAT">2002</I><S N="PTC_UNIQUE">http://www.copyright.gov/legislation/dmca.pdf</S><I N="PTC_DTM_SECT">1001</I><S N="PTC_DOC_ID">application/pdf</S></PTBAG>, lDocumentTypeID == 103, pCard == com.plumtree.server.impl.directory.PTCard, bSummarize == true, pProvider == ): Error in function PTFileAccessorBase.AttachToDocument (pBagDocumentLocation == <?xml version="1.0" encoding="ucs-2"?><PTBAG V="1.1" xml:space="preserve"><I N="PTC_PBAGFORMAT">4000</I><S N="PTC_FILE">C:\Program Files\plumtree\ptportal\6.0\tmp\tmp\{73B926FC-52D7-4675-B0EA-BBFC610A8253}~~~1139465529921.pdf</S><S N="PTC_WEB_URL">http://www.copyright.gov/legislation/dmca.pdf</S><S N="PTC_WEB_CT">Windows-1252</S></PTBAG>, pSession == com.plumtree.server.impl.core.PTSession): Year must be a positive numbercom.plumtree.openfoundation.util.XPIllegalArgumentException: Error in function PTDataSource.ImportDocument (vDocumentLocationBagAsXML == '<?xml version="1.0" encoding="ucs-2"?><PTBAG V="1.1" xml:space="preserve"><I N="PTC_PBAGFORMAT">2002</I><S N="PTC_UNIQUE">http://www.copyright.gov/legislation/dmca.pdf</S><I N="PTC_DTM_SECT">1001</I><S N="PTC_DOC_ID">application/pdf</S></PTBAG>', lDocumentTypeID == 103, pCard == com.plumtree.server.impl.directory.PTCard, bSummarize == true, pProvider == ): Error in function PTDataSource.ImportDocument (vDocumentLocationBagAsXML == <?xml version="1.0" encoding="ucs-2"?><PTBAG V="1.1" xml:space="preserve"><I N="PTC_PBAGFORMAT">2002</I><S N="PTC_UNIQUE">http://www.copyright.gov/legislation/dmca.pdf</S><I N="PTC_DTM_SECT">1001</I><S N="PTC_DOC_ID">application/pdf</S></PTBAG>, lDocumentTypeID == 103, pCard == com.plumtree.server.impl.directory.PTCard, bSummarize == true, pProvider == ): Error in function PTFileAccessorBase.AttachToDocument (pBagDocumentLocation == <?xml version="1.0" encoding="ucs-2"?><PTBAG V="1.1" xml:space="preserve"><I N="PTC_PBAGFORMAT">4000</I><S N="PTC_FILE">C:\Program Files\plumtree\ptportal\6.0\tmp\tmp\{73B926FC-52D7-4675-B0EA-BBFC610A8253}~~~1139465529921.pdf</S><S N="PTC_WEB_URL">http://www.copyright.gov/legislation/dmca.pdf</S><S N="PTC_WEB_CT">Windows-1252</S></PTBAG>, pSession == com.plumtree.server.impl.core.PTSession): Year must be a positive number at com.plumtree.openfoundation.util.XPException.GetInstance(String strErrorMsg, Exception e) at com.plumtree.server.impl.directory.PTDataSource.ImportDocument(String vDocumentLocationBagAsXML, Int32 lDocumentTypeID, IPTCard pCard, Boolean bSummarize, IPTDataSourceProvider pProvider) in e:\buildroot\Release\portalserver\6.0.x\portalobjects\build\x86\src\dotnet\com\plumtree\server\impl\directory\PTDataSource.cs:line 465 at com.plumtree.portalpages.browsing.directory.DirModel.SubmitCardWithPropertyBag(Int32[] arFolderIDs, String strLanguage, Int32 nDocTypeID, Int32 nDataSourceID, IXPPropertyBag pbagCardProp, String strCardName, String strCardDescription) in e:\buildroot\Release\portalui\6.0.x\ptwebui\portalpages\dotnet\prod\src\com\plumtree\portalpages\browsing\directory\DirModel.cs:line 3290 at com.plumtree.portalpages.browsing.directory.DirModel.SubmitCardWithPropertyBag(Int32 nDataSourceID, String strCardName, String strCardDescription, IXPPropertyBag pbagCardProp) in e:\buildroot\Release\portalui\6.0.x\ptwebui\portalpages\dotnet\prod\src\com\plumtree\portalpages\browsing\directory\DirModel.cs:line 3205 at com.plumtree.portalpages.browsing.directory.DirModel.SubmitCard() in e:\buildroot\Release\portalui\6.0.x\ptwebui\portalpages\dotnet\prod\src\com\plumtree\portalpages\browsing\directory\DirModel.cs:line 3051 at com.plumtree.portalpages.browsing.directory.DirRepostControl.PerformAction(Int32 nAction) in e:\buildroot\Release\portalui\6.0.x\ptwebui\portalpages\dotnet\prod\src\com\plumtree\portalpages\browsing\directory\DirRepostControl.cs:line 626   at com.plumtree.uiinfrastructure.form.RepostControl.CheckActionSecurityAndExecute(XPHashtable arguments) in e:\buildroot\Release\uiinfrastructure\6.0.x\dotnet\prod\src\com\plumtree\uiinfrastructure\form\RepostControl.cs:line 184   at com.plumtree.portalpages.browsing.directory.DirRepostControl.CheckActionSecurityAndExecute(XPHashtable arguments) in e:\buildroot\Release\portalui\6.0.x\ptwebui\portalpages\dotnet\prod\src\com\plumtree\portalpages\browsing\directory\DirRepostControl.cs:line 312   at com.plumtree.uiinfrastructure.interpreter.MVCController.HandleRedirect(Redirect redirect, RequestData tempData) in e:\buildroot\Release\uiinfrastructure\6.0.x\dotnet\prod\src\com\plumtree\uiinfrastructure\interpreter\MVCController.cs:line 385   at com.plumtree.uiinfrastructure.interpreter.Interpreter.HandleRequest(IXPRequest request, IXPResponse response, ISessionManager session, IApplication application) in e:\buildroot\Release\uiinfrastructure\6.0.x\dotnet\prod\src\com\plumtree\uiinfrastructure\interpreter\Interpreter.cs:line 387   at com.plumtree.uiinfrastructure.interpreter.Interpreter.DoService(IXPRequest request, IXPResponse response, ISessionManager session, IApplication application) in e:\buildroot\Release\uiinfrastructure\6.0.x\dotnet\prod\src\com\plumtree\uiinfrastructure\interpreter\Interpreter.cs:line 167   at com.plumtree.uiinfrastructure.web.XPPage.Service(HttpRequest httpRequest, HttpResponse httpResponse, HttpSessionState httpSession, HttpApplicationState httpApplication) in e:\buildroot\Release\httpmemorymanagement\2.0.x\dotNET\src\com\plumtree\uiinfrastructure\web\XPPage.cs:line 82   at com.plumtree.portaluiinfrastructure.activityspace.PlumHandler.ProcessRequest(HttpContext context) in e:\buildroot\Release\portalui\6.0.x\ptwebui\portal\dotnet\prod\src\web\PlumHandler.cs:line 37   at System.Web.CallHandlerExecutionStep.System.Web.HttpApplication+IExecutionStep.Execute()   at System.Web.HttpApplication.ExecuteStep(IExecutionStep step, Boolean& completedSynchronously)   at System.Web.HttpApplication.ResumeSteps(Exception error)   at System.Web.HttpApplication.System.Web.IHttpAsyncHandler.BeginProcessRequest(HttpContext context, AsyncCallback cb, Object extraData)   at System.Web.HttpRuntime.ProcessRequestInternal(HttpWorkerRequest wr)   at System.Web.HttpRuntime.ProcessRequest(HttpWorkerRequest wr)   at System.Web.Hosting.ISAPIRuntime.ProcessRequest(IntPtr ecb, Int32 iWRType)Caused by: com.plumtree.openfoundation.util.XPIllegalArgumentException: Error in function PTDataSource.ImportDocument (vDocumentLocationBagAsXML == <?xml version="1.0" encoding="ucs-2"?><PTBAG V="1.1" xml:space="preserve"><I N="PTCPBAGFORMAT">2002</I><S N="PTC_UNIQUE">http://www.copyright.gov/legislation/dmca.pdf</S><I N="PTC_DTM_SECT">1001</I><S N="PTC_DOC_ID">application/pdf</S></PTBAG>, lDocumentTypeID == 103, pCard == com.plumtree.server.impl.directory.PTCard, bSummarize == true, pProvider == ): Error in function PTFileAccessorBase.AttachToDocument (pBagDocumentLocation == <?xml version="1.0" encoding="ucs-2"?><PTBAG V="1.1" xml:space="preserve"><I N="PTC_PBAGFORMAT">4000</I><S N="PTC_FILE">C:\Program Files\plumtree\ptportal\6.0\tmp\tmp\{73B926FC-52D7-4675-B0EA-BBFC610A8253}~~~1139465529921.pdf</S><S N="PTC_WEB_URL">http://www.copyright.gov/legislation/dmca.pdf</S><S N="PTC_WEB_CT">Windows-1252</S></PTBAG>, pSession == com.plumtree.server.impl.core.PTSession): Year must be a positive number at com.plumtree.openfoundation.util.XPException.GetInstance(String strErrorMsg, Exception e) at com.plumtree.server.impl.directory.PTDataSource.ImportDocument(IXPPropertyBag vDocumentLocationBagAsXML, Int32 lDocumentTypeID, IPTCard pCard, Boolean bSummarize, IPTDataSourceProvider pProvider) in e:\buildroot\Release\portalserver\6.0.x\portalobjects\build\x86\src\dotnet\com\plumtree\server\impl\directory\PTDataSource.cs:line 446 at com.plumtree.server.impl.directory.PTDataSource.ImportDocument(String vDocumentLocationBagAsXML, Int32 lDocumentTypeID, IPTCard pCard, Boolean bSummarize, IPTDataSourceProvider pProvider) in e:\buildroot\Release\portalserver\6.0.x\portalobjects\build\x86\src\dotnet\com\plumtree\server\impl\directory\PTDataSource.cs:line 456 ... 18 moreCaused by: com.plumtree.openfoundation.util.XPIllegalArgumentException: Error in function PTFileAccessorBase.AttachToDocument (pBagDocumentLocation == <?xml version="1.0" encoding="ucs-2"?><PTBAG V="1.1" xml:space="preserve"><I N="PTC_PBAGFORMAT">4000</I><S N="PTC_FILE">C:\Program Files\plumtree\ptportal\6.0\tmp\tmp\{73B926FC-52D7-4675-B0EA-BBFC610A8253}~~~1139465529921.pdf</S><S N="PTC_WEB_URL">http://www.copyright.gov/legislation/dmca.pdf</S><S N="PTC_WEB_CT">Windows-1252</S></PTBAG>, pSession == com.plumtree.server.impl.core.PTSession): Year must be a positive number at com.plumtree.openfoundation.util.XPException.GetInstance(String strErrorMsg, Exception e) at com.plumtree.server.impl.directory.accessors.PTFileAccessorBase.AttachToDocument(IXPPropertyBag pBagDocumentLocation, IPTSession pSession) in e:\buildroot\Release\portalserver\6.0.x\portalobjects\build\x86\src\dotnet\com\plumtree\server\impl\directory\accessors\PTFileAccessorBase.cs:line 130 at com.plumtree.server.impl.directory.providers.PTWebProvider.AttachToDocument(IXPPropertyBag spIPBagDocumentLocation, String AccessorID) in e:\buildroot\Release\portalserver\6.0.x\portalobjects\build\x86\src\dotnet\com\plumtree\server\impl\directory\providers\PTWebProvider.cs:line 183 at com.plumtree.server.impl.directory.PTDataSource.InternalImportDocument(IXPPropertyBag vDocumentLocationBagAsXML, Int32 lDocumentTypeID, IPTCard pCard, Boolean bSummarize, IPTDataSourceProvider pProvider, IPTDocumentType spDocumentType) in e:\buildroot\Release\portalserver\6.0.x\portalobjects\build\x86\src\dotnet\com\plumtree\server\impl\directory\PTDataSource.cs:line 1494 at com.plumtree.server.impl.directory.PTDataSource.ImportDocument(IXPPropertyBag vDocumentLocationBagAsXML, Int32 lDocumentTypeID, IPTCard pCard, Boolean bSummarize, IPTDataSourceProvider pProvider) in e:\buildroot\Release\portalserver\6.0.x\portalobjects\build\x86\src\dotnet\com\plumtree\server\impl\directory\PTDataSource.cs:line 436 ... 19 moreCaused by: com.plumtree.openfoundation.util.XPIllegalArgumentException: Year must be a positive number at com.plumtree.openfoundation.util.XPCalendar.ValidateYearRange(Int32 year) at com.plumtree.openfoundation.util.XPCalendar..ctor(Int32 year, Int32 month, Int32 day, Int32 hour, Int32 minute, Int32 second) at com.plumtree.openfoundation.util.XPCalendar.GetInstance(Int32 year, Int32 month, Int32 day, Int32 hour, Int32 minute, Int32 second) at com.plumtree.server.impl.directory.accessors.PTPDFAccessor.convertStringToDate(String val) in e:\buildroot\Release\portalserver\6.0.x\portalobjects\build\x86\src\dotnet\com\plumtree\server\impl\directory\accessors\PTPDFAccessor.cs:line 643 at com.plumtree.server.impl.directory.accessors.PTPDFAccessor.assignPropertyValue(Int32 iRow, Object key, OpenSearchFileProperty prop) in e:\buildroot\Release\portalserver\6.0.x\portalobjects\build\x86\src\dotnet\com\plumtree\server\impl\directory\accessors\PTPDFAccessor.cs:line 363 at com.plumtree.server.impl.directory.accessors.PTPDFAccessor.ReadPDFFieldsFromDoc() in e:\buildroot\Release\portalserver\6.0.x\portalobjects\build\x86\src\dotnet\com\plumtree\server\impl\directory\accessors\PTPDFAccessor.cs:line 292 at com.plumtree.server.impl.directory.accessors.PTPDFAccessor.CheckDocumentValidity() in e:\buildroot\Release\portalserver\6.0.x\portalobjects\build\x86\src\dotnet\com\plumtree\server\impl\directory\accessors\PTPDFAccessor.cs:line 40 at com.plumtree.server.impl.directory.accessors.PTFileAccessorBase.AttachToDocument(IXPPropertyBag pBagDocumentLocation, IPTSession pSession) in e:\buildroot\Release\portalserver\6.0.x\portalobjects\build\x86\src\dotnet\com\plumtree\server\impl\directory\accessors\PTFileAccessorBase.cs:line 113 ... 22 more

  • How to backup a Scale Out File Server

    Can anyone explain to me how i am supposed to backup my Scale Out File Server? It is used to host our VDI environment but we have thrown a few other minimal shares on there for stuff relevant to the VDI setup. Profile Disk/Scripts etc
    Per this topic http://social.technet.microsoft.com/Forums/en-US/15561919-963b-4164-a307-e3696f5d21e1/adding-scaleout-file-servers-to-dpm2012-sp1?forum=dpmfilebackup
    "DPM does not support protecting scale-out file servers (SOFS) for file shares.
    I do not really know how to proceed. To make sure everything is backed up.

    Hi,
    Scale out file server support in DPM is for VM's (and SQL DB's I think), only and to be honest because of the metadata overhead using scale out file server and CSV as a general file server is not a good choice.  To protect VM's running on a scale out
    file server the VM's are backed up from the Hyper-V hosts but the Scale Out file servers do require a DPM agent on them.  SMB 3.0 has a new feature which enables Remote VSS therefore the volumes (shares) are snapped from the Hyper-V hosts.
    This document will give you the basic steps (http://technet.microsoft.com/en-us/library/hh757866.aspx)
    hope this helps,
    Regards,
    Paul

  • How to size a Scale-out File Server

    Hi,
    We are looking to implement a 2-node 2012 R2 Scale-out File Server cluster (using SAS JBOD enclosure) for the primary purpose of storing the VHD files that will be accessed by a 4-node 2012 R2 Hyper-V cluster using 10 gigabit Ethernet (no RDMA).  Our
    environment can be characterised as having a large number of mostly idle VMs that experience sporadic, low intensity use (this is *not* a VDI environment).  We have 2 questions.
    1) To what extent is RAM a consideration for the SoFS servers?  We can't find any documentation to suggest that there are benefits to be gained by having more RAM in the SoFS servers but we don't know if we should go with 8/16/32/64+ GB RAM in each
    of the nodes.
    2) With the need to keep costs down, we don't think RDMA / SMB-Direct NICs are going to be within our reach.  Should we however look to have 2 * dual-port 10 Gbps NICs in both the SoFS & Hyper-V boxes?

    Unless your VMs are read-intensive and you're going to deploy a CSV cache memory requirement for serving mostly idle VMs can be pretty low. However RAM is cheap these days so going for less then 16GB per node does not sound reasonable. Good sample, see:
    Windows Server 2012 File Server Tip: Enable CSV Caching on Scale-Out File Server Clusters
    http://blogs.technet.com/b/josebda/archive/2012/11/14/windows-server-2012-file-server-tip-enable-csv-caching-on-scale-out-file-server-clusters.aspx
    How to Enable
    CSV Cache
    http://blogs.msdn.com/b/clustering/archive/2013/07/19/10286676.aspx
    HYPER-V OVER SMB: SCALE-OUT FILE SERVER AND STORAGE SPACES
    http://www.thomasmaurer.ch/2013/08/hyper-v-over-smb-scale-out-file-server-and-storage-spaces/
    Hope this helped :)
    StarWind VSAN [Virtual SAN] clusters Hyper-V without SAS, Fibre Channel, SMB 3.0 or iSCSI, uses Ethernet to mirror internally mounted SATA disks between hosts.

  • Scale out SSAS server for better performance

    HI
    i have a sharepoint farm
    running performance point service in a server where ANaylysis srver,reporting server installed
    and we have anyalysis server dbs and cubes
    and a wfe server where secure store service running
    we have
    1) application server + domain controller
    2) two wfes
    1) sql server sharepoint
    1) SSAS server ( analysis server dbs+ reporting server)
    here how i scaled out my SSAS server for better performance 
    adil

    Just trying to get a definitive answer to the question of can we use a Shared VHDX in a SOFS Cluster which will be used to store VHDX files?
    We have a 2012 R2 RDS Solution and store the User Profile Disks (UPD) on a SOFS Cluster that uses "traditional" storage from a SAN. We are planning on creating a new SOFS Cluster and wondered if we can use a shared VHDX instead of CSV as the storage that
    will then be used to store the UPDs (one VHDX file per user).
    Cheers for now
    Russell
    Sure you can do it. See:
    Deploy a Guest Cluster Using a Shared Virtual Hard Disk
    http://technet.microsoft.com/en-us/library/dn265980.aspx
    Scenario 2: Hyper-V failover cluster using file-based storage in a separate Scale-Out File Server
    This scenario uses Server Message Block (SMB) file-based storage as the location of the shared .vhdx files. You must deploy a Scale-Out File Server and create an SMB file share as the storage location. You also need a separate Hyper-V failover cluster.
    The following table describes the physical host prerequisites.
    Cluster Type
    Requirements
    Scale-Out File Server
    At least two servers that are running Windows Server 2012 R2.
    The servers must be members of the same Active Directory domain.
    The servers must meet the requirements for failover clustering.
    For more information, see Failover Clustering Hardware Requirements and Storage Options and Validate
    Hardware for a Failover Cluster.
    The servers must have access to block-level storage, which you can add as shared storage to the physical cluster. This storage can be iSCSI, Fibre Channel, SAS, or clustered storage spaces that use a set of shared SAS JBOD enclosures.
    StarWind VSAN [Virtual SAN] clusters Hyper-V without SAS, Fibre Channel, SMB 3.0 or iSCSI, uses Ethernet to mirror internally mounted SATA disks between hosts.

  • Scale Out File Server for Applications using Shared VHDX

    Just trying to get a definitive answer to the question of can we use a Shared VHDX in a SOFS Cluster which will be used to store VHDX files?
    We have a 2012 R2 RDS Solution and store the User Profile Disks (UPD) on a SOFS Cluster that uses "traditional" storage from a SAN. We are planning on creating a new SOFS Cluster and wondered if we can use a shared VHDX instead of CSV as the storage
    that will then be used to store the UPDs (one VHDX file per user).
    Cheers for now
    Russell

    Just trying to get a definitive answer to the question of can we use a Shared VHDX in a SOFS Cluster which will be used to store VHDX files?
    We have a 2012 R2 RDS Solution and store the User Profile Disks (UPD) on a SOFS Cluster that uses "traditional" storage from a SAN. We are planning on creating a new SOFS Cluster and wondered if we can use a shared VHDX instead of CSV as the storage that
    will then be used to store the UPDs (one VHDX file per user).
    Cheers for now
    Russell
    Sure you can do it. See:
    Deploy a Guest Cluster Using a Shared Virtual Hard Disk
    http://technet.microsoft.com/en-us/library/dn265980.aspx
    Scenario 2: Hyper-V failover cluster using file-based storage in a separate Scale-Out File Server
    This scenario uses Server Message Block (SMB) file-based storage as the location of the shared .vhdx files. You must deploy a Scale-Out File Server and create an SMB file share as the storage location. You also need a separate Hyper-V failover cluster.
    The following table describes the physical host prerequisites.
    Cluster Type
    Requirements
    Scale-Out File Server
    At least two servers that are running Windows Server 2012 R2.
    The servers must be members of the same Active Directory domain.
    The servers must meet the requirements for failover clustering.
    For more information, see Failover Clustering Hardware Requirements and Storage Options and Validate
    Hardware for a Failover Cluster.
    The servers must have access to block-level storage, which you can add as shared storage to the physical cluster. This storage can be iSCSI, Fibre Channel, SAS, or clustered storage spaces that use a set of shared SAS JBOD enclosures.
    StarWind VSAN [Virtual SAN] clusters Hyper-V without SAS, Fibre Channel, SMB 3.0 or iSCSI, uses Ethernet to mirror internally mounted SATA disks between hosts.

  • SQL Server Service Pack Install and SSRS Scale Out Environment

    We have an error in our SSRS environment that looks to be addressed in SQL Server 2008 R2 SP3. Our set up is a scale out environment, 2 SSRS app servers and 1 SSRS database server. For simplicity as we are moving to 2012 soon, I would like to only install
    the service pack on one of the SSRS app servers. Anyone aware of any reason besides keeping servers consistent to not do this? I don't want to go through the time to upgrade the SSRS SQL Server if I don't have to.
    Thanks

    Hello,
    If you apply the service pack on the database server first, front-ends may not be able to provide service again until you
    apply the SP to them too. I have no knowledge of other possible issue.
    Remember to backup SSRS databases and encryption key before applying the service pack.
    Hope this helps.
    Regards,
    Alberto Morillo
    SQLCoffee.com

  • When to scale up and/or scale out?

    When do you guys scale up/out, and how do you justify it?
    We have a small farm with a number of new projects we expect to have in production this year. I've been tasked with assessing our current environment, providing recommendations, and performing any scale out Sharepoint configuration.
    I’m reviewing Microsoft’s capacity management and performance testing documentation. I want requests for architecture changes to be as quantifiable as possible, but I'm finding them incredibly frustrating.
    I've checked our current server performance. SQL server's RAM is maxed 90% of the time which, I'm assured by my network guys, is normal for SQL boxes. There are occasional CPU spikes but nothing major.
    I also took a look at the CA Admin Report Library for Crawl Rates, Query Latency, etc, but need something to compare against.
    Any guidance/recommendations greatly appreciated.
    Thanks,
    Scott

    Fist, let's look at your immediate concern.
    High memory use on SQL is a normal sign of a healthy SQL box. i.e. you want SQL to store commonly accessed data in memory so it's not pulling from disk. SQL will try to use as much memory as it can to achieve this goal.
    High memory use on SQL is a normal sign of an unhealthy SQL box when you're seeing lots of page swaps to disk. If the server is constantly caching data from memory to disk this suggests the server needs more memory.
    Microsoft has a good document for troubleshooting performance. Check out Monitoring and maintaining SharePoint Server 2010. 
    The trick with the documentation about performance is every environment is different. What may be unacceptable performance in a farm that serves thousands or users may be completely tolerable and reasonable for one that serves a small organization or department.
    You're not going to find specific guidance from Microsoft along the lines of "if you see CPU at ABC then add a CPU. If you see CPU at XYZ then add a server" because the decision is more nuanced than simply looking at a value.
    There are lots of great guides that others in the community have published that can provide some real-world context to how you can monitor performance for scale.
    Jason Warren
    @jaspnwarren
    jasonwarren.ca
    habaneroconsulting.com/Insights

  • Scale out sharepoint server search to index server and Microsoft SharePoint Foundation Web Application to new server

    Hi
    in my single server these below service are running:
    SharePoint Server Search 
    User Profile Service 
    Microsoft SharePoint Foundation Web Application
    here how i scale out  Microsoft SharePoint Foundation Web Application  to another server
    and  SharePoint Server Search  to new index server.
    adil

    Check here to see how to add servers to the farm:
    http://technet.microsoft.com/en-us/library/cc261752(v=office.15).aspx
    What Inderjeet meant was that if you chose to install SharePoint as a single server instead of a complete farm, you won't be able to add servers to the farm at a later time. You can check if this is the case by checking the registry key he mentioned. If
    that is indeed the case, there's nothing to it but reinstall the farm.
    So that is why you never should choose the single server option, it locks you in and limits future options, while choosing to install sharepoint as a complete farm it's still perfectly acceptable to host it on a single server.
    Kind regards,
    Margriet Bruggeman
    Lois & Clark IT Services
    web site: http://www.loisandclark.eu
    blog: http://www.sharepointdragons.com

  • I have an iMac mid 2010. After Disk utilities said the drive was corrupt, I ran the hardware test (extended) 3 times. No problems. I than ran Disk utility again and got 2 messages in red-keys out of order, and --Volume corrupt, can't be repaired.  I tried

    I have an iMac mid 2010. After Disk utilities said the drive was corrupt, I ran the hardware test (extended) 3 times. No problems. I than ran Disk utility again and got 2 messages in red…keys out of order, and ……Volume corrupt, can't be repaired.
    I tried to do a system install, but message said…disk can't be changed. Will a complete erase and install fix this ?
    Everything is backed-up.

    There's hardware and software.
    Your drive could physically be in good condition but something has caused the file structure to be messed up (it just happens sometimes).  The directory that keep track of files is corrupt.  Disk Utility can help determine if that is the issue.  In some cases it can repair the problem but it cannot do it to a drive that is being used to boot the system while you are doing that.  You have to boot from another drive.  Some things DU cannot repair and you need to try a stronger utility such as Diskwarrior.
    A failing hard drive (one with mechanical issues) can be a cause of file structure issues but this is by no means always the case.  Apart from looking at the SMART diagnosis feature of Disk Utility (and smart isn't totally reliable) there often isn't a way to tell if a drive is failing until it suddenly happens, which is why you need to maintain backups.
    In your situation it does sound serious and not just minor directory corruption.  You can try erasing (formatting) the drive,  You may need to do this if it is really corrupt, in which case there could be underlying physical issues.  I use old technology but if you have the patience you can try erasing the drive with the security option of writing zeros once (don't do more unless you are going on a 3 week vacation).  This makes the drive not simply assume everything is okay by forcing it try every part of the drive by writing data.  If something is gong wrong you may find it marks a lot of bad sectors, or flat out refuses to format.  Then you know you need a new drive.

  • I killed my computer with coffee. It's out of waranty and applecare has run out. It's a 4 year old Macbook Pro. Any idea whether it makes sense to try and fix it?

    I killed my computer with coffee. It's out of waranty and applecare has run out. It's a 4 year old Macbook Pro. Any idea whether it makes sense to try and fix it?

    Nothing works at all. When I spilled the coffee I shut it down, unplugged everything, turned it over and put paper towels beneath it to drain. Made the mistake of turning it back on after a couple of hours. It worked but emitted strange noises from the internal speakers. Turned it back off, but have not been able to turn it back on since. I disassembled the entire macbook yesterday, found coffee stains on the keyboard components. Tried to turn it on and get nothing. The only sign of life is the battery indicator. When I press the button the series of green lights come on, but that's it. Thanks for your responses. Jaylloyd

  • To whom this may concern, I'm having troubles with uploading Flash Files Online throught Adobe Flash Professional and I would to correct the Error Messages I've received as an indication that I need to correct them in order to test a Flash and Shockwave F

    To whom this may concern, I'm having troubles with uploading Flash Files Online throught Adobe Flash Professional and I would to correct the Error Messages I've received as an indication that I need to correct them in order to test a Flash and Shockwave Files I have stored into the Software of Flash!. Could you please help me get this matter resolved? Gerard Hargrove.

    what exactly are you trying to do and attach a screenshot of the error message, if it's in english.

  • DPM 2012 R2, Scale-out and secondary replicas

    Hi all,
    TechNet says scale-out protection for VMs and DPM chaining isn't supported - is this still the case for 2012 R2?
    Our scenario: We have numerous clusters with a ~dozen Hyper-V (2012/2012R2) nodes each, and lots of running VMs. We need to use DPM scale-out to have multiple DPM servers protect the
    clusters, but also need off-site replicas for the data sources.
    Cheers

    Hi,
    There is no change in DPM 2012 R2 with respect to secondary protection for scale out Hyper-v.  Consider backing up to Azure for offsite solution.
    Backing Up DPM using Windows Azure Backup
    http://technet.microsoft.com/en-us/library/jj728752.aspx
    Please remember to click “Mark as Answer” on the post that helps you, and to click “Unmark as Answer” if a marked post does not actually answer your question. This can be beneficial to other community members reading the thread. Regards, Mike J. [MSFT]
    This posting is provided "AS IS" with no warranties, and confers no rights.

  • What does "Scale out deployment (Web Farms)" mean for SP2013 and SQL 2012 SSRS?

    Dear community,
    in the
    Feature List of Reporting Services 2012, it's written that "Scale out deployment (Web farms)" is not possible for SQL 2012 Standard edition. However, does this relate as well to SSRS in SP-integrated mode, or only to native mode?
    In other words: is it possible to deploy SSRS using SQL 2012 STANDARD on more than one server in a SP2013 farm?
    And just to be sure: am I right to assume that in both cases, you can create multiple SSRS Svc Applications (be they all running on the same server) and assign them to different web applications (through their proxy groups)?
    Cheers,
    C

    Hi CHinz, 
    According to your description, you want to configure scale-out deployment for SQL Server Reporting Services 2012 standard edition. Right?
    As you can see in the msdn liabrary, for SQL Server 2012, scale-out deployment is only supported in Enterprise and Business Intelligence editions. This is related to both native mode and sharepoint integrated mode.
    It's not supported to deploy a SSRS 2012 standard edition instance on more than one server.
    Since scale-out deployment is for different report server instance to share one report server database. You can use different web applications to access the same report in different report server instances.
    If you have any question, please feel free to ask.
    Best Regards,
    Simon Hou

Maybe you are looking for

  • Interfaces - (FI/CO)  - GL Account Balances Extract To Legacy

    Hi , I Need to extract data from coss and cosp tables based on selection criteria given in Functional design.Company code and Fiscal Year and Fiscal Period is the selection screen input. From cosp and coss  tables LEDNR,objnr,wrttp,versn,kstar,beknz,

  • What does Safari/Mac do for FTP these days?

    i am trying to bookmark an ftp download site for my website that is of the form: ftp://[email protected]//[email protected] i know this is a mouthful but it is the correct url and when i enter it into Safari it kicks it out into Firefox which I also

  • I am steaming - Thru CS3 - CS4 and now CS5 Fireworks is crap

    All through CS3, CS4 and now CS5 Fireworks on 3 different machines Windows XP, Windows Vista and now Windows 7... simply is a crappily written app. And I have been using Fireworks since Macromedia and these problems have only appeared from Adobe. Is

  • How to download the SAP client into text file

    Hi, I want to download a SAP client to a text file for backup and then upload it again for restoration when needed. Can this be done?

  • Processing EIC follow up via the Universal Worklist (UWL)

    Hi We are implemented EIC (Employee Interationc Centre) ERP and have an issue whereby the follow up appears in both the agent inbox within EIC and the Universal Worklist.   The workflow for follow up is WS61000086.  How do we get the follow up to wor