Setting SSDT on the build server

I’m having issues building the database project on the build server.
I’m trying to setup our build server.  Our application was initial created using VS 2010 and then slowly we migrated to VS 2013. However the path inside all the individual project file (.csproj, .sqlproj) & build project file (.proj)
is still V10.0. Based on research I did, it looks like I have to install Visual Studio 2010 on the build server in order SSDT to build.
So I installed VS 2010 Professional on build server, then trying to install SSDT for 2010. However Microsoft no longer supports SSDT for VS 2010. Website says
“We are no longer updating SSDT for Visual Studio 2010. Projects and DACPACs are fully compatible across shells. Please download the toolset for VS2012 or VS2013 using the links above for continued updates”
So I installed SSDT for VS 2012.
However it installed the SSDT into C:\Program Files (x86)\MSBuild\Microsoft\VisualStudio\v11.0
folder.
But the build is looking into C:\Program Files (x86)\MSBuild\Microsoft\VisualStudio\v10.0
How do I solve this issue? Where can I find SSDT for VS 2010.

Hi lax4u,
Does passing the argument /p:VisualStudioVersion=11.0 to MSBuild when building on the build server help?

Similar Messages

  • Failed to set Subscriptions on the WSUS Server

    We are running SCCM 2012 R2; synchronization of updates is not working.
    The SUP is running on a Windows 2008 R2 server and it is not the Site Server.
    Below I have some log data. I have already tried uninstalling and reinstalling WSUS.
    WCM.log
    Starting WSUS category sync from upstream... SMS_WSUS_CONFIGURATION_MANAGER 1/6/2015 1:14:22 PM 4488 (0x1188)
    Refreshing categories from WSUS server SMS_WSUS_CONFIGURATION_MANAGER 1/6/2015 1:14:27 PM 4488 (0x1188)
    Attempting connection to WSUS server: cobalt.ad.northcentral.edu, port: 8530, useSSL: False SMS_WSUS_CONFIGURATION_MANAGER 1/6/2015 1:14:27 PM 4488 (0x1188)
    Successfully connected to server: cobalt.ad.northcentral.edu, port: 8530, useSSL: False SMS_WSUS_CONFIGURATION_MANAGER 1/6/2015 1:14:27 PM 4488 (0x1188)
    Successfully refreshed categories from WSUS server SMS_WSUS_CONFIGURATION_MANAGER 1/6/2015 1:14:53 PM 4488 (0x1188)
    Attempting connection to WSUS server: cobalt.ad.northcentral.edu, port: 8530, useSSL: False SMS_WSUS_CONFIGURATION_MANAGER 1/6/2015 1:14:53 PM 4488 (0x1188)
    Successfully connected to server: cobalt.ad.northcentral.edu, port: 8530, useSSL: False SMS_WSUS_CONFIGURATION_MANAGER 1/6/2015 1:14:53 PM 4488 (0x1188)
    Category Company:94d731de-22a6-4458-dc4d-b5267de026fc (Adobe Systems, Inc.) not found on WSUS SMS_WSUS_CONFIGURATION_MANAGER 1/6/2015 1:14:53 PM 4488 (0x1188)
    Subscription contains categories unknown to WSUS. SMS_WSUS_CONFIGURATION_MANAGER 1/6/2015 1:14:53 PM 4488 (0x1188)
    Failed to set Subscriptions on the WSUS Server. Error:(-2147467259)Unspecified error SMS_WSUS_CONFIGURATION_MANAGER 1/6/2015 1:14:53 PM 4488 (0x1188)
    STATMSG: ID=6603 SEV=E LEV=M SOURCE="SMS Server" COMP="SMS_WSUS_CONFIGURATION_MANAGER" SYS=ALGAE.ad.northcentral.edu SITE=NCU PID=1204 TID=4488 GMTDATE=Tue Jan 06 19:14:53.974 2015 ISTR0="cobalt.ad.northcentral.edu" ISTR1=""
    ISTR2="" ISTR3="" ISTR4="" ISTR5="" ISTR6="" ISTR7="" ISTR8="" ISTR9="" NUMATTRS=0 SMS_WSUS_CONFIGURATION_MANAGER 1/6/2015 1:14:53 PM 4488 (0x1188)
    Waiting for changes for 59 minutes SMS_WSUS_CONFIGURATION_MANAGER 1/6/2015 1:14:53 PM 4488 (0x1188)
    wsyncmgr.log
    Found active SUP SUP.ad.domain.tld from SCF File.SMS_WSUS_SYNC_MANAGER 
    1/6/2015 1:10:45 PM    4576 (0x11E0)
    Sync failed: WSUS update source not found on site NCU. Please refer to WCM.log for configuration error details.. Source: getSiteUpdateSourceSMS_WSUS_SYNC_MANAGER           
    1/6/2015 1:10:45 PM          
    4576 (0x11E0)
    STATMSG: ID=6703 SEV=E LEV=M SOURCE="SMS Server" COMP="SMS_WSUS_SYNC_MANAGER" SYS=Siteserver.ad.domain.tld SITE=NCU PID=1204 TID=4576 GMTDATE=Tue Jan 06 19:10:45.260 2015 ISTR0="getSiteUpdateSource" ISTR1="WSUS update source
    not found on site NCU. Please refer to WCM.log for configuration error details." ISTR2="" ISTR3="" ISTR4="" ISTR5="" ISTR6="" ISTR7="" ISTR8="" ISTR9="" NUMATTRS=0SMS_WSUS_SYNC_MANAGER         
    1/6/2015 1:10:45 PM          
    4576 (0x11E0)
    Sync failed. Will retry in 60 minutes        
    SMS_WSUS_SYNC_MANAGER 1/6/2015 1:10:45 PM           
    4576 (0x11E0)
    Setting sync alert to active state on site NCU    
    SMS_WSUS_SYNC_MANAGER 1/6/2015 1:10:45 PM     
    4576 (0x11E0)
    Sync time: 0d00h00m00s   SMS_WSUS_SYNC_MANAGER
    1/6/2015 1:10:45 PM          
    4576 (0x11E0)
    WSUSCtrl.log
    Failed to create instance of Microsoft.SystemsManagementServer.WSUS.WSUSServer. error = Unspecified errorSMS_WSUS_CONTROL_MANAGER        
    1/6/2015 12:39:11 PM        
    4888 (0x1318)

    Blank out/remove all of the categories you have configured in ConfigMgr on the Software Update Point. Then run a full synchronization from the console. This will synch up all available categories without trying to subscribe to any. Then, you can go back
    re-select the categories that want (after the full synch completes) and perform another full synch.
    Right now, you are trying to subscribe to a category that doesn't exist.
    Jason | http://blog.configmgrftw.com | @jasonsandys

  • Failed to set Subscriptions on the WSUS Server. Error:(-2147467259)Unspecified error

    Recently restored SCCM 2012 SP1 from Sever 2008 R2, to Server 2012 R2. I had Updates Publisher installed on old 2008 Server with Firefox and Chrome imported into Configuration Manager. Since I have upgraded I am no longer able to sync updates with Microsoft.
    WCM.log says:
    Category Product:44048288-2aac-b2b5-3730-fc020622ea05 (Firefox) not found on WSUS
    Category Product:cc5cb1bb-6b87-94ae-f96a-f758195112a7 (Chrome) not found on WSUS
    Subscription contains categories unknown to WSUS.
    Failed to set Subscriptions on the WSUS Server. Error:(-2147467259)Unspecified error
    STATMSG: ID=6603 SEV=E LEV=M SOURCE="SMS Server" COMP="SMS_WSUS_CONFIGURATION_MANAGER" SYS=wdlsccmpri.aarcorp.com SITE=AAR PID=2788 TID=4668 GMTDATE=Mon Feb 17 18:22:43.654 2014 ISTR0="wdlsccmpri.aarcorp.com" ISTR1=""
    ISTR2="" ISTR3="" ISTR4="" ISTR5="" ISTR6="" ISTR7="" ISTR8="" ISTR9="" NUMATTRS=
    I have gone to deselect these products from syncing in Administration> Site > Configure Site Components >Software Update Point Products Tab.  They are not there. Any help on getting SCCM to sync updates would be greatly appreciated.

    PROBLEM
    =========
    Unable to sync after moving the Primary site server to different OS.
    CAUSE
    =======
    Issues with the product and classification setup in the SUP after the restore.
    RESOLUTION
    ===========
    Ran below query to check if we have reference to Firefox and Chrome categories in the database.
    QUERY:
    select CategoryInstance_UniqueID, CategoryInstanceName
    from fn_ListUpdateCategoryInstances(9) cat
    where cat.AllowSubscription=1 and cat.IsSubscribed=1 and cat.IsParentSubscribed=0
    order by cat.CategoryInstance_UniqueID
    RESULT:  [Along with other products we found Chrome and Firefox]
    Product:44048288-2aac-b2b5-3730-fc020622ea05  Firefox
    Product:cc5cb1bb-6b87-94ae-f96a-f758195112a7 Chrome
    Ran Select Query
    select * from CI_CategoryInstances where categoryinstanceid in
    (select categoryinstanceid  from CI_LocalizedCategoryInstances where LocaleID = 9 and CategoryInstanceName = 'Firefox'or
    CategoryInstanceName = 'Chrome')
    RESULT:
    CategoryInstanceID CategoryInstance_UniqueID CategoryTypeName DateLastModified SourceSite ParentCategoryInstanceID IsDeleted rowversion
    16777523 Product:cc5cb1bb-6b87-94ae-f96a-f758195112a7 Product 2013-11-27 21:05:23.000 AAR 16777522 0 0x0000000006B08673
    16777549 Product:44048288-2aac-b2b5-3730-fc020622ea05 Product 2013-11-27 21:05:23.000 AAR 16777536 0 0x0000000006B08676
    Ran Select Query
    select * from CI_LocalizedCategoryInstances where LocaleID = 9 and CategoryInstanceName = 'Firefox'or
    CategoryInstanceName = 'Chrome'
    RESULT:
    CategoryInstanceID LocaleID CategoryInstanceName rowversion
    16777523 9 Chrome 0x0000000006B08674
    16777549 9 Firefox 0x0000000006B08677
    We did find Chrome and Firefox items in the database.
    We need to delete these items in the database.
    We used below queries to delete the reference from the database.
    delete from CI_CategoryInstances where categoryinstanceid in
    (select categoryinstanceid  from CI_LocalizedCategoryInstances where LocaleID = 9 and CategoryInstanceName = 'Firefox'or
    CategoryInstanceName = 'Chrome')
    delete from CI_LocalizedCategoryInstances where LocaleID = 9 and CategoryInstanceName = 'Firefox'or
    CategoryInstanceName = 'Chrome'
    After deleting performed Scheduled Sync. Was successful

  • [svn:osmf:] 11045: Increasing timer interval in attempt to fix unit test on the build server.

    Revision: 11045
    Author:   [email protected]
    Date:     2009-10-21 02:32:45 -0700 (Wed, 21 Oct 2009)
    Log Message:
    Increasing timer interval in attempt to fix unit test on the build server.
    Modified Paths:
        osmf/trunk/framework/MediaFrameworkFlexTest/org/osmf/composition/TestParallelViewableTrai t.as

    In general theory, one now has the Edit button for their posts, until someone/anyone Replies to it. I've had Edit available for weeks, as opposed to the old forum's ~ 30 mins.
    That, however, is in theory. I've posted, and immediately seen something that needed editing, only to find NO Replies, yet the Edit button is no longer available, only seconds later. Still, in that same thread, I'd have the Edit button from older posts, to which there had also been no Replies even after several days/weeks. Found one that had to be over a month old, and Edit was still there.
    Do not know the why/how of this behavior. At first, I thought that maybe there WAS a Reply, that "ate" my Edit button, but had not Refreshed on my screen. Refresh still showed no Replies, just no Edit either. In those cases, I just Reply and mention the [Edit].
    Also, it seems that the buttons get very scrambled at times, and Refresh does not always clear that up. I end up clicking where I "think" the right button should be and hope for the best. Seems that when the buttons do bunch up they can appear at random around the page, often three atop one another, and maybe one way the heck out in left-field.
    While I'm on a role, it would be nice to be able to switch between Flattened and Threaded Views on the fly. Each has a use, and having to go to Options and then come back down to the thread is a very slow process. Jive is probably incapable of this, but I can dream.
    Hunt

  • Setting ANT_OPTS inside the build.xml

    I am using OAS 10.1.3. For the build process, i am using the ant and jdk inside the OAS itself.
    My problem is that, i need to set the ANT_OPTS everytime before i run the build script.
    I dnt want to do this. Instead, i want to set this somewhere inside the build file itselves.
    I understand that, i can set the ANT_OPTS inside ant.bat in $ANT_HOME and can also set ANT_OPTS in env variables.
    Is there any way, that i can set this option inside the build.xml ?

    Go to ant.jakarta.org. They have numerous tutorials. Do some research first. If you run into an actual problem, then post what it is.
    - Saish

  • Automated build server for Arch? (like the sourceforge build system)

    Has someone considered some kind of automated build system for Arch?
    Something that would work like this:
    - It'd have every library and dependency possible installed.
    - It'd intelligently read the makefiles produced to see what libraries they used and compare them against a lookup table to see what Arch dependencies they then required. Failing that, it could use ldd and a second lookup table that matched libraries to packages.
    - It'd attempt to figure out the target binary to run (again, from the makefiles produced), and then run it. If it worked, it'd be marked as usable. If it didn't work, it'd be marked as needing fixing.
    All of these points can fail, especially in the parsing of the makefiles; in each case, this would be noted by the system and user action could be taken.
    In operation, it wouldn't take away from users managing their own packages. It'd just provide a secure environment to build packages in, and attempt to automate some of the process. In the best cases, the system would theoretically be capable enough to download a package's sourcecode, ./configure it, make it, make a package out of it, get the package verified as usable, then update the repo with it.
    Note the verification step in the previous paragraph: I would never want this to be an automated system. Sure, it sound amazing on paper, and might even work for a little while, but sooner or later something would come crashing down and since repo management is quite a trust-based issue, everyone would freak out and they wouldn't want the build server anymore.
    -dav7
    Last edited by dav7 (2008-10-17 18:41:22)

    Who would have access to upload to such a build server? If it's the general public, then this is a security nightmare, as well as a growth curve nightmare. The monetary investment for a project like this would need to come from somewhere.
    And yeah, something like this has been considered, and a working proof-of-concept has been sitting around for years. http://projects.archlinux.org/?p=pacbuild.git;a=summary . What this kind of project really needs is someone with some distributed computing smarts and dedication (and time) to get it off the ground in a form that will survive past a proof-of-concept barebones implementation.
    One of the largest design challenges would be dependency resolution for batch upgrades. For instance, let's say we update libfoobar, which is depended upon by foo, bar, baz, and batman, the system needs to know that libfoobar needs to be built and installed in order to compile the rest of them against it.

  • SSDT, SSDT-BI, and SQL Server Tooling in VS 2013

    I am upgrading from SQL Server 2008 R2 to SQL Server 2014, and I'm a little confused as to the upgrade path for SQL Server Business Intelligence Studio (BIDS). I'd thought that SQL Server Data Tools (SSDT) was replacing BIDS in subsequent versions of SQL
    Server i.e., 2012 and 2014. However, when I go to install SQL Server 2014, I don't see an option for SSDT. Do I need to download and install SSDT seperately in the form of SSDT 2012, SQL Server Tooling in VS 2013, SSDT-BI for 2013?
    http://msdn.microsoft.com/en-us/data/hh297027
    What's the difference between SSDT and SSDT-BI?
    A-
    Adam

    Hi Adam,
    >> Do I need to download and install SSDT separately in the form of SSDT 2012, SQL Server Tooling in VS 2013, SSDT-BI for 2013?
    In addition to Piotr’s post, you can install SSDT for 2014 via the following method:
    In Visual Studio 2013, go to the Tools menu, select “Extensions and Updates”. On the left of the window there is a section named “Updates”, and under that is “Product Updates”, then look for the Microsoft SQL Server Update for database tooling and click update.
    >> What's the difference between SSDT and SSDT-BI?
    SSDT-BI is a distinct toolset from SSDT or the SQL Server database tooling in Visual Studio, it includes BI project authoring tools and project templates for SSIS/SSAS/SSRS. However, SSDT does not include the BI templates, it is used for building databases.
    Thanks,
    Lydia Zhang
    If you have any feedback on our support, please click
    here.

  • (Deployment) Minimum Software required for BUILD server and IIS WebServer

    We have a simple .Net SDK application written using VB-2005 and the Crystal Reports XI (R2) DEVELOPER package.
    The RPT files are posted on our BOE-XI (R2) server, and are called using the View-On-Demand method from the Page Server (similar to the SDK sample application). At some point in the future, our .Net applications may invoke some of the other SDK content types.
    In our environment, the .NET web-applications are checked-in to Clear Case, and then sent to a stand-alone BUILD server to be compiled. The compiled .NET web-applications are then posted to a stand-alone IIS WebServer (not the IIS server that hosts BOE InfoView/CMC).
    The BUILD server and IIS WebServer have to be as "stripped-down" as possible in order to go through our QAT and Impact Analysis process.
    So far we have had success if we deploy the SDK-only install from the Crystal Reports XI (R2) DEVELOPER package, and the service packs, but it doesn't include all the SDK components for other BOE content types.
    If we deploy the Client-only install from the BOE-XI (R2) package, and the service packs - it doesn't seem to include the CR-Viewer which we need. But it does includes a bunch of extra software (eg. UNIVERSE DESIGNER) on the desktop of the BUILD server and IIS WebServer that our QA guys don't like.
    What are the Minimum Software required for BUILD server and IIS WebServer to support CR and BOE SDKs...?
    Are there any install packages that contain JUST those components, and not all the extra software that is not needed to Build/Deploy in .NET...?
    Thanks in advance for the advice.

    Due to the CrystalDecisions.web.dll for the 2.0 framework being left off the BOE CD you will need to install the Crystal XI R2 runtimes as well as the BOE SDK.
    To get the BOE SDK launch the Enterprise XI R2 installer and choose Server Install - not Client Install.  Choose a Custom install.  You'll be presented with a drop down list for the Enterprise components.  Deselect everything and then look for the Developer Components option.  You may have to drill down into one of the levels to find it.  Under Developer Components you'll see the BusinessObjects Enterprise .NET SDK.  Select it to install it.
    Jason

  • Prevent Scheduled Build from Running for Check-Ins Committed by Build Server

    Hi there,
    I am using TFS 2013.
    I have a scheduled nightly build that runs only when there have been check-ins.
    The problem is that my build process checks-out/in a couple files used for custom versioning. This is done under the machine account of the build server.
    Even though no check-ins were committed by developers - the scheduled build sees the check-ins done by the build server and thus, executes another nightly build.
    Is there a way I can prevent scheduled builds from running if the only check-ins were from the build machine?
    Cheers,
    Keith

    HI Keith,
    I don't think we can achieve this kind of conditional trigger from TFS build definition. We had a similar situation so we disabled the TFS build and started the build using Powershell (windows scheduled task). Below is the powershell script. The scripts
    gets the history for the day and checks for the user passed as parameter if even one changeset is not by the user the script will queue a build. 
    Param(
    $BuildUser = "Account\ToIgnore",
    $TfsUrl = "http://ABCD:8080/tfs/Collection",
    $TeamProjectName = "TeamProject",
    $BuildDefinitionName = "BuildDefinitionName")
    $today = Get-Date
    $yestday = (Get-Date).AddDays(-1)
    $Daterange = $today.ToString("Dyyyy-MM-dd HH:mm:ssZ") + "~" + $yestday.ToString("Dyyyy-MM-dd HH:mm:ssZ")
    $tfsServer = get-tfsserver $TfsUrl
    $Changeset = Get-TfsItemHistory -Server $tfsServer -HistoryItem "$/TeamProject/XYZ" -Version $daterange -User $BuildUser -Recurse
    $Changeset | Out-File C:\Windows\temp\Output.txt
    $TriggerBuild = $False
    foreach ($user in $Changeset)
    if ($user.Owner -ne $BuildUser)
    $TriggerBuild = $True
    break
    if ($TriggerBuild)
    [void][System.Reflection.Assembly]::LoadWithPartialName("Microsoft.TeamFoundation.Client")
    [void][System.Reflection.Assembly]::LoadWithPartialName("Microsoft.TeamFoundation.Build.Client")
    $tfs = [Microsoft.TeamFoundation.Client.TeamFoundationServerFactory]::GetServer($TfsUrl)
    $buildserver = $tfs.GetService([Microsoft.TeamFoundation.Build.Client.IBuildServer])
    [Microsoft.TeamFoundation.Build.Client.IBuildDefinition] $BuildDefinition = $buildserver.GetBuildDefinition($TeamProjectName,$BuildDefinitionName)
    [Microsoft.TeamFoundation.Build.Client.IBuildRequest] $BuildRequest = $BuildDefinition.CreateBuildRequest();
    $buildserver.QueueBuild($BuildRequest) | Out-File C:\Windows\temp\Output.txt
    Hope this help
    Bharath
    Regards, Bharath
    LinkedIn:

  • Not able to set up the preview server

    I've referred ATG documentation for preview server setup, tried to set up the preview on BCC.
    Please find the steps I've followed to set up preview.
    1) Added the store module t as a required module to the versioned module, so that my versioned ear include all the pages of preview in my versioned EAR.
    2) Added the 'preview' layer to the runAssembler while creating the build.
    3) Created the preview configuration files as mentioned below.
    PreviewHost.properties
    ==============
    $class=atg.service.preview.PreviewHost
    hostName=localhost
    port=8380 #port where my BCC is running
    PreviewManager.properties:
    =================
    $class=atg.remote.commerce.assetmanager.preview.service.MerchandisingPreviewService
    # Explicitly sets the document.domain property in both the BCC page
    # and the preview page to "oracle.com"
    previewDomain=localhost
    defaultPreviewUserId=123456 # This is existing user profile id.
    PreviewPath.properties:
    ==============
    $class=atg.service.preview.PreviewPath
    # Explicitly sets the document.domain property in both the BCC page
    # and the preview page to "oracle.com"
    defaultPath=/myAppContext/
    PreviewURL.properties:
    ==============
    $class=atg.service.preview.PreviewURL
    name=PreviewURL
    host=/atg/dynamo/service/preview/PreviewHost
    path=/atg/dynamo/service/preview/PreviewPath
    types=ProductCatalog=product|category
    PreviewURLManager.properties:
    ===================
    $class=atg.service.preview.PreviewManager
    previewURLs=/atg/dynamo/service/preview/PreviewURL
    4) Now created the Ear with the required modules for versioning and preview.
    Application is starting without error, but when I access my application at "http://localhost:8310/myAppContext" I'm getting the following error.
    HTTP Status 403 - An attempt was made to preview a project without logging in as a user with access to publishing assets. If you are running a local preview instance, the preview URLs registered with the PreviewURLManager may be misconfigured. Ensure that the preview URLs are configured with the same host and domain name used to access the BCC server. This preview request URL is: http://localhost:8380/myAppContext/. If you are running a remote preview instance, the Preview feature does not support remote preview on servers running the BCC. Ensure that the remote preview instance is not running any of the BCC related modules. If you received this error and did not try to perform a remote preview, then you may need to disable preview in the current web application by setting the atg.preview context parameter to false in the webapps web.xml file.
    Note: I've enabled the preview for my versioned and non-versioned applications.
    Can anyone suggest me what went wrong in my configuration?. Please provide the steps you have followed to set up the preview for your application.
    Your help is very much appreciated.
    Edited by: 989774 on Feb 22, 2013 7:09 AM

    I've tried this solution, but the problem is not resolved. Is there any alternate solution available?

  • Creation of a Marketing Profile Set based on the standard CRM_MKTTG_BP_BIRTHDATE Infoset thru the WebUI Segment Builder

    Objective: Creation of a Marketing Profile Set based on the standard CRM_MKTTG_BP_BIRTHDATE Infoset thru the WebUI Segment Builder.
    Problems: Standard CRM_MKTTG_BP_BIRTHDATE Infoset DOES NEVER RETURN ANY RESULTS on the segmentation builder of the WebUI client (no matter the filter you specify).
    Other cases:  However if we create a new Data source with data "Extraction from Query" (transaction RS02) based on this standard CRM_MKTTG_BP_BIRTHDATE Infoset and we test the functionality using transaction RSA3, IT DOES RETURN RESULTS (using the same filters as in the segmentation builder)! Of course the segmentation builder works with other attribute lists.
    NEW: IF WE DON’T ESPECIFY ANY SEGMENTATION BASE IN THE PROFILE GROUP, IT DOES RETURN RESULTS!!! (GO TO THE END OF THE DOCUMENT FOR DETAILS)
    Steps followed:
    Review Standard Infoset CRM_MKTTG_BP_BIRTHDATE (Transaction SQ02).Review Data Reading Program code: do we have to change the standard code? We have only remove the "BREAK-POINT" statement.
    Creation of a Data source based on the standard CRM_MKTTG_BP_BIRTHDATE Infoset (Transaction CRMD_MKTDS).
    Creation of an Attribute List based on that standard Data source and selection the Birth date as a filter in the Attribute list (Transaction CRMD_MKTDS).
    On the segment builder of the WebUI set any filter (Example: “TODAY”).
    Steps in detail:
    Review Standard Infoset CRM_MKTTG_BP_BIRTHDATE
    SQ02.
    (Image deleted)
    Review Data Reading Program code (Code attached)
    (Image deleted)
    Creation of a Data source based on the standard CRM_MKTTG_BP_BIRTHDATE Infoset
    Data source: Transaction: SPRO – Customer Relationship Management - Marketing – Segmentation - Maintain Data Sources and Attribute Lists: CRMD_MKTDS
    Create Data Source
    (Image deleted)
    Infoset: CRM_MKTTG_BP_BIRTHDATE.
    Business Partner Field: CRMT_MKTTG_IS_BIRTHDATE-PARTNER_GUID
    (Image deleted)
    Result:
    (Image deleted)
    Attribute List
    Creation of the Attribute List
    (Image deleted)
    Assign Data Source:
    (Image deleted)
    Selection of the Birth date as a filter in the Attribute list:
    (Image deleted)
    Result:
    (Image deleted)
    Segment builder of the WebUI:
    Access to WebUI with marketing role.
    Create Segment:
    Creation of Profile Set
    Graphical modeler
    Create Filter: We have already tried many different filters (we already seen notes referring to this matter).
    Example1: “today”
    Count target group. Here I would like to stand out that once you click on the “Count” option, immediately it shows 0 results. In other words, it doesn’t process any information. With other attribute lists it takes some time to process, no matter the filter or the attribute list you use. However if we create a new Data source with data "Extraction from Query" (transaction RS02) based on this standard CRM_MKTTG_BP_BIRTHDATE Infoset and we test the functionality using transaction RSA3, IT DOES RETURN RESULTS (with “today” filter and other)! So it seems that the problem has to do with the segment builder, not the Infoset.
    Result: it never return any results no matter the filter you specify.
    Other cases:
    Creation of a new Data source with data "Extraction from Query" where you enter the name of your new query
    Creation of a Data source based on the standard CRM_MKTTG_BP_BIRTHDATE Infoset
    Standard Infoset CRM_MKTTG_BP_BIRTHDATE:
    We fill the form and select “Extraction from Query”:
    Press SAVE and define which fields will be available for selection and which of them will be visible.
    RSA3: Extractor checker in order to test the Data source:
    We specify the Data source:
    IT DOES RETURN RESULTS:
    ALV Grid:
    NEW: It seems that it has to do with the Segmentation Basis selected in the creation of a Profile Set, because if we do not select any, it does return results. But we need to create this Profile Set based on a particular Segmentation Basis called BS Usuarios Registrados.
    Segment builder – Profile Set of the WebUISegmentation Basis: we leave it in blanc.
    Graphical modeler
    Create Filter: We have already tried many different filters (we already seen notes referring to this matter).
    Example1: “today”
    Count target group.
    Result: IT DOES RETURN RESULTS!!! àConclusion: the problem is the Segmentation Basis
    Define Segmentation Basis:
    Transaction: SPRO – Customer Relationship Management - Marketing – Segmentation – Classic Segmentation - Define Segmentation Basis.BS Usuarios Registrados:
    Define Reports for Creating Segmentation Bases
    Transaction: SPRO – Customer Relationship Management - Marketing – Segmentation – Classic Segmentation - Define Reports for Creating Segmentation BasesZMK_UTL_MGR_SITES_UREG
    Define Usage for Segments:
    Transacción: SPRO – Customer Relationship Management - Marketing – Segmentation – General Settings - Define Usage for Segments.ZBS_UR à02 Segmentatión Basis
    Attribute List:Category:
    There is no Category 02 Segmentatión Basis!!! Why?? Why is not posible to select Category 02 Segmentation Basis in the creation of the Attribute List???
    Usage ID
    NEW: It seems that it has to do with the Segmentation Basis selected in the creation of a Profile Set, because if we do not select any, it does return results. But we need to create this Profile Set based on a particular Segmentation Basis called BS Usuarios Registrados. Why is not posible to select Category 02 Segmentation Basis in the creation of the Attribute List???

    It seems that the standard Infoset doesn’t work by itself with Segmentation Basis. Two SAP notes have been applied and a new Field Group has been added to the Infoset (in order to distint the Segmentation Base).
    CRMT_MKTTG_IS_BIRTHDATE-MGRGUID
    1966298: MGR_GUID is not populated correctly to infoset read program
    2007687:Segmentation InfoSet queries saved with incorrect Master Group
    For that reason we had to create a new Infoset based on the standard CRMT_MKTTG_IS_BIRTHDATE, but adding this new Field MGRGUID to the Group Fields.
    Issue solved

  • TS3899 I can't SEND email from Telus account in Alberta, Canada? Does anyone know how to set up the Outgoing server? Help! And thanks!

    Can't SEND email from Telus account in Alberta, Canada, unless I go to web mail. Does anyone know how to set up the Outgoing server? Incoming is fine. Outgoing used to work. We changed it when we went to another location, and can't get it back. Telus support can't fix it. Neither smtp.telus.net NOR mail.telus.net works for Outgoing server to send mail. Please help! Thanks.

    iOS: Unable to send or receive email
    http://support.apple.com/kb/TS3899
    Can’t Send Emails on iPad – Troubleshooting Steps
    http://ipadhelp.com/ipad-help/ipad-cant-send-emails-troubleshooting-steps/
    Setting up and troubleshooting Mail
    http://www.apple.com/support/ipad/assistant/mail/
    Server does not allow relaying email error, fix
    http://appletoolbox.com/2012/01/server-does-not-allow-relaying-email-error-fix/
    Why Does My iPad Say "Cannot Connect to Server"?
    http://www.ehow.co.uk/info_8693415_ipad-say-cannot-connect-server.html
    iOS: 'Mailbox Locked', account is in use on another device, or prompt to re-enter POP3 password
    http://support.apple.com/kb/ts2621
    iPad Mail
    http://www.apple.com/support/ipad/mail/
    Try this first - Reset the iPad by holding down on the Sleep and Home buttons at the same time for about 10-15 seconds until the Apple Logo appears - ignore the red slider - let go of the buttons. (This is equivalent to rebooting your computer.)
    Or this - Delete the account in Mail and then set it up again. Settings->Mail, Contacts, Calendars -> Accounts   Tap on the Account, then on the red button that says Remove Account.
     Cheers, Tom

  • Problem with the new server UCS C220 for set IP to CIMC

    Hi
    We’ve a problem with the new server UCS C220.
    We bought two servers UCS C220 M3 for CallManager 8.6 with High Availability.
    When we turn on the server during the boot and when it tells us, oppress F8 to enter at the CIMC and set the IP. But it never enters at the CIMC.
    Then, we configure our DHCP server and our switch, we connect the three gigabyte ports to our switch to give him an IP to the CIMC, so and then can enter via browser, but neither works.
    Note. The dedicated management NIC does not link, the other two ports do make link.
    What do you suggest to put an IP to CIMC and start installing our applications?
    regards

    You may have noticed that there is no DVD rom on the c220. What you need to do is:
    Login into the CIMC from your browser
    Luanch the KVM
    Insert the VMware DVD in your machines drive
    On the KVM pop up there should be a tab to mount the drive, after mounting it click on Macros and choose ctrl_alt+delete to restart.
    After the VMware OS installs press F2 to enter IP.
    Browse to the VMware ip to download the Vsphere client
    Open the Vsphere client, enter the ip of the vmware and the username will be root and no password if you did not set one.
    You can now upload OVA templates or manually create virtua machine from this enviroment.
    Hope this help

  • How can I set up Mavericks Server on a Test Mac without affecting the production server

    I have Mountain Lion server running as on an xserve as our main server, it has developed a few issues particularly related to Profile Manager and Certificates.
    So I was going to use the move to Mavericks as an opportunity to have a fresh start.
    So I want to set up Mavericks Server on a mac mini temporarily to test it and configure all the services before moving it to the Xserve as the main server.
    My question is what is the best way to do this, obviously I can't use the same IP or hostname as the main server if it is connected to the network, and some services can't be set up without a network connection.
    So do I give it a new IP and hostnmame for now and then change it later on, my concern is that changing the hostname might cause issues down the line.
    So what is the best way to set up a fresh server in preperation as a drop in replacement of the current server?

    Ok I got the test server set up fine and cloned it over to a spare partition on the Xserve.
    But when I booted into the new partition several things broke.
    1st the IP address changed itself despite the fact it is set to Manual, The Alert section of Server.app let me repair this.
    Profile Manager no longer worked, the web page just said Profile manager could not be found, was working fine on the test mac
    Tried deleting the OD master and resetting Profile Manager but it still didn't work.
    DNS needed repairing and restarting
    I decided at that point to abandon it and go back to the test mac where everything is still working.
    Any tips for overcoming these issues.
    I have thought maybe it would be best to turn off all server services before making the clone and then turn them on one at a time once it has been restored to the new partition on the xserve
    Also if the certificates that are created when making an OD master were made on the mac mini does that cause a problem when it has been restored to a different machine, should I leave OD off and just set it up when it is in its final place?

  • When I try to set up Icloud control panel in Windows Vista, I get "your setup couldn't be started because the Icloud server stopped responding. I've tried loads of fixes suggested on the net but none worked. Can anyone help me please?

    When I try to set up Icloud control panel in Windows Vista, I get "your setup couldn't be started because the Icloud server stopped responding. I've tried loads of fixes suggested on the net but none worked. Can anyone help me please?

    Hello, ksb2. 
    Thank you for visiting Apple Support Communities.
    We are investigating this issue. This article will be updated as more information becomes available.
    iCloud: iCloud Control Panel setup cannot be started
    http://support.apple.com/kb/TS5178
    Cheers,
    Jason H.

Maybe you are looking for