Analysis Service Process Task Connection Manager

hi there, I was trying to use this Process Task inside SSIS to process Dimensions and partitions. The connection string for the AS connection manager is through a configuration file and value is
 Data Source=10.50.30.71;Initial Catalog=ORDERDW;Provider=MSOLAP.5;
 (Notice that there is no user name and password provided)
To my surprise, when I run it under SSDT 2012, the connection was tested ok and package was actually runnning withour error?  Why as it does not make any sense? Under which credentials is this package running as it definitely needs username/password
to connect to AS server.  Thanks
The package protection level is using default which is encryptSensitiveWithUserKey
--Currently using Reporting Service 2000; Visual Studio .NET 2003; Visual Source Safe SSIS 2008 SSAS 2008, SVN --

To my surprise, when I run it under SSDT 2012, the connection was tested ok and package was actually runnning withour error?  Why as it does not make any sense?
By Default the Windows Credentials are use to logon to SSAS and I guess you have permissions to logon? So where exactly do you see an issue here?
Olaf Helper
[ Blog] [ Xing] [ MVP]

Similar Messages

  • Does analysis service process task honor dynamic connections?

    simple question.. I have a ssis package which contains analysis service process task.. it went well in the dev env. Now, I've deployed to production and connection manager was set by configuration tables.. however, it seems to me that the connection manager
    on production is still pointing to the dev env... any workaround? 
    thanks
    --Currently using Reporting Service 2000; Visual Studio .NET 2003; Visual Source Safe SSIS 2008 SSAS 2008, SVN --

    Hi cat_ca,
    To achieve your goal, we can combine the Environment variable Configuration with SQL Server Configuration. In short, we can create an Environment variable type Configuration to define the connection string for the Connection Manager that is used by a SQL
    Server Configuration to connect to the SQL Server configuration table. To create such an Environment variable Configuration, we need to create a system Environment variable on each server to host the connection string for the corresponding connection manager
    such as:
    Data Source=ServerName;Initial Catalog=TestDB;Provider=SQLNCLI11.1;Integrated Security=SSPI;Auto Translate=False;
    Besides, we need to make sure the Environment variable Configuration is above the SQL Server Configurations in the Package Configurations Organizer list.
    Reference:
    http://amenjonathan.wordpress.com/2010/12/13/creating-a-solid-ssis-etl-solution-part-2-%E2%80%93-package-configurations/ 
    Regards,
    Mike Yin
    If you have any feedback on our support, please click here
    Mike Yin
    TechNet Community Support

  • SSIS Analysis Services Processing Task - which object is being processed?

    Hi
    I've just fallen down the hole dug by a gotcha in the SSIS Analysis Services Processing task (let's call it the ASPT as it's too long to type).
    I was doing testing yesterday, targeting a test database.  The way things are set up here is that we have two structurally-identical databases (let's call them SSASLive and SSASTest) on the same SSAS server.
    The SSIS package includes an Analysis Services Processing task that processes a partition of a measure group in a cube.
    So I changed the properties of the AS Connection Manager in the package to point to database SSASTest.  Then, because I'm paranoid, and don't trust the Editor for the ASPT, I edited the object to be processed, by deleting the existing object from the
    list and re-adding it (going down the tree: database SSASTest, cube TheCube, measure group TheMassiveMeasureGroup, partition CurrentMonth).
    This is an annoying necessity, because: the ASPT editor doesn't allow you to see exactly what the object(s) in the list to be processed are, by e.g. double-clicking on them.  All you can see is the lowest-level object name, which could be just
    "Current Month".  Current Month what?  There's no way of telling.
    Test worked fine.  I made sure the AS Connection manager was reset to point to SSASLive, and deployed the package.
    This morning, the package ran, and processed the partition in database
    SSASTest.  What happened?  Looking in the Code view of the package, I found the XMLA behind the Processing task: the database name is literally specified in the XMLA, and completely ignores the Catalog of the AS connection.  This information
    is not exposed anywhere in the UI view of the package.
    It's a bit like the dangerous annoyance when testing SSIS package Exec SQL tasks, where a previous developer has insisted on fully-qualifying object names: SELECT * FROM LiveDatabase.dbo.ATable.  You're working with a connection to a test database,
    but it makes not difference...
    But at least that problem is clearly visible.  I know that specifying the database name is normal in XMLA tasks: but at least the SSIS editor should allow you to see what the setting is.

    No, the point is that the Connection Manager's Initial Catalog setting (whether hard-coded or set by any of the configuration methods) has no effect on the database addressed by the Analysis Services Processing task.
    Steps to reproduce:
    1. Create an Analysis Services database containing a cube (or select an existing one): let's call it ASDB1.
    2. Make a copy of it on the same server, e.g. by backing up and restoring: let's call it ASDB_Copy
    2. Set up an SSIS package with a Connection Manager pointing to ASDB1
    3. Create an Analysis Services Processing task, using the Connection Manager, processing any particular SSAS object.
    4. Run the package.  The SSAS object in ASDB1 is processed (not the object in ASDB_Copy).  (Confirm by checking the Last Processed date in SSMS, Properties).
    5. Change the Connection Manager's Initial Catalog setting to point to ASDB_Copy.
    6. Run the package.  The SSAS object in ASDB1 is processed, not the object in ASDB_Copy.  The Connection Manager's Initial Catalog setting has no effect on the AS Processing Task.  Examining the Code view of the package confirms that the XMLA
    behind the task specifies the database as well as the AS object. (and this database setting is still ASDB1).
    The fix would be for SSIS to perform validation of AS objects referenced by the AS Processing Task whenever its Connection manager is pointed to a different database, and to update the XMLA accordingly.
    Also, it would be useful to be able to see exactly what the objects listed in the Processing Task are.  The information shown is just not sufficient to identify the object to the developer.  I may have dozens of measure groups with a partition
    called Current Month - but "Current Month" is all I can see in the editor, and there's no way to e.g. double-click on an item in the list and get the full object identification (Cube X, measure group Y, partition Current Month).  Sure,
    I can go and look at the XMLA, but that's fiddly: if there's going to be a wrapper around the XMLA (which is what the AS Processing Task is), then it should be unambiguous.

  • Analysis Services Processing Task - Error processing

    Hi all,
    I have a SSIS package containing an Analysis Services Processing Task. In case it fails, I want to insert the error messages in a table. I have create an SSIS package that fails because of dimension values that are not in the fact table. The package fails and
    shows a number of errors in the debug window:
    Error: 0xC1000007 at Create Cube, Analysis Services Execute DDL Task: Erreur interne : L'opération n'a pas abouti.
    Warning: 0x811F0001 at Create Cube, Analysis Services Execute DDL Task: Erreurs dans le moteur de stockage OLAP : 
    Clé d'attribut introuvable lors du traitement de : Table : 'dbo_SSIS_DATA', Colonne : 'PRODCODE', Valeur : 'HUME'. L'attribut est 'PRODCODE'.
    SSIS package "Step 2 - Create Cube.dtsx" finished: Failure.
    However in the error list window there are 0 errors and 0 warnings.
    So, why doesn't it show an error message in the error list when the package clearly fails. And how can I get the error messages from the log into a table? Thanks in advance.
    Hugo

    Error list does not reflect what is getting to be logged, it is more for package validation.
    But the same errors will show up in the status, and the output window.
    Furthermore, if your log is table bound (e.g. the sysssiserrorlog) and in the log config you chose to log the OnError events then these will be captured automatically.
    Arthur
    MyBlog
    Twitter

  • SQL Server 2005 agent job runs a SSIS package ( Analysis Services Processing Task) fails

     Hi,
    SQL Server 2005 standard edition.
     I have a SSIS package which has a  Analysis Services Processing Task. I have tested the package in BIDS and it runs ok. But when I created a agent job and run it from the job it reports error:
    Code: 0xC0012024     Source: Analysis Services Processing Task      Description: The task "Analysis Services Processing Task" cannot run on this edition of Integration Services.
    It requires a higher level edition.
    This is the result of select @@version
    Microsoft SQL Server 2005 - 9.00.4035.00 (Intel X86)   Nov 24 2008 13:01:59   Copyright (c) 1988-2005 Microsoft Corporation  Standard Edition on Windows NT 5.2 (Build 3790: Service Pack 1) 
    Any idea?

     Hi,
    SQL Server 2005 standard edition.
     I have a SSIS package which has a  Analysis Services Processing Task. I have tested the package in BIDS and it runs ok. But when I created a agent job and run it from the job it reports error:
    Code: 0xC0012024     Source: Analysis Services Processing Task      Description: The task "Analysis Services Processing Task" cannot run on this edition of Integration Services. It
    requires a higher level edition.
    This is the result of select @@version
    Microsoft SQL Server 2005 - 9.00.4035.00 (Intel X86)   Nov 24 2008 13:01:59   Copyright (c) 1988-2005 Microsoft Corporation  Standard Edition on Windows NT 5.2 (Build 3790: Service Pack 1) 
    Any idea?
    Anyway, I have found a work around:
    http://technet.microsoft.com/en-us/library/ff929186.aspx

  • MS SQL Server Analysis Services (SSAS) OLAP Connection - Dimension descriptions

    Hello,
    We are trying to create some reports on BO 4.0 on top of BPC 10 MS information.
    In order to overcome some issues related to standard connector between BO and BPC (XMLA Planning and consolidation 10 for Microsoft platform), we created a new connection based on discussion MS SQL Server Analysis Services (SSAS) OLAP Connection - Business Intelligence (BusinessObjects) - SCN Wiki
    However we have a major issue with this approach. We cannot see on BO dimension member descriptions.
    Is there anything that we missed?
    Could you help?
    Regards,
    Vitor Ramalho

    Bill,
    So, what you need is a model like SQL Server failover cluster instances. (before sql server 2012)
    In SQL Server 2012, AlwaysOn replaces SQL Server failover cluster, and it has been seperated to
    AlwaysOn Failover Cluster Instances (SQL Server) and
    AlwaysOn Availability Groups (SQL Server).
    Since your requirement is not in database level, I think the best option is to use AlwaysOn Failover Cluster Instances (SQL Server).
    As part of the SQL Server AlwaysOn offering, AlwaysOn Failover Cluster Instances leverages Windows Server Failover Clustering (WSFC) functionality to provide local high availability through redundancy at the server-instance level—a
    failover cluster instance (FCI). An FCI is a single instance of SQL Server that is installed across Windows Server Failover Clustering (WSFC) nodes and, possibly, across multiple subnets. On the network, an FCI appears to be an instance of SQL
    Server running on a single computer, but the FCI provides failover from one WSFC node to another if the current node becomes unavailable.
    It is similar to SQL Server failover cluster in SQL 2008 R2 and before.
    Please refer to these references:
    Failover Clustering in Analysis Services
    Installing a SQL Server 2008 R2 Failover Cluster
    Iric Wen
    TechNet Community Support

  • Java Web Services wrapper and connection management.

    We are currently evaluating the use of the SES Java web-services for a custom front end. I've downloaded the sample code and have read through the documentation but I'm confused on a few points.
    We will be OSSO protecting the app, so the flow I think I need to use is (on the OracleSearchService object):
    *setSoapURL
    *proxyLogin
    *setSearchUser
    *doOracleSearch
    (do whatever with results)
    From what I can tell each one of these does a separate call into the actual web service. Does this keep a connection open between these calls and if so how do we need to manage our connections in this case. Is there an explicit close method we should be calling? I see that we can get an OracleSoapHTTPConnection from the undocumented getConnection method on OracleSearchService but I'm unsure if we are supposed to interact with it directly.
    In a related vein how should we managing our OracleSearchService objects? Should we save an instance for each user in our httpSession, or just create a new instance on each call?
    Any insights appreciated,
    Jason

    No Problem at all
    Nomally the Web Service is big different between Web application. or you can build your Web service structure like web application structure. It depends on the web server:
    If the Web Server of hosting company has WEB Service framework, you just need know how to deploy your Web service to Web Service framework and How to deploy to Web Server. In this case, you need deploy it two time, one is for Web Server framework and one is for Web Service framework
    If the Web Server of hosting company doesn't has WEB Service framework, you need build your own Web Service framework into a web application, then you just need know how to deploy your web application with Web service into Web Server framework

  • Dynamically set Analysis Services Connection Manager in a Fooreach loop [SSIS 2008]

    I am storing connection strings for "Analysis Services Connection Manager" in a database table and on run time i am fetching them all, storing in an object and iterating through them inside a Foreach loop. Variable gets updated with a new connection
    string in each iteration but somehow my "Analysis Services Processing Task" fails with "Invalid Connection String" Error.
    Is there any way we can set set Analysis Services Connection Manager Dynamically. I can done this with OLEDB connections and that works. But somehow it's not working for Analysis Services Connection manager. Can anyone please share his/her experience ?
    Many thanks
    Saadat

    See if this helps http://kamakshisuram.wordpress.com/working-with-ssas/ssas-connection-string-dynamically/
    The trick is in using a configuration
    Arthur My Blog

  • Permanently change default error configuration in Analysis Services 2005

    Hi,
    Currently, I am working on a BPC 5.1 application.  The data for this application is loaded(inserted via SQL statement) right to the FACT table and then a full process is run for that cube via an SSIS package using the Analysis Services Processing Task.  Often records are loaded this way where a dimension member for some of the records has not been added to the Account dimension yet.  These records after loading are considered 'orphan records' until the accounts are added to the account dimension.
    This loading process is used because of the volume of records loaded(over 2 million at a time) and the timing of the company's business process.  They will receive data sometimes weeks before the account dimension is updated in BPC with the new dimension members.
    If I try and process the application from the BPC Administration area with these orphan records in the FACT table, the processing stops and an error displays.  Then when I process the cube from Analysis services, an error is displayed telling me that orphan data was found.
    A temporary work-around is to go into the cube properties in Analysis Services 2005, click on Error Configuration, uncheck 'Use default error configuration' and select 'Ignore errors'. Then you can process the application from BPC's Administration page successfully.  But, the problem is that after processing the application successfully, the Analysis Services Error Configuration automatically switches back from 'Ignore errors' to 'Use default error configuration'.
    Does anyone have any suggestions on how to permanently keep the 'Ignore errors' configuration selected so it does not automatically switch back to 'Use default error configuration'?  Prior to BPC 5.0 this was not occurring.
    Also, does anyone know why this was changed in BPC 5.0/5.1?
    Thanks,
    Glenn

    Hi Glenn,
    I understood the problem but I can say that it was a bad migration of appset from 4.2 to 5.0.
    Any way they are using a dts package to import data into our fact table. That's means they have to add another step into that package where they have to do the verfications of records before to insert into fact table. Verfications can be done using the same mechanism from our standard import. Just edit that package and add similar steps into customer package.
    Attention you need somebody with experience developing DTS packages with for BPC to avoid other problems.
    One of big benefits from 5.X compare with 4.2 was the fact that we are able to use optimization schema and aggregations for cubes.
    Heaving that orphan records it is not possible to use optimization schema for cubes and you are not able to create good aggregation into your cube.
    So my idea is to provide all these information to customer and to try to modify that package instead to enable that option which can cause many other issues.
    Sorin

  • Can't connect to analysis services

    Hi,
    After looking through the forum, I haven't seen the problem I encounter.
    I am able to connect to the Database Engine, Integration services, Reporting Services but I can't connect to Analysis Services from SQL Server Management Studio although the service is running.
    I am using SQL Server 2012.
    I have the following error :
    Internal error: An unexpected error occurred (file 'pfshcollation.cpp', line 323, function 'PFCollation::Init'). (Microsoft SQL Server 2012 Analysis Services)
    The progam location is:
       at Microsoft.AnalysisServices.AdomdClient.AdomdConnection.XmlaClientProvider.Connect(Boolean toIXMLA)
       at Microsoft.AnalysisServices.AdomdClient.AdomdConnection.XmlaClientProvider.Microsoft.AnalysisServices.AdomdClient.AdomdConnection.IXmlaClientProviderEx.ConnectXmla()
       at Microsoft.AnalysisServices.AdomdClient.AdomdConnection.ConnectToXMLA(Boolean createSession, Boolean isHTTP)
       at Microsoft.AnalysisServices.AdomdClient.AdomdConnection.Open()
       at Microsoft.SqlServer.Management.SqlStudio.Explorer.ObjectExplorerService.ValidateConnection(UIConnectionInfo ci, IServerType server)
       at Microsoft.SqlServer.Management.UI.ConnectionDlg.Connector.ConnectionThreadUser()
    What can I do to connect?

    Hi Thush,
    According to your description, you get the above error when connecting SSAS in SSMS. Right?
    In Analysis Services, this error turns out to be a the data provider issue. I suggest you re-install the OLE DB Provider for SQL Server 2012.
    On the Installation page, choose New SQL Server stand-alone installation or add features to an existing installation.
    On the Installation Type page, choose Perform a new installation of SQL Server 2012.
    On the Setup Role page, choose SQL Server Feature Installation.
    On the Feature Selection page, click Client Tools Connectivity. This option installs
    Microsoft.AnalysisServices.Xmla.dll
    Do not select any other features.
    Click Next to finish the wizard, and then click Install to run Setup.
    Reference:
    Microsoft® SQL Server® 2012 SP1 Feature Pack
    If you have any question, please feel free to ask.
    Regards,
    Simon Hou
    TechNet Community Support

  • Cannot connect to SQL 2000 Analysis Service cube form 32bit 2010 excel

    Excel 2010 32bit is running on windows 7 64bit PC. User has new PC updated from XP where there was no connection issue.
    Initial Error
    "The following system error occurred: No connection could be made because the target machine actively refused it"
    I installed Office web component 2003 as advised in forums.
    Next Error
    "initialization of the data source failed.  Check the database server or contact your database administrator. Make sure the external database is available, and then try the operation again. If you see this message again, create a new datasource
    to connect tot he database."
    I have trawled the forums for the past 2 days looking for a solution.  Any help would be appreciated.
    Niall

    Hello,
    Please check the below link.May be this can Help You
    https://fawzi.wordpress.com/2013/05/15/sql-2012-analysis-services-error-no-connection-could-be-made-because-the-target-machine-actively-refused-it/

  • How to find Collation setting for Analysis services in Sql server 2000

    Hi
    Can some please guide me how do I check  Collation setting for Analysis services in Sql server 2000
    i need it to install Analysis servcies on my machine

    Hi SMV123,
    We can find the collation settings for Analysis services in SQL Server Management Studio (SSMS). Connect to the SSAS instance from within SSMS->Right click on the instance and select Properties->An Analysis Server properties page will appear. And then
    on the left hand side is a Language /collation option.
    Besides, here is a link about how to install SQL Server Analysis Services 2000 for your reference.
    How to Install SQL Server 2000 Analysis Servicese
    Regards,
    Charlie Liao
    Charlie Liao
    TechNet Community Support

  • SSDT Issues in VS2010 - Analysis Services Project

    Hi,
    I have installed SQL Server 2012 RC0 and CTP4 of the SSDT on my laptop with an existing VS2010 Premium SP1 installation. I hve opened an Analysis Services project and connected to a cube developed on SQL Server 2008 R2 and restored into the 2012 RC0 environment. 
    When I try to open the DSV of the Analysis Services project I get the following error:
    ===================================
    An error prevented the view from loading. (Microsoft Visual Studio)
    ===================================
    The specified module could not be found. (Exception from HRESULT: 0x8007007E) (System.Windows.Forms)
    Program Location:
    at System.Windows.Forms.UnsafeNativeMethods.CoCreateInstance(Guid& clsid, Object punkOuter, Int32 context, Guid& iid)
    at System.Windows.Forms.AxHost.CreateWithLicense(String license, Guid clsid)
    at System.Windows.Forms.AxHost.CreateInstanceCore(Guid clsid)
    at System.Windows.Forms.AxHost.CreateInstance()
    at System.Windows.Forms.AxHost.GetOcxCreate()
    at System.Windows.Forms.AxHost.TransitionUpTo(Int32 state)
    at System.Windows.Forms.AxHost.CreateHandle()
    at System.Windows.Forms.Control.CreateControl(Boolean fIgnoreVisible)
    at System.Windows.Forms.Control.CreateControl()
    at Microsoft.DataWarehouse.Design.ComponentDiagram.CreateDdsView(Control parentControl)
    at Microsoft.AnalysisServices.Design.DataSourceDesignerView..ctor(DataSourceDesigner designer, DataSourceDiagram diagram, IComponent diagramOwnerComponent)
    at Microsoft.AnalysisServices.Design.DataSourceDesignerView..ctor(DataSourceDesigner designer)
    at Microsoft.AnalysisServices.Design.DataSourceDesigner.CreateDataSourceDesignerView(VsStyleToolBar pageViewToolBar)
    at Microsoft.DataWarehouse.Design.EditorWindow.EnsureViewIsLoaded(EditorView view)
    To date I have not been abe to find a solution which is frustrating. Any ideas how I can correct this issue?
    Regards,
    Alan
    Regards, Alan

    The error message that was being returned was:
    An error prevented the view from loading. (Microsoft Visual Studio)
    ===================================
    An error occurred while attempting to start the report preview worker process. (Microsoft.ReportingServices.Designer)
    Given that this scenario worked with previous editions of Reporting Services and BI Development Studio, the error was unexpected.  On further investigation, the same behavior and error could be reproduced by launching a Remote Desktop session in Seamless
    or Remote Applications Integrated Locally (RAIL) mode.
    RAIL extends the RDP protocol presents a remote application running on a RAIL server as a local user application running on the client machine. This causes the remote application to appear as if it is running on the user's local computer rather
    than being presented in the desktop of the remote Server.
    In this scenario, the behavior and the error message are both by design.
    SSDT is a multi-targeted assembly that is built around the .NET 4.0 Framework. Business Intelligence Development Studio (BIDS) is built around the .NET 3.5 Framework and there are differences in the way that the .NET 4.0 Framework deals with child processes
    when the parent process dies. As a result, there was a deliberate decision to change the behavior of the report viewer in preview mode. This is largely because under the .NET 4.0 Code Access Security policy, executing under the current AppDomain is no
    longer supported, necessitating use of a sandbox. You may wish to read Brian Hartman's blog posting "Expression Evaluation in Local Mode" for a more detailed discussion of Visual Studio 2010 defaulting to the .NET 4.0 CAS policy. The net result of
    the change is that when a report is previewed, SSDT spawns a child process to allow previewing of the report. In the RDP using Seamless or RAIL mode, it would still be necessary for SSDT to spin up a new instance of the Reporting Services Preview Processing
    Host (PreviewProcessingService.exe) separate from the SSDT process. In the event that the user were to close SSDT or SSDT were to crash, it would be possible for orphaned instances of the Reporting Services Preview Processing Host application to remain in
    memory and continue execution after SSDT was terminated.
    The change makes it cleaner to close SSDT and eliminates the possibility of orphaned child processes in the event that the SSDT environment is terminated for some reason. Had the change not been made, it would have been possible to allow orphaned
    child applications/processes to continue to execute, using CPU cycles and consuming server memory, until the user logged off the machine. If the user didn’t log off and rather elected to simply disconnect without logging off, any orphaned applications would
    have remained in memory for an indefinite period of time. In the scenario where a machine is being remotely accessed by multiple users, that could potentially mean thousands of orphaned applications would be in memory and actively executing at any given
    time.

  • Analysis Services Execute DDL Task Internal error

    Hi all,
    I need help in solving this sporadic problem...  the dimension below (blanked out as xxxx) is based on a view.  The cube processes fine many times - but fails - abruptly once a week... it's being called from a sql job.
    Any ideas?  Thanks in advance!!
    Error: 2010-06-28 15:45:17.69     Code: 0xC1000007     Source: xxxxxxx Analysis Services Execute DDL Task     Description: Internal error: The operation terminated unsuccessfully.  End Error 
    Error: 2010-06-28 15:45:17.69     Code: 0xC11F000D     Source: xxxxxxx Analysis Services Execute DDL Task     Description: Errors in the OLAP storage engine: An error occurred while the 'xxxxx' attribute
    of the 'xxxxx' dimension from the 'xxxxxx' database was being processed.  End Error  Error: 2010-06-28 15:45:17.69     Code: 0xC11F0006     Source: xxxxxxx Analysis Services Execute DDL Task    
    Description: Errors in the OLAP storage engine: The process operation ended because the number of errors encountered during processing reached the defined limit of allowable errors for the operation.  End Error  Error: 2010-06-28 15:45:17.69    
    Code: 0xC11C0002     Source: xxxxxx Analysis Services Execute DDL Task     Description: Server: The operation has been cancelled.  End Error  DTExec: The package execution returned DTSER_FAILURE (1). 
    Started:  3:45:00 PM  Finished: 3:45:17 PM  Elapsed:  17.328 seconds.  The package execution failed.  The step failed.
    Harsh B

    (From http://msdn.microsoft.com/en-us/library/cc966526.aspx)
    "ExternalCommandTimeout is a server property that is used to set the number of seconds that SSAS should wait to time out when issuing commands to external data sources, such as relational and other OLAP sources."
    "ExternalConnectionTimeout is a server property that is used to set the number of seconds, by default, that SSAS should wait to time out when connecting to external data sources, such as relational and other OLAP sources."
    What you set it to, it's absolutely up to you... But as I see it in your log, your task failed in 17 and 9 seconds... So I think you can multiply the values by 10, I guess the situation won't change... But it is worth to check it at least.
    So... How often your fact table gets new (degenerate) dimension reference values? Is it a process update, or a process full command? I really would like to see it :)
    -- Zoltán Horváth
    -- MCITP SQL Server Business Intelligence Developer 2005, 2008
    -- Please mark posts as answered or helpful where appropriate.

  • Error when trying to view reports or manually processing the TFS data warehouse and analysis services cube

    Hello Guys,
    I am trying to configure the reporting for TFS using SQL Server. But I get following error when viewing any report:
    So I try to manually process the cube to check if it works. I am following this article: https://msdn.microsoft.com/en-us/library/ff400237.aspx
    When I click on GetProcessingStatus and invoke it (with last field set as TRUE) I get following error:
    Please advice as to how to resolve this issue and be able to see the reports.

    I have managed to resolve this issue. Note that for all purposes of this question and answer, mydomain\tfsadmin is a generic user (used to install all software). This is a proof of concept account, for test purposes
    The thing is that while installation of the SQL Analysis service I had given the username: mydomain\tfsadmin (a generic user for testing) as the Analysis service administrator, instead of the 'Domain admin' group.
    Still I have managed to resolve this issue.
    Steps:
    1. Make sure that the user (mydomain\tfsadmin) is a member of Analysis Server -> TFS_Analysis db -> Roles -> TFSWarehouseAdministrator and TFSWarehouseDataReader. (This actually happens automatically when you run the TFS Admin console and configure
    reporting and provide the username that will access the Analysis db, etc. In my case the user is 'mydomain\tfsadmin')
    2. It is a bad practice to manually process the cube (you can do this is make sure that there are no errors, but only after completing the following steps till step 5)
    3. Also make sure that NT AUTHORITY\NETWORK SERVICE is member of the Analysis db -> Roles -> TFSWarehouseAdministrator ->
    this will resolve the error which appears in the 2nd screenshot in the question.
    4. Then you can right click on the Analysis db and run process. If this is throwing error as shown in above reply of mine, then you need to follow https://msdn.microsoft.com/en-us/library/vstudio/ff400237.aspx
    a. http://localhost:8080/tfs/TeamFoundation/Administration/v3.0/WarehouseControlService.asmx
    b. Choose
    ProcessWarehouse, run it.
    c. Choose ProcessAnalysisDatabase, type 'Full' and run it.
    d. Choose GetProcessingStatus,
    enter 'TRUE' in last field and run it.
    e. I don't get any errors at this point.
    5. Now you can connect to the Analysis server via SQL Mgmt Console, right click on Analysis db (TFS_Analysis) and click Process. It all works good.
    6. Now you can browse to the report URL (to get this URL you can open Team explorer 2013, connect to you team project, on RHS click 'Reports', Click 'Go to Site'.
    DONE.

Maybe you are looking for

  • TS2972 How can I stop iTunes on my computer to start whenever I use the Apple TV remote to select a new program?

    Whenever I use the Apple TV remote to choose a new program, iTunes starts on my computer and starts playing from the playlist. How do I prevent that from happening?

  • Why do I get a memory leak when I use the Write Data Storage VI in a loop

    I have a large application that I am collecting various data points from. For the most part they are simple scalar values that need to be logged on a regular basis.  I was attempting to use the new TDM files and the storage VI's  I though everything

  • Problems restoring Ipad 3

    I am having problems with my Ipad 3. I went to use it and suddenly i wasnt able to input my password in the unlock screen. I could press the numbers but nothing would seem to show up in the 4 slots to input my password. The screen wasnt frozen and i

  • Airport Express and Audio

    I'm new to Macs and Airport but so apologies for my ignorance... but I want to know is it possible to send all audio to my airport express connected speakers or can it only be done with itunes. At the moment i have my speakers connected to my airport

  • ABAP Coding issue

    Hi Abapers, I am new to ABAP, I am in the process of learning. I want to do IF KNVK-PAFKT = Z1 select single smtp_addr FROM adr6 INTO v_smtp_addr                           WHERE ADDRNUMBER =   v_adrnr. Endif. But i am able to get the EMail without IF