Analysis Service Database missing post reboot

Hi
Post recycle of Analysis Service we found one of the database went missing. So we tried attaching it from the local folder. But its failing with below error:
The detach log '\\?\E:\SalesCube_Monthly.0.db\SalesCube_Monthly.detach_log' could not be found in the specified database folder.
The database cannot be attached because an error occurred while loading the detach log from the file '\\?\E:\SalesCube_Monthly.0.db\SalesCube_Monthly.detach_log'. One possible reason is that database is already attached in ReadWrite mode to another server instance.
As we did not detach the database, this file detach_log did not get created. Is there any other way, we can attach or recover this database ?

Hi kalisubbu1,
Please follow the steps suggested in this related thread
http://social.msdn.microsoft.com/Forums/en-US/sqlanalysisservices/thread/231ba420-d8da-4945-b1e7-afc185128bb4  to see if it works for you.
You could try stopping the SSAS service on the new server, copying both the <database>.db folder and the <database>.db.xml file to the data folder on the new server and then restarting the SSAS service.
This should work as long as this database is not what is causing the other server not to be able to start and that you do not have any remote partitions or data stored outside of the data folder.
Regards,
Jerry
TechNet Subscriber Supportin forum
If you have any feedback on our support, please contact
 [email protected]

Similar Messages

  • Backup Analysis Services Database

    I am having an issue backing up SQL Server Analysis services Databases.  I have seen some threads on ensuring SQLServerAgent service account has write access to the folder containing the backup, be a sysadmin in SQL Server Instance, and be a server
    administrator in SSAS.  All are currently set.  An receive access Denied when backing up the database via SSMS, or via powershell using SMO/AMO.
    One thing that is odd, I can Backup the SSAS Databases to the folder containing the OLAP Logs without any problems, and the permissions are even more restrictive.
    I have other servers where we backup the SSAS databases just fine, and the Folder permissions, Database and SSAS are the same.  It's just 2 or 3 servers that we are having an issue, and we don't understand why.
    In the grand scheme of things what permissions are required to successfully backup SSAS Databases?
    SQL Server Instance Permissions
    SQL Server Analysis Services Permissions
    File and Folder Permissions
    Any other type of permission not mentioned above
    Any information would be appreciated.
    DJ

    Hi Donsjones,
    Thank you for your question. 
    I am trying to involve someone more familiar with this topic for a further look at this issue. Sometime delay might be expected from the job transferring. Your patience is greatly appreciated. 
    Thank you for your understanding and support.
    Regards,
    Charlie Liao
    TechNet Community Support

  • Duplicating Analysis Services Database

    Hi All,
    I am having an Analysis Services Database cube as DW_cube which i wanted to duplicate as DW_cube1 for some testing purpose..
    i am new to thing can anyone help me out.

    Hello,
    You can copy a database by backup and restore; see
    Backup and Restore of Analysis Services Databases
    Olaf Helper
    [ Blog] [ Xing] [ MVP]

  • Analysis Services(Database Attachment)

    Hai
     I have the database AdventureWorksDW2008R2(),I
    have successfully attached to the database engine but i am unable to attach to Analysis services.when i am trying to attach it the error is showing as
    "The
    folder ' "C:\softwares\downloads\AdventureWorksDW2008R2.zip" ' does not exist, is too long, or contains characters that are not valid or are reserved. "
    Please help me out of this error..
    Thanks
    after several trials i am getting the error like this even i have detached the database from other server instance

    Hi Lavvi,
    In your scenario, you have attached AdventureWorksDW2008R2 database to database engine without any problem. Then AdventureWorksDW2008R2 database is relational database. So we cannot attach a relational database to Analysis Services server.
    In your scenario, you can download AdventureWorks 2008R2 Analysis Services Project from the link below
    http://msftdbprodsamples.codeplex.com/releases/view/88252
    And then open it on SQL Server Business Intelligence Development Studio, then deploy this project to Analysis Services server. Then you will have a sample cube on your server.
    References.
    http://technet.microsoft.com/en-us/library/cc280583.aspx
    http://technet.microsoft.com/en-us/library/ms166576.aspx
    If I have anything misunderstood, please point it out.
    Regards,
    Charlie Liao
    TechNet Community Support

  • SQL Server Analysis Services (PowerPivot) instance returned the following error:

    SQL Server Analysis Services (PowerPivot) instance returned the following error: Error Code: 3241607174 Description: A connection corresponding to the embedded PowerPivot data was not found in the
    Excel workbook, file 'C:\Program Files\Microsoft SQL Server\MSAS10_50.POWERPIVOT\OLAP\Backup\Sandboxes\PowerPivotServiceApplication1\ReadOnlyExclusive-1096b00b-c487-46f9-afea-f717148863a8-Test_Sheet.xlsx'
    Hi Have Installed power pivot to and exsistinf farm standalone farm.
    What Should be the connection in the Excel fiile?
    Thanks
    Ravi
    Ravi

    Hi,
    According to your post, my understanding is that  you got an error with SQL Server Analysis Services.
    Did you have installed PowerPivot for Excel on a client machine and modeled a database as part of this workbook?
    The refresh in question is about refreshing the embedded Analysis Services database with data retrieved from the backend sources (defined when the database was originally modeled) and resaving the updated workbook. It should not be confused
    with Excel / Excel Services options to refresh the data connections in the workbook. That option is about refreshing data stored in Excel’s caches. With regards to PowerPivot, that would be refreshing from the embedded database, not the backend data from which
    the database was sourced. In summary, this feature will not work against a workbook which does not have an embedded database because there is no database to refresh. This is by design and should not be interpreted as indicating that there is something wrong
    with your system.
    More information:
    http://powerpivotgeek.com/2010/02/07/a-connection-corresponding-to-the-embedded-powerpivot-data-was-not-found-in-the-excel-workbook/
    There are some similar articles for your reference.
    http://technet.microsoft.com/en-us/library/ee210712(v=sql.105).aspx
    http://support.microsoft.com/kb/2761246
    http://social.msdn.microsoft.com/Forums/sqlserver/en-US/a324436c-9901-494e-9f9b-314399d65ccb/data-refresh-dont-work-for-serverhealthxlsx
    For the error “The data connection uses Windows Authentication and user credentials could not be delegated. The following connections failed to refresh”.
    There are multiple causes for this error message. The common factor behind all of them is that Excel Services cannot get a valid Windows user identity from a claims token in SharePoint.
    For more information:
    http://technet.microsoft.com/en-us/library/ff487975.aspx
    http://mmman.itgroove.net/2013/03/the-data-connection-uses-windows-authentication-and-user-credentials-could-not-be-delegated/
    Thanks & Regards,
    Jason
    Jason Guo
    TechNet Community Support

  • SSAS(SQl Server Analysis Services).

    Hello Sir,
    We have created an OLAP cube using Business Intelligence Development Studio(BIDS) with
    SQL Server Analysis Services(SSAS) using our data Source which is working fine & for reporting we are connecting the cube with
    MS-Excel through data Connections & we are able to generate the reports with the cube data by filtering & Customizing according to the requirement of reports within the same server instance. But when we want to deploy the cube onto
    the different server (basically on users server) to generate the respective reports we are facing problem. As the end user is not going to have access to the SQL Server Business Intelligence Development Studio how can we deploy the cube. As, we are using ODBC
    connections.
    As this is our first project through SSAS, we kindly need some support to proceed forward. 
    So, Please find the Queries related to the enclosed description as above.
    Queries related to SSAS Cube:
    What are the steps involved to deploy a OLAP cube created in one server to another server?
    Is there any third party tool required for it?
    Thanks in advance.
    Regards,
    Praveen.

    Hi Praveen,
    The following link tells you what you need to know; -
    http://technet.microsoft.com/en-us/library/ms166576(v=sql.105).aspx
    I confirm you do not need a third party tool to deploy a cube and empower your users to access this cube.
    So in summary; -
    Within your Analysis Services solution within BIDs; -
    1) Right click your mouse over the Analysis Services database icon, which is immediately below the solution icon.
    2) Select the sub tab "Deployment".
    3) Enter the name of the server you need to deploy to.
    4) Click "OK".
    5) Right click your mouse (again) over the Analysis Services database icon, which is immediately below the solution icon.
    6) Select "Deploy"
    Kieran Patrick Wood http://www.innovativebusinessintelligence.com http://uk.linkedin.com/in/kieranpatrickwood http://kieranwood.wordpress.com/

  • SQL SERVER 2008 R2 Analysis Services

    Hi 
    We have a SQL Server 2008 R2 Analysis Services and it was running fine but suddenly its not processing the cube. I have checked log the file msmdsrv.log and there its continuously logging the following error message.
    Message: The file '\\?\C:\Program Files\Microsoft SQL Server\MSAS10_50.MSSQLSERVER\OLAP\Data\TRCOLAP.0.db\Fact SL Financial Transaction.0.dim\466.SL Journal Type.kstore' could not be opened. Please check the file for permissions or see if other applications
    locked it. (Source: \\?\C:\Program Files\Microsoft SQL Server\MSAS10_50.MSSQLSERVER\OLAP\Log\msmdsrv.log, Type: 3, Category: 289, Event ID: 0xC1210023) 
    Would really appreciate help in this regard. It our production box
    Thanks
    inazakat

    Hi Inazakat,
    According to your description, your SQL Server Analysis Services database cannot be processed that can be processed without any problems before, right?
    Based on the message, file could not be opened. Please check the file for permissions or see if other applications locked it. It seems that the file is corrupt. This issue can be caused by that the file was deleted by antivirus. You can go to the corresponding
    folder and check if this file exist or not. If this file was deleted or corrupt, you can open the original project on BIDS, and redeploy it to your server. And then process this database.
    Regards,
    Charlie Liao
    TechNet Community Support

  • SSIS Analysis Services Processing Task - which object is being processed?

    Hi
    I've just fallen down the hole dug by a gotcha in the SSIS Analysis Services Processing task (let's call it the ASPT as it's too long to type).
    I was doing testing yesterday, targeting a test database.  The way things are set up here is that we have two structurally-identical databases (let's call them SSASLive and SSASTest) on the same SSAS server.
    The SSIS package includes an Analysis Services Processing task that processes a partition of a measure group in a cube.
    So I changed the properties of the AS Connection Manager in the package to point to database SSASTest.  Then, because I'm paranoid, and don't trust the Editor for the ASPT, I edited the object to be processed, by deleting the existing object from the
    list and re-adding it (going down the tree: database SSASTest, cube TheCube, measure group TheMassiveMeasureGroup, partition CurrentMonth).
    This is an annoying necessity, because: the ASPT editor doesn't allow you to see exactly what the object(s) in the list to be processed are, by e.g. double-clicking on them.  All you can see is the lowest-level object name, which could be just
    "Current Month".  Current Month what?  There's no way of telling.
    Test worked fine.  I made sure the AS Connection manager was reset to point to SSASLive, and deployed the package.
    This morning, the package ran, and processed the partition in database
    SSASTest.  What happened?  Looking in the Code view of the package, I found the XMLA behind the Processing task: the database name is literally specified in the XMLA, and completely ignores the Catalog of the AS connection.  This information
    is not exposed anywhere in the UI view of the package.
    It's a bit like the dangerous annoyance when testing SSIS package Exec SQL tasks, where a previous developer has insisted on fully-qualifying object names: SELECT * FROM LiveDatabase.dbo.ATable.  You're working with a connection to a test database,
    but it makes not difference...
    But at least that problem is clearly visible.  I know that specifying the database name is normal in XMLA tasks: but at least the SSIS editor should allow you to see what the setting is.

    No, the point is that the Connection Manager's Initial Catalog setting (whether hard-coded or set by any of the configuration methods) has no effect on the database addressed by the Analysis Services Processing task.
    Steps to reproduce:
    1. Create an Analysis Services database containing a cube (or select an existing one): let's call it ASDB1.
    2. Make a copy of it on the same server, e.g. by backing up and restoring: let's call it ASDB_Copy
    2. Set up an SSIS package with a Connection Manager pointing to ASDB1
    3. Create an Analysis Services Processing task, using the Connection Manager, processing any particular SSAS object.
    4. Run the package.  The SSAS object in ASDB1 is processed (not the object in ASDB_Copy).  (Confirm by checking the Last Processed date in SSMS, Properties).
    5. Change the Connection Manager's Initial Catalog setting to point to ASDB_Copy.
    6. Run the package.  The SSAS object in ASDB1 is processed, not the object in ASDB_Copy.  The Connection Manager's Initial Catalog setting has no effect on the AS Processing Task.  Examining the Code view of the package confirms that the XMLA
    behind the task specifies the database as well as the AS object. (and this database setting is still ASDB1).
    The fix would be for SSIS to perform validation of AS objects referenced by the AS Processing Task whenever its Connection manager is pointed to a different database, and to update the XMLA accordingly.
    Also, it would be useful to be able to see exactly what the objects listed in the Processing Task are.  The information shown is just not sufficient to identify the object to the developer.  I may have dozens of measure groups with a partition
    called Current Month - but "Current Month" is all I can see in the editor, and there's no way to e.g. double-click on an item in the list and get the full object identification (Cube X, measure group Y, partition Current Month).  Sure,
    I can go and look at the XMLA, but that's fiddly: if there's going to be a wrapper around the XMLA (which is what the AS Processing Task is), then it should be unambiguous.

  • "ApShell_BPU" database in Microsoft Analysis Services

    In the BPC 5.1 Sinlge server install i am not seeing "ApShell_BPU" database in Microsoft Analysis Services, what can be the reason? Have i missed something?
    Please reply me back.
    Thanks and regards
    Abhay

    After a fresh installation of SAP BPC you will have two databases in SQL, APSHELL (storing the real application) and APPSERVER (storing server defaults). In OLAP (MS analysis services) you will have one database called APSHELL.
    The name of the OLAP database will always be exactly the same as the one in SQL.
    So you should have an APSHELL db in AS, not APSHELL_BPU.
    Hope this helps,
    Alwin

  • MS SQL database needs to run mdx query to analysis services to retrieve data from a BPC 10 Netweaver cube(view only)

    We are in the process of migrating from BPC7 SP12 Microsoft sql server to BPC10 netweaver on a sql server with BW7.4 and need to integrate our home grown data warehouse which is on a Microsoft sql server.  The data warehouse currently connects to BPC7 using integration services/analysis services and runs mdx queries to analysis services to retrieve data from the BPC7 cube (view only).  Please provide documentation on how to create this same integration with our data warehouse using BPC10 netweaver on a sql server.

    When you were setting up your ODBC data source for
    the Text driver, did you click on the "Options"
    button? There's a lot of options in there that I
    didn't understand at first glance.Yes I clicked on the options button, but the only thing there is dealing with file extensions. I left it set to the default.
    I have since tried closing my connection after I insert a record, I then try executeQuery, but still no luck. Seems strange that I can write to the file but not read from it. If any thing I'd expect the opposite problem.
    I have also tried using the class "JoltReport" from the sun tutorial instead of my own with the same result.
    Message was edited by:
    Hentay

  • Missing "Windows Service - SQL Server Analysis Services"

    Hi
    In my SharePoint 2013 Central Administration (this is a Project Server 2013 deployment), I got an alert for this error:
    PowerPivot: MidTier process account should have 'Full Read' permission on all associated SPWebApplications. 
    After reading
    this thread, I assumed I have the 2 accounts configured the wrong way, so I check
    this article to see how to change the accounts, but on the section
    "Change the account under which each service runs" it says to select
    "Windows Service - SQL Server Analysis Services", which I can't find like as you can see in the screenshot below. How can I solve this? I'm new to this all, so if you could simplify the explanation I would be thankful, if not, I'll
    just work on it anyway. :) Thanks!

    Hi ,
    In the process of configuring PowerPivot for SharePoint 2010, automatically select the Analysis Services for SharePoint Integration feature. After configuring PowerPivot for SharePoint 2010, we can see
    SQL Server Analysis Services and SQL Server PowerPivot System Service are started on manage services on server. So you can verify whether analysis service is installed for SharePoint on manage
    services on server.
    More information, please refer to the links:
    http://sharepointgeorge.com/2012/installing-sql-server-powerpivot-sharepointstep-step-guide/
    http://technet.microsoft.com/en-us/library/855bd055-5ad3-493f-9c5b-1f5297b2e6e2(v=sql.110)
    As Trevor said, for this issue, you can ignore it. Anyway, you can also check the link:
    http://whitepages.unlimitedviz.com/2012/06/the-health-analyzer-and-powerpivot-for-sharepoint/
    I hope this helps.
    Thanks,
    Wendy
    Wendy Li
    TechNet Community Support

  • [Forum FAQ] How do I create calculated measure using AMO in SQL Server Analysis Services?

    Introduction
    In SQL Server Analysis Services (SSAS), you can create a calculated measure in SQL Server Data Tool (SSDT)/Boniness Integrated Development Studio (BIDS). Sometimes you may need to create calculated measure by using AMO in a C# or VB project.
    In this article, I will demonstrate so how to create calculated measure using AMO in SSAS?
    Prerequisites
    Before create calculated measure using AMO, you need to ensure that the following components were installed in your server.
    The multidimensional database AdventureWorks Multidimensional Model 2012
    A SQL Server with SSIS and SSAS installed
    The AMO libraries installed:
    X86 Package (SQL_AS_AMO.msi)
    X64 Package (SQL_AS_AMO.msi)
    Solution
    Here is the detail steps to create calculated measure using AMO in SSAS.
    Open SSDT and create a new SSIS project.
    Drag Script Task to the design surface.
    Click SSIS-> Variables to open the Variables window and add two variables that used to connect to the server and database.
    Create a connection to connect to SSAS server.
    Rename the connection name to ssas.
    Double click the Script Task to open Script Task Editor.
    Add Connection and Database variables to ReadWriteVariables textbox and then click Edit Script button.
    Add AMO reference in the Solution Explore window.
    Copy the script below and paste it into the script.
    Dim objServer As Server
    Dim objDatabase As Database
    Dim strDataBaseID As String
    Dim objCube As Cube
    Dim objMdxScript As MdxScript
    Dim objCommand As Command
    Dim strCommand As String
    objServer = New Server
    objServer.Connect("localhost")
    objDatabase = objServer.Databases("AdventureWorksDW2012Multidimensional-EE2")
    strDataBaseID = objDatabase.ID
    If objDatabase.Cubes.Count > 0 Then
    objCube = objDatabase.Cubes("Adventure Works")
    If objCube.MdxScripts.Count > 0 Then
    objMdxScript = objCube.MdxScripts("MdxScript")
    objMdxScript = objCube.MdxScripts(0)
    Else
    objCube.MdxScripts.Add("MdxScript", "MdxScript")
    objMdxScript = objCube.MdxScripts("MdxScript")
    End If
    objCommand = New Command
    strCommand = "CREATE MEMBER CURRENTCUBE.[Measures].[Multipy Measures By 3]"
    strCommand = strCommand & " AS [Measures].[Internet Sales Amount] * 3, "
    strCommand = strCommand & " VISIBLE = 1 ; "
    objCommand.Text = strCommand
    objMdxScript.Commands.Add(objCommand)
    objMdxScript.Update()
    objCube.Update()
    End If
    objServer.Disconnect()
    Then you can run this SSIS package to create the calculated measure.
    Applies to
    Microsoft SQL Server 2005
    Microsoft SQL Server 2008
    Microsoft SQL Server 2008 R2
    Microsoft SQL Server 2012
    Please click to vote if the post helps you. This can be beneficial to other community members reading the thread.

    Thanks,
    Is this a supported scenario, or does it use unsupported features?
    For example, can we call exec [ReportServer].dbo.AddEvent @EventType='TimedSubscription', @EventData='b64ce7ec-d598-45cd-bbc2-ea202e0c129d'
    in a supported way?
    Thanks! Josh

  • Creating OLAP report with Crystal Reports and SQL Server Analysis Services 2005

    Post Author: orahc_mao
    CA Forum: Data Connectivity and SQL
    Hi!
    I am currently using the trial version of the Crystal Reports XI and I wanted to do an OLAP report. The problem is I cannot select a cube using "OLAP Connection Browser" (the popup window). I already selected Microsoft OLE DB Provider and entered the server name but still can't connect.
    I don't think the problem is with SQL Server Analysis Services (2005) since Microsoft Excel - Import Data can detect the server as well as the cube I have created.
    I also tried the "Database Expert" of Crystal Reports, created an OLE DB (ADO) connection with "OLE DB Provider for OLAP Services 8.0" driver, entered the server name, checked the integrated security and then I can readily select the database created at SQL Server Analysis Services. However, I still need the OLAP grid create the report but it goes back to my original problem... selecting a cube.
    I hope somebody would help me with this issue.
    Thanks a lot!
    orahc_mao

    Hello,
    I don't recognize those variables as CR ones so it's likely something the original developer is setting in code.
    You'll have to discuss with that person.
    If your have SDK issues then post your question to one of the .NET or Java SDK forums.
    Thank you
    Don

  • Analysis Services Execute DDL Task Internal error

    Hi all,
    I need help in solving this sporadic problem...  the dimension below (blanked out as xxxx) is based on a view.  The cube processes fine many times - but fails - abruptly once a week... it's being called from a sql job.
    Any ideas?  Thanks in advance!!
    Error: 2010-06-28 15:45:17.69     Code: 0xC1000007     Source: xxxxxxx Analysis Services Execute DDL Task     Description: Internal error: The operation terminated unsuccessfully.  End Error 
    Error: 2010-06-28 15:45:17.69     Code: 0xC11F000D     Source: xxxxxxx Analysis Services Execute DDL Task     Description: Errors in the OLAP storage engine: An error occurred while the 'xxxxx' attribute
    of the 'xxxxx' dimension from the 'xxxxxx' database was being processed.  End Error  Error: 2010-06-28 15:45:17.69     Code: 0xC11F0006     Source: xxxxxxx Analysis Services Execute DDL Task    
    Description: Errors in the OLAP storage engine: The process operation ended because the number of errors encountered during processing reached the defined limit of allowable errors for the operation.  End Error  Error: 2010-06-28 15:45:17.69    
    Code: 0xC11C0002     Source: xxxxxx Analysis Services Execute DDL Task     Description: Server: The operation has been cancelled.  End Error  DTExec: The package execution returned DTSER_FAILURE (1). 
    Started:  3:45:00 PM  Finished: 3:45:17 PM  Elapsed:  17.328 seconds.  The package execution failed.  The step failed.
    Harsh B

    (From http://msdn.microsoft.com/en-us/library/cc966526.aspx)
    "ExternalCommandTimeout is a server property that is used to set the number of seconds that SSAS should wait to time out when issuing commands to external data sources, such as relational and other OLAP sources."
    "ExternalConnectionTimeout is a server property that is used to set the number of seconds, by default, that SSAS should wait to time out when connecting to external data sources, such as relational and other OLAP sources."
    What you set it to, it's absolutely up to you... But as I see it in your log, your task failed in 17 and 9 seconds... So I think you can multiply the values by 10, I guess the situation won't change... But it is worth to check it at least.
    So... How often your fact table gets new (degenerate) dimension reference values? Is it a process update, or a process full command? I really would like to see it :)
    -- Zoltán Horváth
    -- MCITP SQL Server Business Intelligence Developer 2005, 2008
    -- Please mark posts as answered or helpful where appropriate.

  • Is SQL Analysis Services black-holing my server's memory?

    I'm running a SQL 2008 SP2 database engine on a server that also hosts analysis services and reporting services.  The server has 32GB of memory.  When I run performance monitor against it, I can see only 2.5GB in use:  see the included screen
    shot.  Yet, the server reports that out of that 32GB, only 208 MB are free.
    How can I discover which other process or 'gremlin' is sucking the missing 30 GB of memory?

    How can I discover which other process or 'gremlin' is sucking the missing 30 GB of memory?
    Then check the other processes.
    And yes, SQL Server and SSAS have a different strategy in utilizing the memory (RAM) and depending on the current workload it can happen, that e.g. SSAS requires the memory used by SQL Server and this can cause performance problem. If you use SQL Server
    as data warehouse for SSAS, then you will see this effect extremely during cube processing.
    Olaf Helper
    [ Blog] [ Xing] [ MVP]

Maybe you are looking for

  • 10.7.5 update has deleted my Entrourage email account.

    I have just completed the 10.7.5 update and when the computer restarted post-update, I found that my entire Entourage email account has been deleted. No emails, no saved folders, no inbox and I can't receive anything unless I set up a new account. Af

  • Playback on

    Ok well I have a Creative Labs Nomad Jukebox Zen Xtra with 40GB of space. Now I know you can put all the songs that are on the MP3 player onto your PC but what about just playing the songs from your MP3 to your pc? I have it all hooked up to my USB p

  • Sqlldr encounters ORA-01801 when NLS_LANG=.UTF8

    Hi, When I set NLS_LANG=.UTF8 and then run SQL Loader, I get the following error: ORA-01801: date format is too long for internal buffer My CTL file characterset is UTF8 and the content in the DAT file is also in UTF8 format. In addition, the table s

  • Outlook Connector Without Mail Server

    Installed Java Connector (v 7.2.310.1) for MS Outlook 2002 with SP3. I have different usernames and passwords for the mail server and the calendar server. Used fake server name for the mail server in the deployment program, and kept getting the 'Unab

  • Fan like noise - Macbook Pro Retina 13 inch - late 2013 model

    If I open Frontline Commando (game app) my brandnew Mackbook pro make fan like noise. (annoying and scary) Could anyone advice please? my sys configuration: Model - Macbook Pro Retina 13 inch - late 2013 model RAM - 8 GB Processor - 2.4 GHz intel cor