Creating SSAS cube on top of a view

Hi,
I have a view which doesn't have ID columns. I want to create an SSAS cube on top of that view. Can someone please suggest how this can be achieved?

Hello,
You need also dimensions in order to create cube.So you need to figure out what are your dimensions.There is a article below by Microsoft.You can refer it.Hope this helps.
http://technet.microsoft.com/en-us/magazine/ee677579.aspx

Similar Messages

  • How to create ssas cube by using MySQL Database

    please tell me step by step process to create ssas cube by using MySQL database in my system, i have Sqlserver 2008 enterprise edition and MySql5.0

    There is an OLEDB provider for MySQL which you can get from here
    https://cherrycitysoftware.com/ccs/Providers/ProvMySQL.aspx otherwise you can also use SSIS to push the data straight into the Analysis Services database without needing to stage it in SQL Server. Also, as you can load data into AS using XMLA you could
    also write your own loader to extract data using ODBC and push it into AS using XMLA, essentially what I suspect SSIS does. However those latter solutions don't allow you to create a database on top of MySQL because you need an OLEDB (or .net) provider for
    that.
    In the simplest case, install the OLEDB provider and then in AS create a Data Source connection using that provider. Once you have done that you should be able to create a Data Source View using that connection enabling you to import the schema definitions
    for the tables/views in the MySQL database. From there you build dimensions and cubes etc. about which there is plenty of information on the web.
    http://bi-logger.blogspot.com/

  • How to create ADF BC components like EO from  "View with INSTEAD OF trigger

    I have a "View with INSTEAD OF trigger" on a external schema. is it possible to create ADF EO on top of this view in my local schema?. If possible, then is it possible to insert/update that external table using ADF standard data controls and Application module?. I'm trying to see if it's possible with standard ADF controls without calling pl/sql API to insert/update that external table. any ideas are appreciated.
    Regards,
    Surya

    http://stegemanoracle.wordpress.com/2006/03/15/using-updatable-views-with-adf/

  • Query accessing SSAS cube is taking more time in power view report

    Hi,
    My SSAS server is having 16 Core's . I have created a power view in sharepoint site, which is accessing ssas cube.
    The report is taking more time to show the result.  I traced the DAX query using SQL Server profiler. When I ran this DAX query in SSAS server directly, then its taking less time but when the same query running from sharepoint server (power view) , it
    is taking more time. Number of records in the fact table is 300M . 
    Also i tracked the %process time in performance monitor, mostly the line is below 20%. I think the CPU is not fully utilized..
    Could you please help me to improve the performance of my report.
    Regards,
    Arun

    Hi Arun,
    According to your description, you create a PowerView report in SharePoint site connect to SSAS cube, the problem is that it take long time to show the result in the report, now you want to improve report performance, right?
    In your scenario, you said that it takes more time to return the result in PowerView report than run the query directly in SSMS. For a report, its run time contain retrieval data time and render report time. So it takes more time to return the result in
    PowerView report than run the query directly in SSMS. And there are 300M records in the fact table, the performance can be caused by large data. Here is a blog which describes tracks down Power View performance problems.
    http://blogs.msdn.com/b/psssql/archive/2013/07/29/tracking-down-power-view-performance-problems.aspx
    Regards,
    Charlie Liao
    TechNet Community Support

  • Performance issue in browsing SSAS cube using Excel for first time after cube refresh

    Hello Group Members,
    This is a continuation of my earlier blog question -
    https://social.msdn.microsoft.com/Forums/en-US/a1e424a2-f102-4165-a597-f464cf03ebb5/cache-and-performance-issue-in-browsing-ssas-cube-using-excel-for-first-time?forum=sqlanalysisservices
    As that thread is marked as answer, but my issue is not resolved, I am creating a new thread.
    I am facing a cache and performance issue for the first time when I try to open a SSAS cube connection using Excel (using Data tab  -> From Other Sources --> From Analysis Services) after daily cube refresh. In end users system (8 GB RAM but around
    4GB available RAM), for the first time, it takes 10 minutes to open the cube. From next run onwards, its open up quickly within 10 secs.
    We have daily ETL process running in high end servers. The configuration of dedicated SSAS cube server is 8 core, 64GB RAM. In total we have 4 cube DB - out of which for 3 is full cube refresh and 1 is incremental refresh. We have seen after daily cube
    refresh, it takes 10 odd minutes to open the cube in end users system. From next time onwards, it opens up really fast with 10 secs. After cube refresh, in server systems (32 GB RAM, around 4GB available RAM), it takes 2 odd minutes to open the cube.
    Is there, any way we could reduce the time taken for first attempt ?
    As mentioned in my previous thread, we have already implemented a cube wraming cache. But, there is no improvement.
    Currently, the cumulative size of the all 4 cube DB are more than 9 GB in Production and each cube DB having 4 individual cubes in average with highest cube DB size is 3.5 GB. Now, the question is how excel works with SSAS cube after
    daily cube refresh?
    Is it Excel creates a cache of the schema and data after each time cube is refreshed and in doing so it need to download the cube schema in Excel's memory? Now to download the the schema and data of each cube database from server to client, it will take
    a significant time based on the bandwidth of the network and connection.
    Is it anyway dependent to client system RAM ? Today the bigest cube DB size is 3.5 GB, tomorrow it will be 5-6 GB. Now, though client system RAM is 8 GB, the available or free RAM would be around 4 GB. So, what will happen then ?
    Best Regards, Arka Mitra.

    Could you run the following two DMV queries filling in the name of the cube you're connecting to. Then please post back the row count returned from each of them (by copying them into Excel and counting the rows).
    I want to see if this is an issue I've run across before with thousands of dimension attributes and MDSCHEMA_CUBES performance.
    select [HIERARCHY_UNIQUE_NAME]
    from $system.mdschema_hierarchies
    where CUBE_NAME = 'YourCubeName'
    select [LEVEL_UNIQUE_NAME]
    from $system.mdschema_levels
    where CUBE_NAME = 'YourCubeName'
    Also, what version of Analysis Services is it? If you connect Object Explorer in Management Studio to SSAS, what's the exact version number it says on the top server node?
    http://artisconsulting.com/Blogs/GregGalloway

  • Powerview modifying a calculated measure in SSAS cube

    We are running SQL Server 2012 and Sharepoint 2013 we are having problems with Powerview modifying a calculated measure in the underlying multidimensional cube. We have a powerview report with multiple pages that allows a user to utilize a global filter
    to view the financial performance of particular customers on multiple reports/pages. When the user picks a different customer, one of the calculated measures, in this case gross margin percent, is essentially stuck. It does not change when the customer selection
    changes. It remains the gross margin percent as the previous customer. All of the other measures change appropriately. Measures such as sales, expenses, ebitda and the other measures that are columns in the relational sql server db feeding the cube are fine.
    Only the calculated measures defined in ssas are not working. I then tested the molap cube on a clean connection to a new excel sheet and was shocked to find that the same value the power view report is stuck on is also displayed in the excel file!  It
    has essentially modified the data in the cube!  The only way to clear the value and display the correct calculation is to re deploy the cube. Any one here know what to do to resolve this?

    Hi Osu75,
    According to your description, you create a PowerView report using SSAS cube as the data source in SharePoint site, the problem is that when changing the selection in the global filter, it change the calculated measure in the underlying multidimensional
    cube, right?
    In your scenario, which script are you using when creating this calculated measure?  Generally, report is used to display data from underlying database, it's really strange that the PowerView report change the calculated measure in the underlying multidimensional
    cube.
    From a support perspective this is really beyond what we can do here in the forums. If you cannot determine your answer here or on your own, consider opening a support case with Microsoft. Visit this link to see the various support options that are available
    to better meet your needs:
    http://support.microsoft.com/default.aspx?id=fh;en-us;offerprophone
    Regards,
    Charlie Liao
    If you have any feedback on our support, please click
    here.
    Charlie Liao
    TechNet Community Support

  • SSRS Expression Current month in a Fiscal calendar dimension from a SSAS cube

    How to create an expression for a default SSRS parameter to a SSAS cube when using i Fiscal Year calendar - and where I want to show current Calendar month and year (for example February 2014) 
    This is an example of my Time dimension for February 2014:
    [Time].[FiscalYMD].[Fiscal_Year].&[2013].&[10]
    The fiscal calendar goes from May - May - which means that April 2014
    would look like this:
    [Time].[FiscalYMD].[Fiscal_Year].&[2013].&[12]
    and May 2014 would look like this:
    [Time].[FiscalYMD].[Fiscal_Year].&[2014].&[01]
    Please advise

    Hi HCMJ,
    To represent the fiscal calendar, you shift the current date by 4 months back. Then you can do:
    ="[Time].[FiscalYMD].[Fiscal_Year].&[" & Year( DateAdd( "m", -4, Today() ) ) & "].&[" & Month( DateAdd( "m", -4, Today() ) ) & "]"
    It just gets the year and month of '4 months ago'. I have not tested it, but I guess it should work.
    Regards
    Andrew Borg Cardona

  • How to Create a Cube Based on other Cubes?

    I want to create a cube that is sourced from other cubes. My situation is I have a number of cubes based on individual products and I want to create a cross-product cube that combines the individual product cubes at a common dimensional grain.
    e.g :
    CUBE_1 (DIM_1, DIM_2, DIM_3)
    CUBE_2 (DIM_1, DIM_2, DIM_3, DIM_4)
    CUBE_3 (DIM_1, DIM_2, DIM_3, DIM_4, DIM_5)
    CROSS_CUBE (DIM_1, DIM_2, DIM_3)
    I see if your cubes are in the same AW then you can access measures from other cubes with a calculated measure. But I'm struggling to understand two things:
    1. If I want to build a cube that is only sourced from other cubes and not any database tables then what do I put in the cube mapping?
    2. If I create a calculated measure that accesses another cube's measure does Oracle OLAP automatically match up the dimensionality? e.g. If I create a calculated measure in CROSS_CUBE that adds together a measure from CUBE_1 and a measure from CUBE_2 then will the new measure return the correct results across  DIM_1, DIM_2 and DIM_3?
    To do this in a relational star schema you have to do a Drill-Across: http://www.kimballgroup.com/2005/06/design-tip-68-simple-drill-across-in-sql/
    Is what I'm trying to do possible or will I need to combine the cube views into another view and use this as my cross cube source?
    Thanks in advance.
    OLAP Analytic Workspace 11.2.0.4.0
    Oracle OLAP API 11.2.0.4.0
    OLAP Catalog 11.2.0.4.0

    You can create a cube in AWM which consists only of calculated measures.  These calculated measures can reference other cubes provided that
    1) the cubes are in the same Analytic workspace as the calculated measure cube and
    2) the calculated measure cube is dimensioned by the same dimensions as all the cube measures it references in its calculated measures.
    For instance, if your calculated cube is dimensioned by dim1, dim2, dim3 and dim4, it can have calculated measures which refer to measures in cubes that are dimensioned by any subset of the dimensions (i.e. dim1 and dim3, or dim1, dim2, dim3 and dim4, etc).
    The calculated measure cube does not need to be mapped because it only contains calculated measures that reference measures in other cubes which I am assuming are mapped and built. 
    --Ken

  • SSAS cubes - fragments

    Hi,
    Hoping someone can help. I have defined a logical model over the top of a set of SSAS cubes in a single SSAS database. This consists of a summary cube and detailed cubes that relate to individual products. I would like to have a "Product" dimension that will allow the user to drill between the products and for dimensions at lower levels of granularity (i.e. that are only relevant to one product) data should be sourced from the relevant detailed cubes. Unfortunately although the summary cube has a product dimension the detailed cubes don't - when I try to map the detailed cube to the product dimension using an expression I get the following error message: "Target data source does not support the Complex Expression operation."
    e.g.
    Product Dimension has 3 products in it: Bikes, Cars and Boats. There is a summary cube that contains all 3 products and has the Product dimension defined.
    There are 3 detailed cubes for Bikes, Cars and Boats that don't have a product dimension. In my mappings for the Product Name attribute of the Product dimension I have the following:
    Logical Table Source..... Mapped As
    SummaryCube..... Product.Product
    BikeCube... 'Bikes'
    CarCube... 'Cars'
    BoatCube.. 'Boats'
    Hope this explains it. I get the error message whenever I'm trying to get data from the detailed level cube.
    Thanks,
    Rob

    Hi Platt,
    By default, the default account for AS services 'NT Service\MSSQLServerOLAPService' has only public permission to few databases. If you haven't change the AS services account, or haven't grant the permission for this account, then you cannot deploy
    the AS project to server.
    So in your scenario, ensure the service "SQL Server Analysis Services" is running with a proper Log On As user. This user should have administrator permission on Analysis Services and the proper permissions on the database engine. To avoid this
    issue, you can grant the SELECT permission for the default account for AS services to data source DB. Or you can change the services account to this domain user who has the corresponding to the data source DB.
    http://jetreportskb.host4kb.com/article/AA-00238
    http://msdn.microsoft.com/en-in/library/ms175371.aspx
    Regards,
    Charlie Liao
    TechNet Community Support

  • Using FORMAT_STRING from SSAS cube - not working consistently

    Here's the deal.  
    I've got an SSAS cube and I am scoping the format of the measures based on dimension members. 
    Like this.
    scope(( [Account].[Level 5].&[I000900000] )); 
    format_string(This) = "(#,0,);#,0,";   
    end scope; 
    I'm then calling this format in the SSRS report using typical SSRS trickery and hacks.  
    Well. When I run the report for one date (this is a GL financial cube), this works perfectly.  When I run the report for a different date, it all of a sudden doesn't work.  When I view the data in excel through a pivot table, everything is fine.
    It's driving me crazy.  Has anyone else experienced something like this.

    Hi Baracus,
    Thank you for your question. 
    I am trying to involve someone more familiar with this topic for a further look at this issue. Sometime delay might be expected from the job transferring. Your patience is greatly appreciated. 
    Thank you for your understanding and support.
    Regards,
    Charlie Liao
    TechNet Community Support

  • SSRS Parameter Drop downs are greyed out after changes to SSAS cube

    Hi
    I am facing a weird problem. I have 3 SSRS reports that are based on an SSAS cube. Few days back I deployed some changes to cube and everything works ok on dev environment. When I deployed the same changes to production, SSRS reports don't work. I don’t
    get any error or anything when I click open the report but all the parameter boxes are greyed out. If I restore SSAS database from dev to production then SSRS reports work fine and I can select parameter drop downs. What could possibly wrong here. I have no
    idea to what to troubleshoot here.
    Thanks
    Nik

    Hi Nik,
    Are you using render the report from a report snapshot? In contrast with reports that run on demand, it is not possible to specify a different parameter value for a report snapshot when the report is open. So the parameter textbox is grayed out. If the users
    want to view the report with different parameter, we can select “Always run this report with the most recent data” option in “Processing Options properties” page.
    Besides, if the issue is not in this scenario, we can try to edit the report with Report Builder to check whether the report is ok. Such as delete the parameters, then recreate them to check again.
    If there are any misunderstanding, please elaborate the issue for further investigation.
    Thanks,
    Katherine Xiong
    Katherine Xiong
    TechNet Community Support

  • Slow parameter from SSAS cube

    Hi,
    I've got a SSRS report which use a SSAS cube as source. In the report there are 5 parameters which are build in the Report builder 3.0 using the designer.  The report server is a SQL server 2008 R2 with the latest updates.
    The issue that I am having is when running the report if I would like to choose another value from one of my parameters then I open the parameters (drop-downbox)and select another value in the box - then it's like the drop-down box hangs for 3-4
    seconds when I click out of the box before the value is appearing in the report, when I use the Report builder Tool. If I do the same in IE then the reports render for 10-15 seconds each time a new value is choosen.
    Any Idea how to improve this?
    The Parameter consist of only 5-6 values and it performs very fast if I execute the Query in the designer
    I've attached the Query as it looks like for the parameter(hidden dataset):
    WITH MEMBER [Measures].[ParameterCaption] AS [Company].[Company Name].CURRENTMEMBER.MEMBER_CAPTION MEMBER [Measures].[ParameterValue] AS [Company].[Company Name].CURRENTMEMBER.UNIQUENAME MEMBER [Measures].[ParameterLevel] AS [Company].[Company Name].CURRENTMEMBER.LEVEL.ORDINAL
    SELECT {[Measures].[ParameterCaption], [Measures].[ParameterValue], [Measures].[ParameterLevel]} ON COLUMNS , [Company].[Company Name].ALLMEMBERS ON ROWS FROM ( SELECT ( STRTOSET(@ReportingCurrencyReportingCurrency, CONSTRAINED) ) ON COLUMNS FROM [SALES])
    (the issue is the same for all my parameters)

    Hi HCMJ,
    In your scenario, you said that "the reports render for 10-15 seconds each time a new value is choosen".
    Report server performance is affected by a combination of factors that include hardware, number of concurrent users accessing reports, the amount of data in a report, and output format.
    To find out how long it takes to process a report and which process phase take more time, we can review the report server execution log. We can use the report execution log to find out how often a report is requested, what formats are used the most, and
    how many milliseconds of processing time is spent on each processing phase.
    In addtional, i'd suggest you enable SQL Sever profiler to monitor the queries fired by the report, once you find some queries took a very long time to run, consider creating the smaller cube partition or optimzing the query by adding index or
    partition to improve the query performance.
    Regards,
    Charlie Liao
    TechNet Community Support

  • Creating a cube using OEM

    I have an Oracle 10g database that has 5 tables in it. these tables are linked together through keys. Now, I am curious to know if I can create a cube with two or more dimensions based on columns from the same table, and also have the measure(s) for that cube be from the same table?? Would I need to create a view of that table and use that view as the "fact table"?
    It seems like the cubes built in a lot of these samples use dimensions where each of those dimensions each come form a separate table. Does it always have to be that way?
    Thanks for any input.

    I'm not sure exactly why you want to do this.
    I guess I didn't explain very clearly :)
    We (i.e. the company I work for) creates software (in C++) which needs to work with a lot of data which is in an Oracle-DB.
    As the analysis done by the SW would profit (performance-wise) from the layout of the data in an cube, we want to get the data in a cube.
    But information like the number of dimensions isn't known beforehand (as that's sth. that's dependent on the site the customer uses the SW on, and can
    change with time), the software has to do the creation of dims and the cube itself.
    (For the curious ones: http://www.centerpoint.eu.com/ )
    And as OLAPI isn't available for C++ (at least afaics), I'd really want to do this in PL/SQL.
    If you do use standard form, you then allow yourself the
    option of using the various tools later
    That's really not needed :)
    You can always build the original AW using
    AWM, then save the build script it generates, and use
    that to do subsequent builds.
    Is that build-script in OLAP DML? How do I save it?
    BTW, in addition to the SPL INFILE command, there is
    the DBMS_AW.INFILE plsql routine, which saves you
    some quotes :-).
    dbms_aw.execute('infile ''work_dir/my.inf''') vs
    dbms_aw.infile('work_dir/my.inf')
    I hate superfluous quotes :)
    Thanks Jim!

  • Cache and performance issue in browsing SSAS cube using Excel for first time

    Hello Group Members,
    I am facing a cache and performance issue for the first time, when I try to open a SSAS cube connection using Excel (using Data tab  -> From Other Sources --> From Analysis Services) after daily cube refresh. In end users
    system (8 GB RAM), for the first time, it takes 10 minutes to open the cube. From next run onwards, its open up quickly within 10 secs.
    We have daily ETL process running in high end servers. The configuration of dedicated SSAS cube server is 8 core, 64GB RAM. In total we have 4 cubes - out of which for 3 is full cube refresh and 1 is incremental refresh. We have seen after
    daily cube refresh, it takes 10 odd minutes to open the cube in end users system. From next time onwards, it opens up really fast with 10 secs. After cube refresh, in server systems (16 GB RAM), it takes 2 odd minutes to open the cube.
    Is there, any way we could reduce the time taken for first attempt ?
    Best Regards, Arka Mitra.

    Thanks Richard and Charlie,
    We have implemented the solution/suggestions in our DEV environment and we have seen a definite improvement. We are waiting this to be deployed in UAT environment to note down the actual performance and time improvement while browsing the cube for the
    first time after daily cube refresh.
    Guys,
    This is what we have done:
    We have 4 cube databases and each cube db has 1-8 cubes.
    1. We are doing daily cube refresh using SQL jobs as follows:
    <Batch xmlns="http://schemas.microsoft.com/analysisservices/2003/engine">
    <Parallel>
    <Process xmlns:xsd="http://www.w3.org/2001/XMLSchema" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:ddl2="http://schemas.microsoft.com/analysisservices/2003/engine/2" xmlns:ddl2_2="http://schemas.microsoft.com/analysisservices/2003/engine/2/2" xmlns:ddl100_100="http://schemas.microsoft.com/analysisservices/2008/engine/100/100" xmlns:ddl200="http://schemas.microsoft.com/analysisservices/2010/engine/200" xmlns:ddl200_200="http://schemas.microsoft.com/analysisservices/2010/engine/200/200">
    <Object>
    <DatabaseID>FINANCE CUBES</DatabaseID>
    </Object>
    <Type>ProcessFull</Type>
    <WriteBackTableCreation>UseExisting</WriteBackTableCreation>
    </Process>
    </Parallel>
    </Batch>
    2. Next we are creating a separate SQL job (Cache Warming - Profitability Analysis) for cube cache warming for each single cube in each cube db like:
    CREATE CACHE FOR [Profit Analysis] AS
    {[Measures].members}
    *[TIME].[FINANCIAL QUARTER].[FINANCIAL QUARTER]
    3. Finally after each cube refresh step, we are creating a new step of type T-SQL where we are calling these individual steps:
    EXEC dbo.sp_start_job N'Cache Warming - Profit Analysis';
    GO
    I will update the post after I receive the actual im[provement from UAT/ Production environment.
    Best Regards, Arka Mitra.

  • SSAS Cube - Throws "attribute key not found" when the key exists

    Hi,
    I created a cube with 2 fact tables ( CabecNF and ItensNF ) and with some dimensions, in this case Cliente.
    When I proccess the cube it throws this error:
    Internal error: The operation terminated unsuccessfully.
    Errors in the OLAP storage engine: The process operation ended because the number of errors encountered during processing reached the defined limit of allowable errors for the operation.
    Errors in the OLAP storage engine: An error occurred while processing the 'ITENSNF' partition of the 'ITENSNF' measure group for the 'SCA' cube from the SCA_Cube_ database.
    Errors in the OLAP storage engine: The process operation ended because the number of errors encountered during processing reached the defined limit of allowable errors for the operation.
    Server: The operation has been cancelled.
    Errors in the OLAP storage engine: An error occurred while processing the 'CABECNF' partition of the 'CABECNF' measure group for the 'SCA' cube from the SCA_Cube_ database.
    Errors in the OLAP storage engine: The attribute key cannot be found when processing: Table: 'dbo_CLIENTE', Column: 'CODIGO', Value: '6571'. The attribute is 'CLIENTEID'.
    Errors in the OLAP storage engine: The attribute key was converted to an unknown member because the attribute key was not found. Attribute CLIENTEID of Dimension: CLIENTE from Database: SCA_Cube_, Cube: SCA, Measure Group: ITENSNF, Partition: ITENSNF, Record: 152957.
    I understand that is not finding the ClienteID 6571. But I've checked on my database and he exists, and he has itens in the CabecNF wich has itens on ItensNF.
    I've already customized the error configuration on the dimensions, but it didn't solve the problem.
    What could be causing this behavior?
    Thanks.
     

    i did nt get what are trying say by this:
    "First the Cliente dimension reported 500 rows skipped due to having the same name(even the name not being a key), so I changed the dimension and set the "name" property as not aggregable. Re-processed the dimension, and this time no skipped rows"
    1.) were u able to do a full process of the dimension successfully?
    if yes can you go and check the view or data source query that is making this dimension  have the value for the column 'CODIGO' 6571
    Yes, the process of the dimension was successful. At first successful with warnings, and then successful without warnings.
    But I found the problem: The Cliente dimension has relathionship with the Seguimento dimension, and the Cliente 6571 had NULL at that relationship. I fixed that and the cube was processed successfully.
    It was a bit hard to find the problem because the error didn't mentioned anything about the Seguimento dimension.
    Thanks for all your help guys!

Maybe you are looking for