Connect a Date Dimension to a cube without relationship

Hi everybody,
I would like to answers to one business requirements.
I create a cube that models the following event : a customer send a product from an agency to another customer who receives it in another agency.
So I have a fact table with only two measures
Amount
Count
which is connected to these dimensions
Product
Sending Date
Receiving Date
Sender (Customer)
Receiver (Customer)
Sender (Agency)
Receiver (Agency)
The users would like to analyse the following KPI, at a specific date :
Number of transactions sent, the amount
Number of transactions received, the amount
Number of transactions pending, the amount
To answer this business requirement, I have added a new date dimension in the cube with no relation, so that the user can select a date from this independent dimension and get the different KPIs. 
But I don't get any results.
Is it a good model ? How to make it possible for the user to use the independent Date Dimension to perform analyzes of different KPI ?

Is it a good model ? How to make it possible for the user to use the independent Date Dimension to perform analyzes of different KPI ?
Hi Meal,
According to your description, you want to know if is it possible for the user to use the independent Date Dimension to perform analyzes of different KPI, right?
As per my understanding, we cannot do this without relationship between dimension tables and the independent Date table. However, we can add the relationship between the added date table and the fact table Sending Date and Receiving Date column. Please refer
to the link below to see the details.
http://msdn.microsoft.com/en-us/library/ms175427.aspx
Regards,
Charlie Liao
TechNet Community Support

Similar Messages

  • How to add a new dimension in ASO cube without losing all data

    Hi,
    I have an ASO cube with 18 dimensions and I want to add a new one (regular). When I add this new dimension it asked me that the all data has to cleared before the restructure can take place. If I click yes then I would lose all data regardless of the fact if I have a level0 export or not. Because when I tried to reload the export back into the cube it gave me an error cause all the dimensions were not present in the export. Does anyone know how would I be able to accomplish this task?
    Thanks
    fikes

    You can try this work around solution.
    1) Export Lev0 data
    2) Clear the cube
    3) Add new dimension Regular.
    4) Add a default member Reg00 in the new dimension and save.
    5) Open the Lev0 data in a text editor, Insert "Reg00" in the begining of the file and Enter.
    6) Lev0 file will have first record "Reg00" and the remaining export file will be from the second record.
    7) Load the modified file with out rule file.
    You can verify that all your history data will be loaded w.r.to the Reg00 of the New dimension Regular.

  • Performance Point - 1 Dashboard, 2 cubes, 1 Time intelligence filter, filtering 2 different date dimensions on different cubes!

    Hi
    Hope you are well.
    I have a challenge. I am working with Performance Point and a number of different cubes.
    In this case I need to include graphs and charts from 2 different cubes on one dashboard page.
    The challenge I have, is to be able to use a Time Intelligence filter from one cube, eg, [TimeSold].[Calendar hierarchy] on Cube 1, and for this to filter graphs from the second cube (second cube [Time].[Calendar hierarchy]), on one dashboard page. 
    Is there any way I can achieve this using MDX or Parameters within performance point?
    I cannot amend the cubes or dimensions. I have to work with the front end tools only.

    Hi ShebUK,
    I don't think we can achieve this requirement on the front end, but you can discuss this issue at the following forum:
    http://social.technet.microsoft.com/Forums/en-US/home?category=performancepointserver
    I'm not familiar with PerformancePoint Services and not sure that we can implement linked dimension to achieve this. A linked dimension is one that exists in one Analysis Services database, but reused in another Analysis Services database of the same version
    and compatibility level. For more information, please see:
    Define Linked Dimensions: http://msdn.microsoft.com/en-us/library/ms175648.aspx
    Hope this helps. 
    Regards,  
    Elvis Long
    TechNet Community Support

  • Updating metadata/data in ASO reporting cube from BSO

    Hi Experts,
    I am creating an ASO reporting cube and would like to know the best way to automate the update of metadata and data from a BSO cube without the use of a partition? This is my first time working with ASO. Any insight or examples would be much appreciated. Please advise.
    Thanks in advance!

    Thank you for the responses!
    The dimensions will be the same. I will be exporting level0 data then loading it into the ASO cube. I believe I should be able to use the same load rule that I am using for the BSO cube, correct? Also, I am planning on only loading the 'Final' versions of 'Budget' and 'Forecast'. In my data export I believe I can put a condition to only export the 'Final' version, please correct me if I'm wrong?
    One last thing, I would like to keep 2 years of 'Actuals' and 1 year of 'Budget' and 'Forecast'. At year end I would to clear previous 'Budget' and 'Forecast' and 2nd prior year after the current 'Budget' and 'Forecast' are closed. Can this be handled in a separate script? Please advise.
    Thanks again for the responses.

  • Date Dimension not updating correctly

    Hi
    My current date dimension underlying data was due to expire soon so I set about trying to create a new one.
    I got hold of the script that it was based and went about re-creating the data like for like.
    I pointed the date dimension at this new data structure and it all processed fine.
    When I was doing some testing I noticed althought the total figures were fine, some of the year figures were a little off compared to the underlying data.
    I found that dates such as 31/12/2012 were being classified as year 2013. The month was also being classified as January but the date was still showing as 31 December 2012
    I switched the date dimension back to the old structure and all year numbers were fine.
    I decided to recreate the Date Dimension in the cube using the same structure as the current date dimension and when it processed with the new data structure all the numbers are correct.
    I tried clearing cache but I am baffled as to why the new data structure won't work with the existing Date Dimension but will work with a newly created date dimension in the same cube?
    Regards

    Hi
    I seemed to have solved it (although I am still testing), there was a data diffrence in the week key and that it was over lapping with the end of the year so I changed it to match what I already had.
    It still doesn't explain that before making the change when I created a new dimension it seemed to work ok.
    I had a look at the XML and the only difference that I could see was that the original date dimension had a Time Binding that was generated on the server or the data source.

  • How to Maintain Cube without Aggregating the incoming data

    Hi All,
    I have a process which loads a cube and then aggregates the incoming data. The PL/SQL procedure is as below.
    CREATE OR REPLACE PROCEDURE load_actual AS
    xml_clob clob;
    xml_str varchar2(4000);
    isAW number;
    begin
    DBMS_LOB.CREATETEMPORARY(xml_clob,TRUE);
    dbms_lob.open(xml_clob, DBMS_LOB.LOB_READWRITE);
    dbms_lob.writeappend(xml_clob, 183, ' <BuildDatabase Id="Action4" AWName="FIN_AW.FINAW" BuildType="EXECUTE" RunSolve="true" CleanMeasures="false" CleanAttrs="false" CleanDim="false" TrackStatus="true" MaxJobQueues="0">');
    dbms_lob.writeappend(xml_clob, 42, ' <BuildList XMLIDref="ORG_ACTUAL.CUBE" />');
    dbms_lob.writeappend(xml_clob, 18, ' </BuildDatabase>');
    dbms_lob.close(xml_clob);
    xml_str := sys.interactionExecute(xml_clob);
    dbms_output.put_line(xml_str);
    end;
    As per my requirement I have to flip sign for some leaf level dimensions after the cube load and then re run the aggregate. If I go with this script I will end up aggregating my data twice.
    My question is how do I modify this script so that the records are loaded in the cube but the aggregate does not run.
    I tried by changing RunSolve="true" parameter as below but still it is solving the data.
    dbms_lob.writeappend(xml_clob, 184, ' <BuildDatabase Id="Action4" AWName="FIN_AW.FINAW" BuildType="EXECUTE" RunSolve="false" CleanMeasures="false" CleanAttrs="false" CleanDim="false" TrackStatus="true" MaxJobQueues="0">');
    My cubes are compressed, partitioned and I am using Oracle 10.2.x.
    Please provide your suggestions if I missing something here ?
    Thanks,
    Vishal

    10g Cubes have a property called 'AutoSolve'. As far as I understand, if this is set to YES, then the aggregation will happen automatically whenever you build the cube. I may be missing something, but the RunSolve property of the BuildDatabase command does not seem to be used.
    In AWM the "AutoSolve" property appears to correspond to the check box
    "Use Default Aggregation Plan for Aggregation"
    in the "General" tab for the cube. In order to uncheck this you also need to uncheck "Use Compression" on the "Implementation Details" tab.
    As an example, I created a cube named C1 with the box checked and then exported the XML.
    <Cube Name="C1" LongName="AMERICAN::C1" ShortName="AMERICAN::C1" PluralName="AMERICAN::C1" Id="C1.CUBE" Schema="GLOBAL" UseGlobalIndex="false" SparseType="COMPRESSED" DefaultDataType="NUMBER" AutoSolve="YES">
    I then created cube C2 with it unchecked.
    <Cube Name="C2" LongName="AMERICAN::C2" ShortName="AMERICAN::C2" PluralName="AMERICAN::C2" Id="C2.CUBE" Schema="GLOBAL" UseGlobalIndex="false" SparseType="STANDARD" DefaultDataType="NUMBER" AutoSolve="NO">
    The compression check is enforced by AWM. I hand modified the XML for cube C1 to switch AutoSolve to NO and used it to recreate the cube without errors (as a compressed cube). I don't know if there will be any issues downstream.

  • How to tune performance of a cube with multiple date dimension?

    Hi, 
    I have a cube where I have a measure. Now for a turn time report I am taking the date difference of two dates and taking the average, max and min of the date difference. The graph is taking long time to load. I am using Telerik report controls. 
    Is there any way to tune up the cube performance with multiple date dimension to it? What are the key rules and beset practices for a cube to perform well? 
    Thanks, 
    Amit

    Hi amit2015,
    According to your description, you want to improve the performance of a SSAS cube with multiple date dimension. Right?
    In Analysis Services, there are many tips to improve the performance of a cube. In this scenario, I suggest you only keep one dimension, and only include the column which are required for your calculation. Please refer to "dimension design" in
    the link below:
    http://www.mssqltips.com/sqlservertip/2567/ssas--best-practices-and-performance-optimization--part-3-of-4/
    If you have any question, please feel free to ask.
    Simon Hou
    TechNet Community Support

  • How to delete aggreagetd data in a cube without deleting the Aggregates?

    Hi Experts,
    How to delete aggreagetd data in a cube without deleting the Aggregates?
    Regards
    Alok Kashyap

    Hi,
    You can deactivate the aggregate. The data will be deleted but structure will remain.
    If you switch off the aggregates it wont be identified by the OLAP processor. report will fetch the data directly from the cube. Switching off the aggreagte won't delete any data,but temporarly the aggregate will not be availbale as if it is not built on the info cube. No reporting is not possible on swtiched off aggregates. The definition of the aggregate is not deleted.
    You can temporarily switch off an aggregate to check if you need to use it. An aggregate that is switched off is not used when a query is executed.This aggregate will be having data from the previous load's. If the aggregate is switched off means it wont be available for reporting but data can be rolled up into it.
    If u deactivate the aggregates the data will be deleted from the aggregates and the aggregate structure will remain the same.
    The system deletes all the data and database tables of an aggregate. The definition of the aggregate is not deleted.
    Later when you need those aggregate once again you have to create it from scratch.
    Hope this helps.
    Thanks,
    JituK

  • Date dimension fragmentation from 2 different database connections

    Hi Everyone,
    In obiee 10g, is it possible to implement fragmentation on a date dimension from 2 different physical database connections (not schemas)?
    For example:
    Physical connection & dimension 1: physical1.obiee_rpd.w_mcal_day_d.date
    physical connection & dimension 2: physical2.obia_rpd.w_mcal_day_d.date
    the fact table will be: physical1.obiee_rpd.w_fact_f.fact
    The 2 physical dimension tables will be joined to the 1 fact table in the physical layer.
    I then create a BMM with 2 LTS : physical1.obiee_rpd.w_mcal_day_d and physical2.obia_rpd.w_mcal_day_d.
    I then create 1 logical column in that BMM which sources both physical1.obiee_rpd.w_mcal_day_d.date and physical2.obia_rpd.w_mcal_day_d.date.
    Then, i set up fragmentation to use physical1.obiee_rpd.w_mcal_day_d.date when physical1.obiee_rpd.w_mcal_day_d.date < '04/30/2012' and use physical2.obiee_rpd.w_mcal_day_d.date when physical1.obiee_rpd.w_mcal_day_d.date >= '05/01/2012'.
    The problem is that the query that answers generates always uses physical1.obieerpd.w_mcal_day_d.date , regardless of the date filter applied!_
    Any tips?
    Thanks!

    Hy,
    Can you try the following.
    Define:
    - one logical date dimension with two logical table sources (physical1.obiee_rpd.w_mcal_day_d.date and physical2.obia_rpd.w_mcal_day_d.date)
    - one logical fact table with two Logical Table Source (LTS). And in each LTS, define the same logical level in the content tab and don't forget to check the checkbox "This source should be combined with other sources at this level".
    Cheers
    Nico
    Edited by: gerardnico on Jun 25, 2012 4:23 PM Date and not datum (Dutch)

  • Modify Cube without clear data

    Hi all,
    By request I changed a cube i.e. add several key figures into this cube. Following this change the related update rule and DTP also changed. Now I want to transport this changed cube into Productive system but don't want to delete all data in the cube before this transport, is this possible and what are the precautions ?
    Thanks in advance

    Dear Leon Ouyang,
    We can transport the INfo cube without data deletion into production system with out any issues.But only thing is it will take more time (time depends on the data volume i.e. if there is huse data it will take more time).
    With my previous observation ,i found for 25 million records it took 1hour15 mins of time to overwrite the exesting info cube in production.
    Here is the logic behind the object transport which is already exist in production:
    When ever we are transporting obects from one system(dev sys) to another system (prod sys) ,at the time of transport all the data will be transferd in to buffer and revert back to original object (by system it self) after the objects got sucess fully imported.
    Below are the Precaution we have to take while transporting the info cube are:
    1) Make sure that is no local request created for the transport objects( in your scenario info cube,DTP,Transformation..etc).
    2) Data loads should not happen while transportion.
    3) Make sure that transport path from  dev to prod is perfect(we can check in se09 transaction).
    4) Better to Transport object in from top to bottom i.e. 1st info cube and then depending transformation,dtp's etc..
    5) Monitor the transport job wile transoporting.we can find job with RDD* IN Sm37 transaction.
    Hope this may give you some input for transporting objects...
    Pls let me know if you require any further info..
    Cheers,
    Maruthi

  • Compare data in R/3 with data in a BW Cube after the daily delta loads

    Hi Friends,
    How can I compare data in R/3 with data in a BW Cube after the daily delta loads? Are there any standard procedures for checking them or matching the number of records?

    Hi Sunil,
    If you want to check the records daily instead of checking the data in R/3 manually ......
    You can try this...
    If you have staging DSO(level 1) that means whatever data is in source system load it to Staging DSO without any routines or any modifications.
    Now load this DSO data to Cube or DSO(level 2) as per your requirement with routines etc.
    Now Staging DSO contains Source system data.
    Now the level 2 Cube or DSO contains BW data with some modifications.
    Now create a Multiprovider based on level 1 and level 2 data targets.
    Now create a report on which keyfigures you want to test the data.
    In Multiprovider there is a field called 0infoprovider in data packet dimension.
    you can drag this infoprovider to the columns and restict your keyfigures with level 1 and level 2 data targets.
    In the first column you can see the level 1 DSO data ( source system data),in the 2nd column you can see the BW data.
    Now create a formula which gives the diffrence b/n level 1 and level2.
    that is R/3 data - BW data.
    If the diffrence is zero both R/3 and BW data are same.
    if the diffrence is not eqaul to zero check whether any routine is there or not.

  • Data in DSO and CUBE

    Hi Experts
    Please..Please update me on in detail if possible with example ......Key Fields and Dimensions
    How Data Records will be processed in DSO and  Cube
    I tried to search but i can't able to find what i am looking for
    My Requirment:
    I am extracting Employee Benefits records from Non SAP Source System (Oracle Tables)
    In the Oracle there is no Date field avaliable and there won't be any history avaliable in oracle once the Benefits of an employee is changed the old record will be overwritten to new one so they can't track the employee benefits history...but in BW i need to store/show history of employee benefits by Cal day
    Oracle Table Fields
    Location (Primary Key)
    Emp No (Primary Key)
    Insurance Type
    Call Allowance Type
    Annual Leave (No of day Days)
    Pension Schem
    Company Laptop (Yes/No)

    hi,
    key fields are the primary keys of ur ods tables. based on the key field values the data fields will get overwritten.
    suppose if the key fields values were same then the data fields will get overwritten. but the changes were captured in the change log table, hence the history of changes is available.
    Dimensions- are the prmary keys of cube. but in cube only addition of records will happen as default not overwrite as in ods.
                       dimension ids were generated based on the sid values of characteristics.
    maximum of 16 key fields and dimensions can be there in ods and cube respectively.
    for ur case, include 0calday field in the key fields and use the routine to update the field with system date(SY-Date). this keeps track/ else without date also , change log maintains history for each load.
    have to add 0recordmode to the communication structure of the ods(infosource).
    Ramesh

  • SSAS Date Dimension attribute not displaying all members when browsed singly

    I have an SSAS Cube database that holds data from 2009 to 2013 (5 years)
    There is a Date Dimension in the database which holds a "Calender Period" Hierarchy with the attributes CalenderYear >> Quarter >> MonthYear >> Day .
    I am connecting to the cube via Microsot EXCEL
    Whenever i try to browse this hierarchy alone in the spread sheet, the first attribute as per the hierarchy shown above i.e. Calender Year does not display all the members, that is confusing the users to assume that the cube is not processed with the latest
    data, which is not the case. However as soon as i try to browse any measure with this dimension, the data automatically shows for the latest dates of 2013 as per the expectations.
    The perspective that is selected while connecting to the Cube in this case is the one that is by default the one including all the Cube Objects.
    Whenever i try to browse the same cube for the same Dimension Hierarchy but in a separate user perspective i get the latest three Calender Years i.e. 2011,2012 and 2013 but not the complete years 2009 to 2013.
    When i was browsing this earlier by the original base perspective (which has all the cube objects visible) the Calender Years displayed were 2009,2010,2011&2012 but not 2013 which should be there.
    I checked the base view in the DB that is used to create this Dimension, that is fetching the complete data for 2009 to 2013.
    I have also verified that this issue is still there if i try to browse the cube from the SSMS.
    Can anyone help me to get all the years from 2009 to 2013 (the years for which my cube has data) become visible as soon as i drop it alone into the Spreadsheet?
    Thanks 
    Suvrat

    Thanks for your response Muthukumaran.
    I did not know about this default measure property of the SSAS cube, it would be nice if you could also tell me how to check which is the default measure for a cube.
    I understand your point that if a particular measure group does not have data for a year then that particular year will be skipped as per the default excel settings, which when disabled allowed all the years to be displayed, but in that case every time the
    users connect to the cube via excel, they will have to configure this property.
    I would like to share the way i resolved this issue:
    I went ahead and simply did a Process Full for the Date dimension. The years started to display by default as per the expectations. Now after observing this change, as per my understanding I believe this issue was due to the CACHE of the cube not being updated
    for the date dimension. As after each incremental load of the DWH we processed the cube & while processing the cube the dimensions were set for a Process Update and the Measure Groups were set for a Process Full. This caused this issue, and we could resolve
    this once the Date Dimension was processed full.
    Thanks
    Suvrat Rai

  • Dates against rows and columns in pivot require two date dimensions???

    Hi,
    I have an Excel pivot table connected to a SSAS tabular cube.Now For this Report i want the row label area of pivot to contain Shipping Month and column label area contains the Order Months.Do i need two date dimensions for this purpose?? What is the way
    to do such kinds of Reports where the row as well as columns contain date dimension???
    Thanks.

    Hi Nimo0102,
    Based on your description, you want Shipping Month to be shown on row label, and Order Months to be shown on the column label, right? When connecting to a data base on an Excel pivot table, if we want to the dimension appear on the row or column label, we
    can drag the dimension to ROWS or COLUMNS. So in your scenario, if you have Shipping Month Order Months on your cube, you can drag it to ROWS or COLUMNS directly. Otherwise, I am afraid you need to create them on the cube.
    If I have anything misunderstood, please point it out.
    Regards,
    Charlie Liao
    TechNet Community Support

  • Adjusting dimensions in the cube

    hi,
    One of our cubes have more than 40 million records. But the reports on it are running slow. so i checked the fact dimension ratios, then they are not efficient. So, I need to adjust the dimensions in such a way that they are efficient. In order to do that i have to delete and load the data again which is very huge...
      Is there any way that without deleting the data can i adjust the dimensions in the cube? Please let me know.
    Thanks in advance....

    How do you say that? Did you try to do some checks in RSRT. May be some indexing or aggregates might help.Tru to use reporting agent.  If not, try this in your dev syste or sandbox.
    1. Send the data to another cube ( may be you have to make a new cube as copy of the old cube).
    2. Model another cube the way you wish to remodel the cube.
    3. Create update rules and schedule the package.
    4. Take care of things like partioning, cache, use of multi provider, etc.
    5.  Perform stress and performance test.
    6.  Copy all the queries fro the old cube to this cube.
    7.  Collect all the objects and transport.
    If you are in 7.0, then you have some options. But in 3.5 it is bit hard.
    Ravi Thothadri

Maybe you are looking for