Historic and Current data for Master data bearing objects

Hi All,
We are trying to implement type 2 dimensions for all the master data bearing characteristics, where we require historic and current data available for reporting. the master data can have a number of attributes and all of them can be time dependent. We are not getting any 'datefrom' or 'dateto' from the source system.
For example:
For Example: The table below shows data entering BI at different dates.
Source Data day of entering BI
MasterID ATTR1 ATTR2
123506 Y REWAR day1
123506 N REWAR day4
123506 Y ADJUST day4
123506 N ADJUST dayn
The field 'day of entry into BI' is only for your understanding; we do not get any date fields from the source. SID is the field we are generating for uniqueness. It is a counter. EFF_DATE would be the current date for all the data. EXP_DATE would be 31.12.9999 until the attributes change. SID and MasterID together would be the key.
On day 1 the following data enters BI,
day 1
SID MasterID ATTR1 ATTR2 EFF_DATE EXP_DATE
1 123506 Y REWAR 2/10/2009 12/31/9999
On day 4, 2 data records enter with same PID,
SID MasterID ATTR1 ATTR2 EFF_DATE EXP_DATE
2 123506 N REWAR 2/13/2009 12/31/9999
3 123506 Y ADJUST 2/13/2009 12/31/9999
the EXP_DATE of the record of day 1 needs to be changed to current date.
Also there are two records entering, so latest record would have EXP_DATE as 31.12.9999. And the EXP_DATE of the first record on day 4 should change to the current date.
so the following changes should happen,
CHANGE
SID MasterIDATTR1 ATTR2 EFF_DATE EXP_DATE
1 123506 Y REWAR 2/10/2009 2/13/2009
CHANGE
SID MasterID ATTR1 ATTR2 EFF_DATE EXP_DATE
3 123506 Y ADJUST 2/13/2009 2/22/2009
On day n, one data record enters with same PID,
SID MasterID ATTR1 ATTR2 EFF_DATE EXP_DATE
4 123506 N ADJUST 2/22/2009 12/31/9999
The change is ,
CHANGE
SID MasterID ATTR1 ATTR2 EFF_DATE EXP_DATE
3 123506 Y ADJUST 2/13/2009 2/22/2009
The data expected in P-table is as below, on Day n or after Day n, untill any other record enters for this MasterID,
1 123506 Y REWAR 2/10/2009 2/13/2009
2 123506 N REWAR 2/13/2009 2/13/2009
3 123506 Y ADJUST 2/13/2009 2/22/2009
4 123506 N ADJUST 2/22/2009 12/31/9999
Has anyone worked on type 2 dimensions earlier? Or any ideas to implement this logic would be appreciated.
Regards,
Sudeepti

Compound the Master ID with eff date and other attribute as superior objects
so you will get P-table as
ATTR1   ATTR2   MAT ID  
1 2/10/2009 2/13/2009 123506 Y REWAR
2 2/13/2009 2/13/2009 123506 N REWAR
3 2/13/2009 2/22/2009 123506 Y ADJUST
4 2/22/2009 12/31/9999  123506 N ADJUST

Similar Messages

  • Is there any way we can set default value for a Date Attribute to current date in Master Data Services

    Is there any way we can set default value for a Date Attribute to current date in Master Data Services.
    I as well wants to know that is there any posibility to show Calendar control while input data into respective date attributes.
    Thanks.

    Hi Anagha,
    So far i havent found any way to set todays date by default from MMI, but i guess this flow should work as workaroud
    1. Add buisness rule which can set a default value when Date = NULL/Blank
    2.get the entity table ,use -select EntityTable from mdm.tblEntity where Name = '<enter entity name>'
    3.Go to that table and add a after update trigger like
    if uda_<entityid>_<attributeid(Date attribute)> = default value
    update uda_<entityid>_<attributeid(Date attribute)> =getdate() where id = <LastUpdatedRow>
    I will check on this too from my side.
    By the way AFAIK i dont think so calendar control integration is possible .

  • Is there any deffirence for master data from flat fime bi 7.0 and 3.5?

    hi friends,
    is there any deffirence for master data from flat fime bi 7.0 and 3.5?
    regards
    suneel.

    you can follow the same 3.x dataflow eventhough u upgrade.
    the only diff
    3.x - would be emulated DS
    7.0 - RSDS
    3.x - file path declared in infopack
    7.0 - is declared in DS (that would be inherted by infopack)
    Infosource - optional - works fine on both the versions
    creating a new DS, it has to be RSDS .. no way out

  • DBconnect for master data

    Hi everbody,
    we are extracting customer master data from an Iseries-system cia DBconnect.The relevant info object 'Customer' has attributes and texts.
    Generation of the data source and loading of data worked without problems for the info object attributes (after generating the data source in the source system tree with setting 'master data attributes for 'data source type').
    Now i would like to extract from the same data source the info object -texts. Problem is that the system does not allow to assign for the same info source 'customer' the source system for a second time: 'No data sources for source system 'Iseries' available/message RSAR383.
    This is although I did generate the data source (below the source system tree) a second time with the setting 'master data texts' (for data source type).
    While it is normal that I can connect the data source only once to an info source for transaction data, I would expect to be able to use the data source twice for master data when differentiating between 'text' and 'attributes'.
    Has anybody an idea on this topic?
    Thanks and best regards
    Marcus

    Hi Joe, Hi Uday,
    thanks for the further clarification on the topic. The problem is now solved following your recommendations. From the source system we receive attributes and texts in one view. The corresponding data source was generated with type  'texts'. I had to manually enter the text fields 0TXTSH, 0TXTMD, 0TXTLG in info source and had to maintain mapping in the communication structure.On info provider level it was possible to create two update rules (type attribut and type text). Then-using one info package- the upload of attributes and texts worked fine.
    Thanks again for your support
    Best Regards
    Marcus
    I tried to reward due points, but I receive a message Rewarding the message failed
    Please bear with me, I will figure that out and reward the points.
    Message was edited by: Marcus Theviot

  • DTP for Master data

    Hi Guys,
    How to extract master data and hierarchies from ECC into BI infoobjects?
    Do we follow the same procedure for loading master/hierarchy data as for transaction data - i.e. Load into PSA and then create Transformations and DTPs to load into infoobjects? If so,  what is the best approach as far as Full/Delta loads.
    1. Is it ok to do a full load into PSA (using infopackage) and then do a full load always (The process chains for master data would be running nightly).
    or
    2. Should we have 2 infopackages (full and delta) into PSA and then multiple DTPs. (The process chains for master data would be running nightly).
    Regards.

    Indicator: Only Get Delta Once
    Source requests of a DTP for which this indicator is set are only transferred once, even if the DTP request is deleted in the target.
    Use
    If this indicator is set for a delta DTP, a snapshot scenario is built.
    A scenario of this type may be required if you always want an InfoProvider to contain the most up-to-date dataset for a query but the DataSource on which it is based cannot deliver a delta (new, changed or deleted data records) for technical reasons. For this type of DataSource, the current dataset for the required selection can only be transferred using a 'full update'.
    In this case, a DataStore object cannot usually be used to determine the missing delta information (overwrite and creation of delta). If this is not logically possible because, for example, data is deleted in the source without delivering reverse records, you can set this indicator and perform a snapshot scenario. Only the most up-to-date request for the DataSource is retained in the InfoProvider. Earlier requests for the DataSource are deleted from the (target) InfoProvider before a new one is requested (this is done by a process in a process chain, for example). They are not transferred again during the DTP delta process. When the system determines the delta when a new DTP is generated, these earlier (source) requests are seen as 'already fetched'.
    Setting this indicator ensures that the content of the InfoProvider is an exact representation of the source data.
    Dependencies
    Requests that need to be fetched appear with this indicator in the where-used list of the PSA request, even is they have been deleted. Instead of a traffic light you have a delete indicator.
    Get Data by Request
    This indicator belongs to a DTP that gets data from a DataSource or InfoPackage in delta mode.
    Use
    If you set this indicator, a DTP request only gets data from an individual request in the source.
    This is necessary, for example, if the dataset in the source is too large to be transferred to the target in a single request.
    Dependencies
    If you set this indicator, note that there is a risk of a backlog: If the source receives new requests faster than the DTP can fetch them, the amount of source requests to be transferred will increase steadily.

  • How to do Stastical Setup for Master data Loading

    Hi All
    Thanks a lot for all guys,who are helping me a lot.
    I want to do stastical setup for Master data of sales order.I know how to do for transaction data.
    But what is the procedure for MAster data in SBIW-Tcode.
    Regards
    Balaji

    Hi,
    <i>Actually first we should load data to Master data table and then to Transaction load data.By this reporting performance will be good.Right?</i>
    Actually this will decrease the time of transaction data loading .So it improves the loading performance ,  not reporting performance.
    <i>Now I got a doubt from where this Master Data is been loading? and how to load this data.</i>
    2LIS_11_VAITM and 2LIS_11_VASCL are the DS for Transaction data , not for master data.To load the master data, first of all find out the master bearing characterstics in the data target.And then search for teh DS of master data in R/3 with tcode RSA6 .You will use characterstic(info object) as the search option here.
    For example 0materail is the Master characterstic.
    And you can see the DS for this Characterstic in RSA6(if it is not there, you have activate it from RSA5)
    0MATERIAL_ATTR
    0MATERIAL_LGEN_HIER
    0MATERIAL_LKLS_HIER
    0MATERIAL_TEXT
    With rgds,
    Anil Kumar Sharma .P

  • About delta loading for master data attribute

    Hi all,
    We have a master data attribute loading failed which is a delta loading. I have to fix this problem but I have two questions:
    1. how can I find the those delta requests because I need to delete all these failed requests first, am I right ? Master data is not like cube or ods that we can find the requests in Manage, for master data how can we find them ?
    2. Could you please let me know the detailed procedures to perform this delta loading again ?
    Thanks a lot

    Hi...
    1. how can I find the those delta requests because I need to delete all these failed requests first, am I right ? Master data is not like cube or ods that we can find the requests in Manage, for master data how can we find them ?
    Look.....for master data.....no need to delete request from the target..........just make the status red.......and repeat the load.....But problem is that master data sometimes does'nt support Repeat delta..........if u repeat......then load will again fail with Update mode R.........in that case u hav to do re-init.......
    1) delete the init flag.......(In the IP scheduler >> in the top Scheduler tab >> Initialization option for source system)
    2) Init with data transfer(if failed load picks some records)..........otherwise .....init without data transfer.....if the last delta failed picking 0 records.......
    3) then Delta.......
    2. Could you please let me know the detailed procedures to perform this delta loading again ?
    1) Make the QM status red.........to set back the init pointer.......
    2) Repeat the load.....
    After that.........if again load failed with Update mode R.....
    1) delete the init flag.......
    2) Init with data transfer(if failed load picks some records)..........otherwise init without data transfer.....
    3) then Delta.......
    Regards,
    Debjani.....

  • Unable to generate "Export data source" for Master data infoobject

    Hi Gurus,
    I'm in BW 3.5
    I need to generate "Export data source" for a master data info object.
    I followed below steps:
        I ticked the checkbox "Characterist. is export data source" in the master data Info object maintenance screen.
        In the Infosource area, for the desired Infoobject --> Right click --> Generate export datasource --> a dialog box displays saying "The generation of DataSource  was successful"
    Problem:
    Even after refreshing, etc.. I couldn't find the generated datasource 8**** for the above master data. I also tried to replicate the datasource from the myself source system, but couldn't find it even there. I don't know where the generated export data source is hiding ???
    Could anyone please help ????
    Thanks,
    Venkat

    Hi Venkat,
    I guess that after creating an export datasource for infoobject this DS should appear in the RSA1-Source systems - MYSELF - datasource overview (with name starting with 8). Since it's a datamart, it should appear also as system generated infosource. And in this case you would be able to assign in URs of data target the appropriate infosource (make sure that in Infoproviders tab of RSA1 you have inserted your infoobject as a data target).
    However, there are some bugs related to export data source. CHeck if you see you export data source in RSA6.
    If you don't see it - consider applying OSS Note #816892 - "30BSP26: Export DS for master data is not generated".
    Note #876845 - "30BSP29: InfoObject change: Export DataSource not adjusted" is also may be useful.
    -Vikram

  • Error while creating Data Source for master data attributes

    Hi BI Experts,
    Well its been some time for me that I have been part of Extraction in BI.I primarily handled reporting in my last assignments.
    I was trying extraction with flat files in SAP BI 7(new to sap bi 7 but very much familiar with BW3.5) but failed in the activity during master data attributes and text upload in infoobject (say IOSP_Mat).
    Here is the procedure that I did after creation of characteristic IOSP_Mat.I created a source system for flat file followed by data source for Master data attributes, i selected all the parameters correctly.i.e. csv file format, data seperator as   ,
    and other settings, now when i am trying to look at the proposed data in the next tab using Load example data.its not showing the desired result.The columns that I have maintained in Flat File is  MAT_NUMBER and MAT_NAME (with say 100 data in the file)
    same is the result when I am trying to load the text data too columns maintained are
    (LANGUAGE MAT_NUMBER Short Description)(same 100 data).
    now i used to rsa1old transaction  to upload the file using 3.5 version.i created info source for master data/text/hierarchies for  IOSP_Mat
    now when trying to upload it using info package for master and text data,I observe its(the data) not maintained in the characteristic IOSP_Mat.
    When I monitored ,I figured the data has not been even uploaded to the PSA level.
    Can you BI experts tell me the answer for this.
    Thanks,
    Srijith

    apologies to all of you for late response,
    was busy with some other activities.
    I don't remember the exact message,but I remember it was not loaded to even the PSA level.I will try it again and post the exact message.
    Thanks again for your quick response.
    Once again sorry to all of you for my late response
    Thanks,
    Sri

  • Creation of OLAP Variable for Master data

    Dear BW Experts,
    How to Create OLAP Variable for Master data  in SAP BW? What is the use of OLAP Variable?

    The entry column Type (Variable) on the Select Data tabpage indicates a variable time frame in which the data should be requested. The variables are not replaced by concrete values until a request is made. There are eight options:
    Variable            Data request        Description
    0
                          Yesterday
                                            (12am – 12am)
    1
                         Last week
                                            (Monday-Sunday)
    2
                           Last month
                                            (1. (First day - last day of previous month)
    3
                           Last quarter
                                           (1. (First day - last day of previous quarter)
    4
                           Last year
                                          (01. (January 1 – Dec 31 of previous year)
    5
                      User time selection
                                               See below
    6
                       ABAP routines
                                               See below
    7
                         <b>OLAP variables</b>
                                                See below
    <b>Time selections with the help of the variables 0 to 4.</b>
    The variables 0-4 only refer to date fields of the data type DATS.
    <b>User time selection</b>
    Type 5 gives you a free selection of all fields.
    If you select this variable for a field, confirm the selection, and then choose Detail for Type, an additional dialog box appears. In this dialog box you can freely limit the values of the fields and determine the next period value and the number of periods until a repetition.
    Variables are placeholders for values and are not replaced by concrete values until a data request is made. You can also use a Variable for selections when requesting data for an InfoObject, if you choose type 7, confirm the selection, and then choose Detail for Type.

  • Error handling for master data with direct update

    Hi guys,
    For master data with flexible update, error handling can be defined in InfoPackege, and if the load is performed via PSA there are several options - clear so far. But what about direct update...
    But my specific question is: If an erroneous record (e.g invalid characters) occur in a master data load using direct update, this will set the request to red. But what does this mean in terms of what happens to the other records of the request (which are correct) are they written to the master data tables, so that they can be present once the masterdata is activated, or are nothing written to masterdata tables if a single record is erroneous???
    Many thanks,
    / Christian

    Hi Christian -
    Difference between flexible upload & Direct upload is that direct upload does not have Update Rules, direct upload will have PSA as usual & you can do testing in PSA.
    second part when you load master data - if error occurs all the records for that request no will be status error so activation will not have any impact on it i.e. no new records from failed load will be available.
    hope it helps
    regards
    Vikash

  • Why to use chnage  pointers for  master data idoc why not  the  change idoc

    Hi Gurus,
    I have  one doubt about  Idoc.
    When changes to master  data has to be sent Change pointers are configured and  used. (CDPOS & CDHDR). In case of  the  transaction data change change idoc  is  used as is the case with orders (we  use ORDCHG for  ORDERS message type to send  the  chnage details to a  order which was already sent to other system.
    Why we can't use  change  idoc or  message type whatever, instead of  Chnage pointers in case of  master data or  vice versa in case  of transaction data.
    Your valuable  input  will be rewarded with suitable  points!!
    -B S B.

    Hi,
    It was a good question ... SAP seems never designed the change pointer to handle transactional data IDocs. Only for master data IDocs distribution purposes.
    It may due to volume data ... the transactional data can change many times within a day compare with master data which hardly change or once a while.
    Regards,
    Ferry Lianto

  • Ignore duplicate records for master data attributes

    dear  experts ,
                   how & where can i  enable "ignore duplicate records" when i am running my DTP to load data
                     to master data attributes.

    Hi Raj
    Suppose you are loading master data to InfoObject and in PSA you have more than one records for Key.
    Let's assume you are loading some attributes of Document NUmber ( 0DOC_NUMBER) and in PSA you have multiple records for same document number. In PSA this is not a problem, as Key for PSA table is technical key and which is different for every record in PSA. But this is a problem for InfoObject attribute table as more than  one records for primary key ( here 0DOC_NUMBER) will create Primary Key consraints in database.
    This issue can be easily avoided by selecting "Handle Duplicate Records" in DTP . You will find this option under *update" tab of DTP.
    Regards
    Anindya

  • Effect of 'No update' for master data texts

    Hello all,
    I am using flexible update - update rules - for master data text. I want to confirm effect of 'No update' in following scenario
    In the fisrt set of update rules -from first source system -I have mapped both short text and long text
    Key  |  short text| long text
    ......| overwrite| overwrite
    K1    | some SHTX| some LGTX
    In the second set of update rules, I have selected 'No update' for Long text
    Key  |  short text | long text
    ......|overwrite| NO UPDATE
    K1    | some SHTX |
    My loading sequence is first source system followed by second source system
    what should be my ultimate long text and why?
    Many thanks in advance!
    Regards
    Sanjyot

    Hi surekha,
    your long text will not get changed.
    Take the example you mentioned
    Key | short text.....| long text
    ........| overwrite......| overwrite
    K1 | some SHTX.. | some LGTX
    In the second set of update rules, I have selected 'No update' for Long text
    Key | short text... | long text
    ........|overwrite......| NO UPDATE
    K1 | some SHTX. |
    In the second set of update rules, you mentioned "No UPdate", So, when a record with the same key is found n Masterdata, it is not going to make any changes to the field Long text.
    Where as short text will get overwritten.
    Take some data for ex with the above mentioned type
    data from source 1 : 2000 Maggi Noodles
    data from source 2 : 2000 Feasters
    when the data from first source system arrives, data will be 2000 maggi noodles
    when the data from second source suystem arrives, as there exist this records with same key(2000), it checks for the other fiels
    short text will be over written with the value in second system (according to rules its in overwrite mode)
    Long text will not be updated as the rule says no update.
    hope this helps,
    Cheers,
    Srinath.
    Cheers,
    Srinath.

  • Why we are not using "Message Control" for Master data ?

    Hi friends,
    We are using Message control as output determination for Transactional data, <b>what is the main reason for not using the msg control for Master data.</b>Please explain with details...
    I am very thankful to them...
    Thanks and Regards.
    <i><b>Naveen Kumar. Chamala</b></i>

    Hi Naveen,
       Listing down those I know -
       1. Master data is expected to be changed/created less frequestly.
       2. Transactional data is expected to be chnaged/created more frequestly.
       3. Transactions to create/change/display master data do no have message control integration, since you are not supposed to print/fax/mail master data to your partners.
       4. Distribution of master data in generally supposed to happen within Enterprise boundry and typically you will have one central server maintaining master data which will be distributed to multiple downstream operation system (Some thing where MDM fits). So ideally when distributing master data using ALE you will be sending it to multiple systems.
       5. Distributing transactional data PO/SO etc is same as sending fax/print which fits with the requirement of message control.
    Hope this helps,
    Regards,
    Sanjeev

Maybe you are looking for