Best Design for Master data

hi,
  I would like to get your suggestions on the data modelling:
  I have material, mat_plant and mat_sales.I need to create reports on these master data. But when i created a infoset on those 3 master data, one of my key figure "amount" is becoming 10 times like 1 line for material 10 lines for mat_plant and as a result the report is showing the key figure values 10 times of original value.
Please advise what would be the best way to go...
Thanks in advance...

Adi,
   You will get 10 lines only. I guess you have 10 plants for that perticular Plants or sales.
It should come reporting requirements. Please let us know what do you want to report?
relation between 3 master data objects is... 1:N:N (0Material:0MAT_PLANT:0mat_sales).
You can't avoid it... restrict by plant....
Nagesh Ganisetti.

Similar Messages

  • Best Practice for Master Data Reporting

    Dear SAP-Experts,
    We face a challenge at the moment and we are still trying to find the right approach to it:
    Business requirement is to analyze SAP Material-related Master Data with the BEx Analyzer (Master Data Reporting)
    Questions they want to answer here are for example:
    - How many active Materials/SKUs do we have?
    - Which country/Sales Org has adopted certain Materials?
    - How many Series do we have?
    - How many SKUs below to a specific season
    - How many SKUs are in a certain product lifecycle
    - etc.
    The challenge is, that the Master Data is stored in tables with different keys in the R/3.
    The keys in these tables are on various levels (a selection below):
    - Material
    - Material / Sales Org / Distribution Channel
    - Material / Grid Value
    - Material / Grid Value / Sales Org / Distribution Channel
    - Material / Grid Value / Sales Org / Distribution Channel / Season
    - Material / Plant
    - Material / Plant / Category
    - Material / Sales Org / Category
    etc.
    So even though the information is available on different detail  levels, the business requirement is to have one query/report that combines all the information. We are currently struggeling a bit on deciding, what would be the best approach for this requirement. Did anyone face such a requirement before - and what would be the best practice. We already tried to find any information online, but it seems Master data reporting is not very well documented. Thanks a lot for your valuable contribution to this discussion.
    Best regards
    Lukas

    Pass a reference to the parent into the modal popup. Then you
    can reference anything in the parent scope.
    I haven't done this i 2.0 yet so I can't give you code. I'll
    post if I do.
    Oh, also, you can reference the parent using parentDocument.
    So in the popup you could do:
    parentDocument.myPublicVariable = "whatever";
    Tracy

  • Best design for historical data

    I'm searching a way to design some historical data.
    Here my case:
    I have millions of rows defining a population, each on 20-30 different tables.
    Those rows are related to one master subject area, let say it is a person.
    The details table are defining status in a period of time (marital status, address, gender (yes that may change in some cases!, program, phase, status, support, healthcare, etc). They have all two attributes that define the time period (dateFrom, dateTo), and one or many attributes that define the status of the person (the measure).
    I know we need a weekly situation for this population since 1998 to yet.
    Problems are those:
    1) the population we will analyze will count some 20 different criteria's (measures), may be we will divide those in some different datamarts to avoid to much complexity
    2) we will drill down from year to week, and will accomplish comparison like year to year, year to previous year, year to date, etc at each level of time hierarchy (year, quarter, month, week).
    The question is :
    1) do we need to transform our data at a week level for each person to determine the status at this level (the data may be updated and will cause this transformation to be refreshed) ?
    1) do we need to aggregate for each level because mixed situation exists due to fact that more than one status may exists for same person at a given time period (more than one status may exists in march by example, how to interpret it ?), that will cause to some logic to be applied, isn't ?
    We will be glad to hear some recommendation/idea about this !
    Thank You

    I would try to get some exact user requirements, with the dataset you have described there are millions of combinations of answers that can be defined and it will be difficult for you to fulfill them all in one place, especially if there are slowly changing dimensions involved.
    The user requirements will determine what levels you transform the data at.

  • SAP BW 3.5 APD for master data

    Gurus,
    We are having issues with SAP BW 3.5 APD transformation before full upload for Item master data.  Due to the high volume encountering short dumps. We cannot extend memory space to fix this short dump.  What we are trying to do is initilize the values before the full load to make consistent with SAP so that BW item cost values is in sync with SAP.  There was no issue in development and only in test system.
    So work was moved to test.
    Is there any suggestion for this problem or alternative to something else?  Please help.
    Thanks,
    Di

    Hi Abhishek,
    We need to correct values of custom cost related fields of master data in BW that pull data from SAP.  If any Item  with cost field in SAP is not maintained any longer, we found the solution to initilize every record with APD before doing full upload in BW to be in sync. Version  is BW 3.5. We are facing performance issue.  So tried different ways like delete all the records that is already zero before reading data from the internal table.  We still get short dump when process chains runs with APD program. I am not sure how to pull data in smaller chunks with filters here.  Any suggestions please? Design is master data, filter columns, routines and master data. Source and target are the same.
    Thanks,
    Di

  • Why to use chnage  pointers for  master data idoc why not  the  change idoc

    Hi Gurus,
    I have  one doubt about  Idoc.
    When changes to master  data has to be sent Change pointers are configured and  used. (CDPOS & CDHDR). In case of  the  transaction data change change idoc  is  used as is the case with orders (we  use ORDCHG for  ORDERS message type to send  the  chnage details to a  order which was already sent to other system.
    Why we can't use  change  idoc or  message type whatever, instead of  Chnage pointers in case of  master data or  vice versa in case  of transaction data.
    Your valuable  input  will be rewarded with suitable  points!!
    -B S B.

    Hi,
    It was a good question ... SAP seems never designed the change pointer to handle transactional data IDocs. Only for master data IDocs distribution purposes.
    It may due to volume data ... the transactional data can change many times within a day compare with master data which hardly change or once a while.
    Regards,
    Ferry Lianto

  • Requirements for master data template

    HI,
    Please let me know the basic requirements for master data template. What all the infotype to be involved in collecting master data template.
    Thanks and regards,
    Aylwin

    Hi Alwyn,
    If you dealing with india payroll definately Yes subject to you business process again.
    Even though it is indian payroll some item may not applicable
    For exmple other income resourece will be  if you go to manufaturing industries you may face problem even though they got some they don't want reveal it to the employee.
    in such that infotype is not required as per your business process.
    Some states like UP there is no profession tax. Genrally we say profession tax need to be configure you need that master data.
    If such case that might be wrong for that you need to give all the info
    use PA30 t code go in side check what are the mandatory fields and take them as master data fields and put it in template.
    If your client is able to give other field you can very well add them no harm in it.
    But make sure you filled mandatory fields
    That is the reason i was asking you to tell which coutry payroll you using
    Best Regards

  • Upload Image for Master Data Product - SAP CRM GUI

    Hi,
    I see in DEMOS image for master data product TX  COMMPR01, when you open transaction image is in top of the screen with the other data.
    I create a new product and add an image in Tab Document, however image only is load when I pushed botton preview.
    Some idea?
    Regards,

    Hi Wilbert,
    Well it should work if you have loaded the image at the right place in the Tab Document. Please verify the same.
    What I did to check this :-
    Go to Tab Document -> follow the tree structure Document -> Picture -> Full Size.
    The Right click on Full Size ->create -> Document -> Import. And I imported a picture form my desktop. I did the same for Thumbnail. I can see the image in the top right corner of the screen in TX COMMPR01.
    Also In order to view the complete picture I just Right click on the file name -> Document -> Display and I can see the full picture.
    <b>Reward points if it helps!!</b>
    Best regards,
    Vikash.

  • DBconnect for master data

    Hi everbody,
    we are extracting customer master data from an Iseries-system cia DBconnect.The relevant info object 'Customer' has attributes and texts.
    Generation of the data source and loading of data worked without problems for the info object attributes (after generating the data source in the source system tree with setting 'master data attributes for 'data source type').
    Now i would like to extract from the same data source the info object -texts. Problem is that the system does not allow to assign for the same info source 'customer' the source system for a second time: 'No data sources for source system 'Iseries' available/message RSAR383.
    This is although I did generate the data source (below the source system tree) a second time with the setting 'master data texts' (for data source type).
    While it is normal that I can connect the data source only once to an info source for transaction data, I would expect to be able to use the data source twice for master data when differentiating between 'text' and 'attributes'.
    Has anybody an idea on this topic?
    Thanks and best regards
    Marcus

    Hi Joe, Hi Uday,
    thanks for the further clarification on the topic. The problem is now solved following your recommendations. From the source system we receive attributes and texts in one view. The corresponding data source was generated with type  'texts'. I had to manually enter the text fields 0TXTSH, 0TXTMD, 0TXTLG in info source and had to maintain mapping in the communication structure.On info provider level it was possible to create two update rules (type attribut and type text). Then-using one info package- the upload of attributes and texts worked fine.
    Thanks again for your support
    Best Regards
    Marcus
    I tried to reward due points, but I receive a message Rewarding the message failed
    Please bear with me, I will figure that out and reward the points.
    Message was edited by: Marcus Theviot

  • DTP for Master data

    Hi Guys,
    How to extract master data and hierarchies from ECC into BI infoobjects?
    Do we follow the same procedure for loading master/hierarchy data as for transaction data - i.e. Load into PSA and then create Transformations and DTPs to load into infoobjects? If so,  what is the best approach as far as Full/Delta loads.
    1. Is it ok to do a full load into PSA (using infopackage) and then do a full load always (The process chains for master data would be running nightly).
    or
    2. Should we have 2 infopackages (full and delta) into PSA and then multiple DTPs. (The process chains for master data would be running nightly).
    Regards.

    Indicator: Only Get Delta Once
    Source requests of a DTP for which this indicator is set are only transferred once, even if the DTP request is deleted in the target.
    Use
    If this indicator is set for a delta DTP, a snapshot scenario is built.
    A scenario of this type may be required if you always want an InfoProvider to contain the most up-to-date dataset for a query but the DataSource on which it is based cannot deliver a delta (new, changed or deleted data records) for technical reasons. For this type of DataSource, the current dataset for the required selection can only be transferred using a 'full update'.
    In this case, a DataStore object cannot usually be used to determine the missing delta information (overwrite and creation of delta). If this is not logically possible because, for example, data is deleted in the source without delivering reverse records, you can set this indicator and perform a snapshot scenario. Only the most up-to-date request for the DataSource is retained in the InfoProvider. Earlier requests for the DataSource are deleted from the (target) InfoProvider before a new one is requested (this is done by a process in a process chain, for example). They are not transferred again during the DTP delta process. When the system determines the delta when a new DTP is generated, these earlier (source) requests are seen as 'already fetched'.
    Setting this indicator ensures that the content of the InfoProvider is an exact representation of the source data.
    Dependencies
    Requests that need to be fetched appear with this indicator in the where-used list of the PSA request, even is they have been deleted. Instead of a traffic light you have a delete indicator.
    Get Data by Request
    This indicator belongs to a DTP that gets data from a DataSource or InfoPackage in delta mode.
    Use
    If you set this indicator, a DTP request only gets data from an individual request in the source.
    This is necessary, for example, if the dataset in the source is too large to be transferred to the target in a single request.
    Dependencies
    If you set this indicator, note that there is a risk of a backlog: If the source receives new requests faster than the DTP can fetch them, the amount of source requests to be transferred will increase steadily.

  • About delta loading for master data attribute

    Hi all,
    We have a master data attribute loading failed which is a delta loading. I have to fix this problem but I have two questions:
    1. how can I find the those delta requests because I need to delete all these failed requests first, am I right ? Master data is not like cube or ods that we can find the requests in Manage, for master data how can we find them ?
    2. Could you please let me know the detailed procedures to perform this delta loading again ?
    Thanks a lot

    Hi...
    1. how can I find the those delta requests because I need to delete all these failed requests first, am I right ? Master data is not like cube or ods that we can find the requests in Manage, for master data how can we find them ?
    Look.....for master data.....no need to delete request from the target..........just make the status red.......and repeat the load.....But problem is that master data sometimes does'nt support Repeat delta..........if u repeat......then load will again fail with Update mode R.........in that case u hav to do re-init.......
    1) delete the init flag.......(In the IP scheduler >> in the top Scheduler tab >> Initialization option for source system)
    2) Init with data transfer(if failed load picks some records)..........otherwise .....init without data transfer.....if the last delta failed picking 0 records.......
    3) then Delta.......
    2. Could you please let me know the detailed procedures to perform this delta loading again ?
    1) Make the QM status red.........to set back the init pointer.......
    2) Repeat the load.....
    After that.........if again load failed with Update mode R.....
    1) delete the init flag.......
    2) Init with data transfer(if failed load picks some records)..........otherwise init without data transfer.....
    3) then Delta.......
    Regards,
    Debjani.....

  • Unable to generate "Export data source" for Master data infoobject

    Hi Gurus,
    I'm in BW 3.5
    I need to generate "Export data source" for a master data info object.
    I followed below steps:
        I ticked the checkbox "Characterist. is export data source" in the master data Info object maintenance screen.
        In the Infosource area, for the desired Infoobject --> Right click --> Generate export datasource --> a dialog box displays saying "The generation of DataSource  was successful"
    Problem:
    Even after refreshing, etc.. I couldn't find the generated datasource 8**** for the above master data. I also tried to replicate the datasource from the myself source system, but couldn't find it even there. I don't know where the generated export data source is hiding ???
    Could anyone please help ????
    Thanks,
    Venkat

    Hi Venkat,
    I guess that after creating an export datasource for infoobject this DS should appear in the RSA1-Source systems - MYSELF - datasource overview (with name starting with 8). Since it's a datamart, it should appear also as system generated infosource. And in this case you would be able to assign in URs of data target the appropriate infosource (make sure that in Infoproviders tab of RSA1 you have inserted your infoobject as a data target).
    However, there are some bugs related to export data source. CHeck if you see you export data source in RSA6.
    If you don't see it - consider applying OSS Note #816892 - "30BSP26: Export DS for master data is not generated".
    Note #876845 - "30BSP29: InfoObject change: Export DataSource not adjusted" is also may be useful.
    -Vikram

  • Error while creating Data Source for master data attributes

    Hi BI Experts,
    Well its been some time for me that I have been part of Extraction in BI.I primarily handled reporting in my last assignments.
    I was trying extraction with flat files in SAP BI 7(new to sap bi 7 but very much familiar with BW3.5) but failed in the activity during master data attributes and text upload in infoobject (say IOSP_Mat).
    Here is the procedure that I did after creation of characteristic IOSP_Mat.I created a source system for flat file followed by data source for Master data attributes, i selected all the parameters correctly.i.e. csv file format, data seperator as   ,
    and other settings, now when i am trying to look at the proposed data in the next tab using Load example data.its not showing the desired result.The columns that I have maintained in Flat File is  MAT_NUMBER and MAT_NAME (with say 100 data in the file)
    same is the result when I am trying to load the text data too columns maintained are
    (LANGUAGE MAT_NUMBER Short Description)(same 100 data).
    now i used to rsa1old transaction  to upload the file using 3.5 version.i created info source for master data/text/hierarchies for  IOSP_Mat
    now when trying to upload it using info package for master and text data,I observe its(the data) not maintained in the characteristic IOSP_Mat.
    When I monitored ,I figured the data has not been even uploaded to the PSA level.
    Can you BI experts tell me the answer for this.
    Thanks,
    Srijith

    apologies to all of you for late response,
    was busy with some other activities.
    I don't remember the exact message,but I remember it was not loaded to even the PSA level.I will try it again and post the exact message.
    Thanks again for your quick response.
    Once again sorry to all of you for my late response
    Thanks,
    Sri

  • Creation of OLAP Variable for Master data

    Dear BW Experts,
    How to Create OLAP Variable for Master data  in SAP BW? What is the use of OLAP Variable?

    The entry column Type (Variable) on the Select Data tabpage indicates a variable time frame in which the data should be requested. The variables are not replaced by concrete values until a request is made. There are eight options:
    Variable            Data request        Description
    0
                          Yesterday
                                            (12am – 12am)
    1
                         Last week
                                            (Monday-Sunday)
    2
                           Last month
                                            (1. (First day - last day of previous month)
    3
                           Last quarter
                                           (1. (First day - last day of previous quarter)
    4
                           Last year
                                          (01. (January 1 – Dec 31 of previous year)
    5
                      User time selection
                                               See below
    6
                       ABAP routines
                                               See below
    7
                         <b>OLAP variables</b>
                                                See below
    <b>Time selections with the help of the variables 0 to 4.</b>
    The variables 0-4 only refer to date fields of the data type DATS.
    <b>User time selection</b>
    Type 5 gives you a free selection of all fields.
    If you select this variable for a field, confirm the selection, and then choose Detail for Type, an additional dialog box appears. In this dialog box you can freely limit the values of the fields and determine the next period value and the number of periods until a repetition.
    Variables are placeholders for values and are not replaced by concrete values until a data request is made. You can also use a Variable for selections when requesting data for an InfoObject, if you choose type 7, confirm the selection, and then choose Detail for Type.

  • Best practice for heirachical data

    First off, I have to say that JMX in Java 6 is terrific stuff. Bundling jconsole in with Java has made JMX adoption so much easier for us.
    Now, to my question. We have read-only hierarchical data (think a DOM tree) that we would like to publish via JMX. What is the best practice? We see two possibilities:
    1. Publish each node of the tree with it's own object name and type. This will allow jconsole to display the information in the tree control.
    2. Publish just the root of the tree with an object name and type and then use CompositeType to describe the nodes of the tree. This means you look at the tree in the "Attribute Value" panel of jconsole.
    Is there any best practices for such data? We have implemented #2 and it works but we are wondering if long term this might lead to unforeseen consequences.
    Thanks in advance.
    --Marty                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                       

    Path,
    I did go with #1 and it worked out great. Every node in our tree is an ObjectName node. Works very well for us.
    --Marty                                                                                                                                                                                                                                                                   

  • Best practice for sharing data with model window

    Hi team,
    what would the best practice for sharing data with a modal
    window be ? I use a modal window to display record details from a
    record list, but i am not quite sure how to access the data from
    the components in the main application in the modal window.
    Any hints would be welcome
    Best
    Frank

    Pass a reference to the parent into the modal popup. Then you
    can reference anything in the parent scope.
    I haven't done this i 2.0 yet so I can't give you code. I'll
    post if I do.
    Oh, also, you can reference the parent using parentDocument.
    So in the popup you could do:
    parentDocument.myPublicVariable = "whatever";
    Tracy

Maybe you are looking for