Master data and impact of loads from transaction data

If I have the following scenario:
1. Load to cube which contains a dimension which has an info object 0CUSTOMER
2. The transaction data is being loaded from a non SAP system and 0CUSTOMER is loaded on a daily basis from SAP system
3. As part of the transaction load some Customers have a value which is not existing in the 0CUSTOMER info object
Do these new customers also get loaded to the 0CUSTOMER master data i.e. the SIDs are added to the SID table for 0CUSTOMER or is it just loaded as part of the cube and stored in the specific dimension
Thanks

Hi,
they are also automatically loaded to 0customer, as the sid of that record is needed in the cube. Possibly, if there is no relation to your R/3 customers, you have some initial records in your customer master (means all attributes as well as the texts are initial).
regards
Siggi

Similar Messages

  • Master data tables with unwanted records from transaction data upload

    Hi Friends,
      I have a master data table for infoobject 'C' with compounding characteristics 'A' & 'B'.  I upload this master data with values given below:
        <i><u> A,              B,              C,           Short text,                        Long text</u></i>
           <b>  <b>P,          10,           BBB,         Apple,                              Big Apples
             Q,             20 ,           XYZ  ,       Tomatoes    ,                    Red Tomatoes</b></b>
      When I load data into ODS from a source system, I may not necessarily have data for all these 3 fields in these transaction record.  Example:
      <i><u>     A,                B,             C,             D,            E</u></i>    
         <b> P                -1            FFF</b>          20           30            
         <b> Q                10           GGG        </b> 10           40
       The problem is when I upload the above transaction data, it populates the <b>master data table</b> too with these two new records <b>1 -1 FFF</b> and  <b>2 10 GGG</b>, which I would like to avoid.
       Is there any way?
       Will assign full points to anyone who helps me here.
       Thanks,
       JB

    Hi JB,
    If you want to load transactional data and still want to prevent the population of the master data table, I don't think it is possible, as it is goes aginst the data consistency in the warehouse.
    However, if you can afford not to load transactional data for such cases, you can activate referential integrity check for the infoobject C. Then neither transactional data nor masterdata enter the datawarehouse until you maintain masterdata yourself for the infoobject C.
    hope this helps.

  • SAP Best Practice: Problems with Loading of Transaction Data

    Hi!
    I am about to implement SAP Best Practices scenario "B34: Accounts Receivable Analysis".
    Therefore I load the data from SAP ERP IDES system into SAP NetWeaver 2004s system.
    My problems are:
    when I try to load the Transaction data for Infosources 0FI_AR_4 and 0FI_AR_6 I get the following errors/warnings:
    when I start the "schedule process" the status getting still "yellow 14:27: 31(194 from 0 records)"
    On the right side I see some actions that are also "yellow", e.g. "DataStore Activation (Change Lo ): not yet activated".
    As a result I cannot see any data in tcode "RSRT" executing the queries "0FIAR_C03/...".
    The problems there
    1) Input help of the web template of query don't contain any values
    2) no data will be shown
    Can some one help me to solve this problem?
    Thank you very much!
    Jürgen

    Be in the monitor window where u got the below issue
    when I start the "schedule process" the status getting still "yellow 14:27: 31(194 from 0 records)"
    and go to environment in the menu options TransactRFC--->in the sourcesystem...
    give the logon details and enter and from there give the correct target destination as ur BI server and execute...
    if u find some idoc's pending there push that manually using F6..
    and come back to ur load and refresh....
    if still it doen't turn green u can manully change status to red in STATUS tab and come to processing tab and expand ur processing details and right click on ur data packet which was not yet updated and select manual update...
    it shows busy status and when it comes out of that once again refresh...
    rgds,

  • OFI_AP_4  Delete all data and REINIT the load to FIAP ODS and Cube

    Hi bw GURUS I need a quick help, I have requirement to Delete all data and REINIT the load to FIAP ODS and Cube  i have data source ofi_ap_4, please tell me what should i do as i know LO extraction but whats is the transaction code to run the set job for FIAP like for Purchasing I have OLI3BW now tell me what will be the steps to reinit the load to fiap ods and cube.
    i know this much as below
    Delete the data from cube and ODS
    Delete the Init in Cube AND ods NOW reinit i mean load the init, but now problem is before doing that i need to run the set up job or tell me what are the steps please
    its very important since i am doing in production
    thanks
    Soniya

    You are right by your approach to update from PSA to ODS as u have wrritten before. Just got to PSA tree select the PSA request and choose update Immediatly or scedule for update.
    (and if u have 0FI_GL_4 but u r not using it ..it is fine. I  was telling that in case u have that and you are trying to use somethin with 0FI_AP_4)

  • AR aging report--from transaction date

    Hi
    Presently the 7/4 bucket aging report shows current and invoices no of days past due i.e. 31 days due past due date.
    Is there a report or any way by which this aging report can be generated to show aging from transaction date.
    Thanks
    Ninad

    Hi ninad
    Aging Reports by default are generated based on Transaction date only and not based on GL Date. Please explain your requirement clearly
    Regards
    Sivakumar

  • How to get master data records that do not have transaction data in a query

    Hi,
    How to get master data records that do not have transaction data in a query output. Can we create a query or any other way to get the master data records that do not have transaction data?

    Hi,
    Create a multiprovider which includes transactional data target and master data info object. Make sure that identification for this master data info object is ticked on both the provider.
    Create report on this multiprovider , keep the master data info object in rows , and now you should able to see all the values which are there in master data info object irrespective of transaction happened or not .
    Next you may create condition showing only zero keyfigure values , ie. master data without any transaction.
    Hope that helps.
    Regards
    Mr Kapadia

  • Why should we load header data first and then we load item level data?

    Hi BW guru`s,
    I have small confusion about the data loading.
    Why should we load header data first and then we load item level data?
    Is there any particular reason?
    Scenario: 1st I have uploaded 2LIS_11_VAHDR sales document header data from R/3 to BW by using LO-Cockpit extraction. Then I loaded 2LIS_11_VAITM.This is the norma procedure which we use to follow.
    I have a question. If I load 2LIS_11_VAITM data 1st from R/3 to BW and then I will load 2LIS_11_VAHDR by using LO-Cockpit extraction. what will happen?
    Regards,
    Venkat
    Edited by: VENKAT BOORUGADDA on Aug 12, 2008 11:51 AM

    There is no difference in doing it the other way.
    The load sequence will come into play only during activation where if you map the same fields from the two datasources, you might want to have the previous value overwritten by data from the next datasource.
    That is when you should care about loading one datasource before the other.
    To your question it is not arule that header data should come first.

  • Loading from 3 data sources

    hi guru's,
    Iam having a scenario that I have to load the data from 3 different Data sources into single Z-Cube my question to you is what are the advantages and disadvantages for this scenario.
    regards
    lakshmi

    Hi,
    The scenario is common and only thing is
    1) How are you managing the loads,records are not getting doubled ...as cube has no overwriting facilyt and every records sits on the top of other one...doesnt amtter if they are same or different.
    So you have to make sure that you are getting accurate information from the data source.
    2)Delta from the data source should be accurate and should be in accordane with the struture of the infocube.
    Not all the kinds of deltas can be loaded into the cube and has to be loaded through DSO.
    3)You have to take care of delta management issues as well in case of data load failures as well which becomes tough to manage if there is no DSo in between.
    I would always advice to go for any DSO in betweem if you are loading to the cube.
    Thanks
    Ajeet

  • I am resurrecting my old iMacG5 for my grandson. Appleworks has question mark in dock and does not load from bundled software on original installation discs. Can anyone help?

    I am resurrecting my old iMacG5 for my grandson. Appleworks has question mark in dock and does not load from bundled software on original installation discs. Can anyone help?

    I'm sorry no-one has answered you yet - I'm afraid I don't know the answer to the question: but I wanted to put into your mind that starting up with AppleWorks at this stage isn't a good idea. The program was 'end-of-lifed' years ago, it won't run on Lion, and some of the document types created in it - Databases in particular - cannot be opened at all by any other program whatever. Word Processing documents can be opened in Pages, and Spreadsheets in Numbers (with some reservations in both cases), but building up a collection of AppleWorks documents at this stage is going to lead to problems later on.
    It's difficult to know what to recommend for a PowerPC machine that would still be compatible with modern ones. An older version of Microsoft Office, perhaps, if you can find one which will run on PPC; but most of the modern Word Processors are Intel only. This is something you would need to investigate.

  • How to export a data as an XML file from oracle data base?

    could u pls tell me the step by step procedure for following questions...? how to export a data as an XML file from oracle data base? is it possible? plz tell me itz urgent requirement...
    Thankz in advance
    Bala

    SQL> SELECT * FROM v$version;
    BANNER
    Oracle DATABASE 11g Enterprise Edition Release 11.1.0.6.0 - Production
    PL/SQL Release 11.1.0.6.0 - Production
    CORE    11.1.0.6.0      Production
    TNS FOR 32-bit Windows: Version 11.1.0.6.0 - Production
    NLSRTL Version 11.1.0.6.0 - Production
    5 rows selected.
    SQL> CREATE OR REPLACE directory utldata AS 'C:\temp';
    Directory created.
    SQL> declare                                                                                                               
      2    doc  DBMS_XMLDOM.DOMDocument;                                                                                       
      3    xdata  XMLTYPE;                                                                                                     
      4                                                                                                                        
      5    CURSOR xmlcur IS                                                                                                    
      6    SELECT xmlelement("Employee",XMLAttributes('http://www.w3.org/2001/XMLSchema' AS "xmlns:xsi",                       
      7                                  'http://www.oracle.com/Employee.xsd' AS "xsi:nonamespaceSchemaLocation")              
      8                              ,xmlelement("EmployeeNumber",e.empno)                                                     
      9                              ,xmlelement("EmployeeName",e.ename)                                                       
    10                              ,xmlelement("Department",xmlelement("DepartmentName",d.dname)                             
    11                                                      ,xmlelement("Location",d.loc)                                     
    12                                         )                                                                              
    13                   )                                                                                                    
    14     FROM   emp e                                                                                                       
    15     ,      dept d                                                                                                      
    16     WHERE  e.DEPTNO=d.DEPTNO;                                                                                          
    17                                                                                                                        
    18  begin                                                                                                                 
    19    OPEN xmlcur;                                                                                                        
    20    FETCH xmlcur INTO xdata;                                                                                            
    21    CLOSE xmlcur;                                                                                                       
    22    doc := DBMS_XMLDOM.NewDOMDocument(xdata);                                                                           
    23    DBMS_XMLDOM.WRITETOFILE(doc, 'UTLDATA/marco.xml');                                                                  
    24  end;                                                                                                                  
    25  /                                                                                                                      
    PL/SQL procedure successfully completed.
    .

  • Problem with getting current date and time using oracle.jbo.domain.Date

    I`d like to get current date and time using oracle.jbo.domain.Date method getCurrentDate(), but it always return current date and 12:00:00. I also need to get the current time.

    I think you should use java.sql.Timestamp domain.
    (And set database type to TIME or DATETIME.)
    Jan

  • Has anyone else had a problem with data? after updating my phone to the new update, i have all this data, and keep getting notifications that my data is getting high. i never had this problem before.

    has anyone else had a problem with data? after updating my phone to the new update, i have all this data, and keep getting notifications that my data is getting high. i never had this problem before.

    fair enough.  No need for any unnecessary posts either.  You issue had been addressed ad nauseum already in this forum had you bothered to search this forum (as forum etiquette would dictate) before posting.
    In any case, I hope that your problem was solved.

  • Master data from transaction data

    Hi All,
    I am getting data from R/3. I am also getting data from another web based application which takes its basic master data from R/3.
    Now from the transaction data from the web based application(coming to BW as flat file), I am getting a field Material Class ID which desnot exist in R/3. Can I get the data of this field as master data in BW and that too as mapped to the Material from R/3.
    That is to say: can I get the Material Class ID data as master data in relation to Material from R/3?
    If anyone has any idea regarding this, please help me. Thanks in advance.
    Regards,
    Vinay

    Not sure i understand you properly..but,
    If Material and Material Class ID has the same 'meaning' then with conversions or without conversions.. you can store in the same 'material'.
    But if they are differed and you have got a reference data availble you can create a attribute as mentioned and map material and material class id. So
    1) Create only Material Class ID and LOAD this master data first from flat file.
    2) Enhance the material master data to add this as new attribute.
    3) Load the material master data and while loading fetch the material class id in the routine with the reference rules..
    Regards,
    Hari

  • Conversion of Sales Center (0Plant) from Transactional data to master data.

    Hi All,
    We have an requirement like this.
    Currently in my reports sales center (0Plant) pulling data from tranansactional data.Now due to some problem this transaction data needs to be changed to pull data from master data from 0PLANT attribute of ZOUTNBR will give me the routine to pull data from master data tables.
    Regards,
    Shivu.

    Ans

  • How to edit PSA and continue the load with corrected data from PSA.

    Hi,
    Can you guide me on how to edit data in PSA and then continue a load?
    i.e. if you can provide me with the steps on how to fix a load problem if a load fails. The load fails and Monitor shows red, so how do I fix them in the PSA and then allow the load continue with the corrected data into the data target.
    I will appreciate the details steps involved.
    Thanks

    Hi
    First select that request then turn to Red and delete the Request , delete the qeruest then system allow to edit the psa data ,selete psa then selet packet then selet record double click modify the data, dont forget after completing save the data. after competing modification then select psa right click select- start update immediatly......
    thanks
    ram

Maybe you are looking for

  • Why wont my iPhone 4 sync certain playlists? iOS 5.1.1 and iTunes 10.6.3

    Hello! Recently, my iPhone 4 wont sync any new playlists. When I tick the box for the playlist I want to sync, the data boxes move and the song count increases, but when I click sync, it does nothing and the playlist isn't visible on the iPhone. So f

  • MIGO IN THIRD PARTY

    IN THIRD PARTY TRANSACTION CAN WE HAVE STASCTICAL MIGO FOR ANALYSIS PURPOSE? IF YES PROVIDETHE STEPS FOR CONFIGURATION.

  • FUNCTIONAL MODULE CONVERT_OTF GIVING JUNK DATA IN UNICODE SYSTEM

    Hi Gurus, After Unicode conversion of our QAS server we are facing problem in PDF  conversion through FM   CONVERT_OTF . as it is giving  all junk data . We checked the spool and in spool it is showing correct output. We checked the FM in debug mode

  • How to start streaming from server side after applet initialized on client

    Hi, I am using JSP for on demand streaming server. I have included an applet in jsp page which start new player on client side after on streamreceive event. But my problem is how to give call to server that applet on client side has been initialized

  • SD Card in Drive Mode

    When I use Drive mode on Windows XP, the SD card, which previously was visible, is not showing up on my computer, only the Life Drive shows up. Thanks Post relates to: LifeDrive