Loading Master Data with Flexible Update

Hi,
I have created an Infoobject - Business Partner (Master Data bearing characteristics). The attributes to this are Region, Sales Person, Industry code which are master bearing characteristic as well.
I need to load data to the business partner from a csv file.
The layout of the CSV file is -
BP number, BP text (long Text), Region Code, Region Desc (Med Text), Sales Person Code, Sales Person Desc (Med Text), Industry Code, Industry description (Med Text).
How do I define the infosource to load this data.
Appreciate any help with this.
Thank You,
Prashanth

Hi,
First of all, I need to note that there are two kind of Infosources: with direct and flexible update.
If you choose a direct IS, then in the creating of the IS you just enter the name of the infoobject where you are going to load data. The system will create a IS (comm. structure). Enter this IS for changing. The system will propose you the communication structure,  click on a bottom icon “Transfer structure/Transfer rules”, choose as a ‘Source system’ your flat file system. Agree with the system when it asks you to save assignments (up to 3 times). Activate TRs. Then click in the field for Datasource. You may see there another datasources (for texts, attributes and hierarchies). Choose one by one another datasources with their activation.
Now you can create an infopackage for a load – you can choose what kind of package it is going to be – for loading texts, attributes or hierarchies.
Note that in this case the structure of the flat file is proposed by the system and you need just prepare the flat file corresponding to the proposed structure (different for each of 3 possible datasources). Execute infopackage.
If you use a flexible IS, then you may insert into comm. structure the fields that you think you may need in the master data. Note that here you may have not only attributes but TEXTS also. Save a comm. structure. Assign a flat file source system and activate a TRs.
After that in RSA1-dataproviders tab – right click – insert IO as a data target – choose your IO. Refresh the screen. You’ll see up to three data targets. Create update rules for each of them. In URs map the fields in the IS with the fields in URs.
Best regards,
Eugene

Similar Messages

  • Error handling for master data with direct update

    Hi guys,
    For master data with flexible update, error handling can be defined in InfoPackege, and if the load is performed via PSA there are several options - clear so far. But what about direct update...
    But my specific question is: If an erroneous record (e.g invalid characters) occur in a master data load using direct update, this will set the request to red. But what does this mean in terms of what happens to the other records of the request (which are correct) are they written to the master data tables, so that they can be present once the masterdata is activated, or are nothing written to masterdata tables if a single record is erroneous???
    Many thanks,
    / Christian

    Hi Christian -
    Difference between flexible upload & Direct upload is that direct upload does not have Update Rules, direct upload will have PSA as usual & you can do testing in PSA.
    second part when you load master data - if error occurs all the records for that request no will be status error so activation will not have any impact on it i.e. no new records from failed load will be available.
    hope it helps
    regards
    Vikash

  • Master Data InfoSource (Flexible Update)

    Hello Everyone,
    I have a small doubt regarding a Master Data InfoSource (Flexible Update). I see that there are two data targets for this infosource. They are Attribute data target and text data target.
    My question is, do i have to create 2 separate InfoPackages to load the 2 data targets (one each for attribute and text) or a single infopackage will be enough for both.
    I created an infopackage and chose Attribute data target. When i checked the contents, i see that even the text data target is populated with data even though i didn't choose it in the infopackage. Also, i get multiple entries for each record (time period 1000-1899 record and time-period 1900-9999 record). How to get only a single entry.
    Could someone help.
    Thanks.

    Hi Sachin,
    This post should help to clarify why different records exist with different to and from dates:
    Re: Master data
    Hope this helps...

  • Master data at flexible update

    Hi sdn
    I have one dought , Master data overwrites this we all know but if master data is at flexible update level, then can we  load the data daily full update by process chain or not?
    if we  load daily by process chain then what about the requests which are accumulating daily in that particulat master data infoobject at flexible update level?
    pls guide I'm in confusion
    rubane

    Data are also overwritten; the request could be usefull to undo a load and get back to previous master data situation.
    Hope it helps.
    Regards

  • Change communication structure for master data with direct update

    Hi All,
    I am having a problem with a change I want to make to some master data. I have added the attribute to the characteristic, but when I have gone to change the communication structure, its not possible (the add line button is greyed out)
    I can see the new infoobject in the datasource/trans structure, but not in the comms structure (And yes it is in change mode   ).
    The master data uses direct update, and I have read that this causes some hassles in changing the comms str.
    Can someone please give me some steps in doing this??
    Thanks
    Ryan

    Hi Ryan,
    Why u have nothing to map is ?
    U r datasource trasfer structure from the source ssytem has nothing new ..
    1)TO ur Info object u can only write the routing or give any formula or anythign like the trasfer rules oftions that have when u click on the IO icon..
    2) Enhave the structure of the datasoruce that is getting from the  R3 source system...
    3) Replicate it ..
    Then only u will be able to find the extra field at the Trasfer structure... to map to the IO which have been added..
    hope it helps
    regards
    AK

  • Loading master data with date-fields

    Hello,
    we are loading with the extractor 0crm_mktelm5_attr data to the data target 0crm_mktelm (infoobject).
    The fields "from" and "to" are initial in psa. But the Update in the master data abort with the failure:
    "Invalid "to" date '00000000 ' " and "Invalid "to" date '00000000 ' ".
    The data-fields in psa are initial, and in other BW-Systems, we have not problems with the extractor 0crm_mktelm5_attr and initial date-fields.
    Have any body an idea??
    Thanks
    Murat

    Hello Shashank,
    thanks,
    the Infosource.mapping was false.
    regards
    Murat

  • How to load master data with text into infoObject in BI 7.0

    Hi all;
    I have an infoobject 0employee ( this is for employee number ,8 character long), with medium text option.
    I would like to load Employee name as a text option for employee number.
    How do I load this mater data into this 0employee object,
    Please help.
    I did look for something similar in the forum and there are many of them, but I couldn't find anything that for BI 7.0 and really clear help for me. Sorry to ask a question that been asked before.
    thanks.
    Raiine

    hi rainnie,
    check this link t oload master data in BI 7
    Re: BI 7.0 flat File extraction
    as you are using SAP defined infoobject you can start from the creation of datasource.
    text format for the flat file is
    LANGUAGE,INFOOBJECT,TEXT
    EN,0EMPLOYEE,RAINNIE
    You have to create ,infopakage, transformation ,dtp for both the attribute and text.
    hope this helps
    cheers
    Sunil

  • Loading Master Data with compounding attributes

    Hello All,
    I have a characteristic, "Commission Code", compounded to "Product" and "Commission Code" is an attribute of the InfoProvider, "Sales Rep".
    Now, I'm not able to load data into the master data for "Sales Rep" from a flat file because as I get the same sales rep with two different commission codes I get an error caused by "duplicated keys":
    below an extract of the flat file which goes in error :
    SalesRep (key)     ///     CommissionCode     ///     Product
    1000                    ///       10%                      ///    ProdA
    1000                    ///        20%                      ///     ProdB
    how to load the master data in the proper way ? Where I'm wrong ?
    Thanks in advance.

    Hi ,
    Sales rep infoobject should be compounded by commision code. I think you will get the compound of commision code (product) as well in the compound.
    This will make the records unique and you would be able to load same sales rep with two different commision code.
    So your key then would be sales rep, commsion code and product together.
    Hope it helps,
    Regards,
    Sunmit.

  • How to laod Master Data using Flexible Update

    Hi,
    I have created 35 Characteristics and now i need to create a structure using 35 characteristics.
    I am unable to use the Direct Update because it is asking the Reference Object and Attributes
    But the 35 characteristics are not attributes.
    So how to create the structure and load the data into all 35 characteristics.
    Client is sending me the single file where it contains 35 fields.
    Thank you.

    Murali,
    Master data object process type is not there because in BW, master data is conformed or is shared by other infoproviders. Due to this dependency, master data loading is a bit more complex. You need to drop all the data from all the infproviders which are using the master data you want to delete.
    -Saket

  • 36 duplicate record  found. -- error while loading master data

    Hello BW Experts,
    Error while loading master data
    ( green light )Update PSA ( 50000 Records posted ) No errors
    ( green light )Transfer rules ( 50000  „³ 50000 Records ): No errors
    ( green light )Update rules ( 50000 „³ 50000 Records ): No errors
    ( green light ) Update ( 0 new / 50000 changed ): No errors
    ( red light )Processing end: errors occurred
         Processing 2 finished
    ƒÞ     36 duplicate record found. 15718 recordings used in table /BIC/PZMATNUM
    ƒÞ     36 duplicate record found. 15718 recordings used in table /BIC/XZMATNUM
    This error repeats with all the data-packets.
    what could be the reason of the error. how to correct the error.
    Any suggestions appreciated.
    Thanks,
    BWer

    BWer,
    We have exactly the same issue when loading the infoobject 0PM_ORDER, the datasource is 0PM_ORDER_ATTR.
    The workaround we have been using is 'Manual push' from the details tab of the monitor. Weirdly we don't have this issue in our test and dev systems, even in Production, this doesn't occur somedays. We have all the necessary settings enabled in the infopackage.
    Did you figure out a solution for your issue, if so please let me know and likewise I will.
    -Venkat

  • What are the tables will update while loading Master data ?

    Hello Experts,
    What are the tables will update while loading Master data ? And requesting you to provide more information about Master data loading and its related settings in the beginning of creation infoobjects. 

    It depends upon the type of Master data u r loading....
    In all the master data loadings, for every new value of master data an SID will be created in the SID table /BI*/S<INFOOBJECT NAME> irrespective of the type of master data.
    But the exceptional tables that get updated depending on the type of master data are.....
    If it is a time Independent master data then the /BI*/P<INFOOBJECT NAME> table gets updated with the loaded data.
    If it is a time dependent master data then the /BI*/Q<INFOOBJECT NAME> table gets updated with the loaded data.
    If the master data is of time Independent Navigational attributes then for every data load the SID table will get updated first and then the /BI*/X<INFOOBJECT NAME> table gets updated with the SID's created in the SID table (NOT WITH THE MASTER DATA).
    If the master data is of time dependent navigational attributes then for every data load the SID table will get updated first and then the /BI*/Y<INFOOBJECT NAME> table gets updated with the SID's created in the SID table (NOT WITH THE MASTER DATA).
    NOTE: As said above, For all the data in P, Q, T, X, Y tables the SID's will be created in the S table /BI*/S<INFOOBJECT NAME>
    NOTE: Irrespective of the time dependency or Independency the VIEW /BI*/M<INFOOBJECT NAME> defined on the top of /BI*/P<INFOOBJECT NAME> & /BI*/Q<INFOOBJECT NAME> tables gives the view of entire master data.
    NOTE: it is just a View and it is not a Table. So it will not have any physical storage of data.
    All the above tables are for ATTRIBUTES
    But when it comes to TEXTS, irrespective of the Time dependency or Independency, the /BI*/T<INFOOBJECT NAME> table gets updated (and of course the S table also).
    Naming Convention: /BIC/*<InfoObject Name> or /BI0/*<InfoObject Name>
    C = Customer Defined Characteristic
    0 = Standard or SAP defined Characteristic
    * = P, Q, T, X,Y, S (depending on the above said conditions)
    Thanks & regards
    Sasidhar

  • Error 0 update while loading master data

    Hi,
        Iam facing a problem while loading master data.Data has comes to PSA but it getting failed to load into infoobject it is showing the error " error in the 0 update".
    Ihave tried for manual update, even though it is showing same problem .masterdata 0material is copy back from production to Quality.presently iam uploading data in quality.can anybody give your valuable suggestions.
    Thanks & Regards,
    Mahi.

    I've a compounding characteristic in this characteristic.
    And the master data loading was fine yesterday but today it doesn't work.
    When I check with rsrv this characteristic, a dump with message:
    Error in ABAP application program.
    The current ABAP program "SAPLRRSI" had to be terminated because one of the
    statements could not be executed.
    This is probably due to an error in the ABAP program.
    In program "GP47HH4A9VROVK98YR0TT9S7TQ5 ", the following syntax error occurred
    in the Include "GP47HH4A9VROVK98YR0TT9S7TQ5 " in line 25:
    "Field "/BIC/MZCATWRT" is unknown. It is neither in one of the specifie"
    "d tables nor defined by a "DATA" statement. "DATA" statement."
    So I think something is wrong in this Char.
    But what can I do for that?
    Thanks a lot for all idea.
    Juan

  • Use of "with master data & with out master data" at DTP update level

    Hello experts,
    In DTP, I check "with out master data". When I try to send corresponding transactional data, It is showing SID related error. Can anybody suggest what is the use of "with master data & with out master data" at DTP level.
    Thanks in advance,
    Zakir.

    HI
    HI in DTP level If you set this indicator, the system terminates the update of the request if no values are available for a data record.
    Load the relevant master data before you load the transaction data.
    If you set this indicator, the system terminates activation if master data is missing and produces an error message.
    If you do not set this indicator, the system generates any missing SID values during activation.
    In DataStore maintenance, if you do not set the SIDs Generation upon Activation indicator, the No Update without Master Data indicator in the DTP has no effect.
    thx
    vijju

  • Is it possible to make a delta load for a Master data with Standard DS

    I have a full load bringing huge data for master data with standard datasource.
    I want to run a delta due to huge no. of records but when I create a new Infopackage it dont give a option for delta update.
    Are delta loads specific to only standard or customized DS's or any other reason behind that ?

    I kind of understand what you are asking about, but I am unclear as to how it pertains to our BO SDK.
    You are wanting to find the differences between a large dataset and another large dataset.
    I am not sure what an Infopackage is.
    Are you using the BO Enterprise SDK or some other product?
    Jason

  • C# load xml data with attributes to datagridview; update data and save back to the xml

    Hi guys,
    i currently have this XML as data
    <DEALS>
    <DEAL ID="22" ISBN="0-7888-1623-3" Screenplay="Mary Kaplan" Title ="Mr Gentleman" Director = "Jonathan Jones">
    <ALLOCATION DEMANDE="5000" CODE="72" PRICE="25.00">2500</ALLOCATION>
    <ALLOCATION DEMANDE="7000" CODE="75" PRICE="35.00">4000</ALLOCATION>
    </DEAL>
    <DEAL ID="23" ISBN="1-7988-1623-3" Screenplay="Joe Doe" Title ="Nothing Much" Director = "Listentome">
    <ALLOCATION DEMANDE="3300" CODE="72" PRICE="15.00">2500</ALLOCATION>
    </DEAL>
    </DEALS>
    I load the data with the code below:
    i use xDocument to load the DealData.xml and then put it in ToString.
    //outside of the form
    public static class XElementExtensions
    public static DataTable ToDataTable(this XElement element)
    DataSet ds = new DataSet();
    string rawXml = element.ToString();
    ds.ReadXml(new StringReader(rawXml));
    return ds.Tables[0];
    public static DataTable ToDataTable(this IEnumerable<XElement> elements)
    return ToDataTable(new XElement("Root", elements));
    //in the form
    private void button2_Click(object sender, EventArgs e)
    var xDocument = XDocument.Load("DealData.xml");
    string txtxml = xDocument.ToString();
    StringReader reader = new StringReader(txtxml);
    XDocument doc1 = XDocument.Load(reader);
    var res = doc1.Descendants("DEAL").ToList().Where(x => x.Attribute("ID").Value == "22").Descendants("ALLOCATION");
    dataGridView1.DataSource = res.ToDataTable();
    Now my question is:
    I would like to update the data from the DataGridview with the Attribute("ID").Value == "22" and save the data back to the XML.
    For example, after I load my data on the datagridview, we only pickup anything that is in the node
    <DEAL ID="22" ISBN="0-7888-1623-3" Screenplay="Mary Kaplan" Title ="Mr Gentleman" Director = "Jonathan Jones">
    <ALLOCATION DEMANDE="5000" CODE="72" PRICE="25.00">2500</ALLOCATION>
    <ALLOCATION DEMANDE="7000" CODE="75" PRICE="35.00">4000</ALLOCATION>
    </DEAL>
    I updated the datagridview below with DEMANDE = 9000 CODE = 66 PRICE = 24.77 AND ALLOCATION = 1200
    I want be able to extract the data back in the xml as a child of the DEAL ID = "22"
    So that the XML file looks like this 
    <DEALS>
    <DEAL ID="22" ISBN="0-7888-1623-3" Screenplay="Mary Kaplan" Title ="Mr Gentleman" Director = "Jonathan Jones">
    <ALLOCATION DEMANDE="5000" CODE="72" PRICE="25.00">2500</ALLOCATION>
    <ALLOCATION DEMANDE="7000" CODE="75" PRICE="35.00">4000</ALLOCATION>
    <ALLOCATION DEMANDE="9000" CODE="66" PRICE="24.77">1200</ALLOCATION> <-! this is the new line !->
    </DEAL>
    <DEAL ID="23" ISBN="1-7988-1623-3" Screenplay="Joe Doe" Title ="Nothing Much" Director = "Listentome">
    <ALLOCATION DEMANDE="3300" CODE="72" PRICE="15.00">2500</ALLOCATION>
    </DEAL>
    </DEALS>
    Is there a way to achieve that?
    I have been searching and reading in the books but i cannot find a solution for this.
    Thank you
    Please do not forget to click “Vote as Helpful” if the reply helps/directs you toward your solution and or "Mark as Answer" if it solves your question. This will help to contribute to the forum.

    I would think of something like the below, the id is passed as static you need to change this
    protected virtual void button1_Click(object sender, EventArgs e)
    //Get DataTable from DGV datasource
    DataTable dt = new DataTable();
    dt = (DataTable)dataGridView1.DataSource;
    string file1 = "<?xml version=\"1.0\" encoding=\"utf-8\" ?><DEALS><DEAL ID=\"22\" ISBN=\"0-7888-1623-3\" Screenplay=\"Mary Kaplan\" Title =\"Mr Gentleman\" Director = \"Jonathan Jones\"><ALLOCATION DEMANDE=\"5000\" CODE=\"72\" PRICE=\"25.00\">2500</ALLOCATION><ALLOCATION DEMANDE=\"7000\" CODE=\"75\" PRICE=\"35.00\">4000</ALLOCATION></DEAL><DEAL ID=\"23\" ISBN=\"1-7988-1623-3\" Screenplay=\"Joe Doe\" Title =\"Nothing Much\" Director = \"Listentome\"><ALLOCATION DEMANDE=\"3300\" CODE=\"72\" PRICE=\"15.00\">2500</ALLOCATION></DEAL></DEALS>";
    StringReader reader = new StringReader(file1);
    XDocument doc1 = XDocument.Load(reader);
    //Remove all elements related to the id being populated into the Grid
    doc1.Descendants("DEAL").ToList().Where(x => x.Attribute("ID").Value == "22").Descendants("ALLOCATION").Remove();
    //loop through datatable and create new xelement to be added to the xdocument
    foreach (DataRow dr in dt.Rows)
    XElement xe = new XElement("ALLOCATION",
    new XAttribute("DEMANDE", dr[0].ToString()),
    new XAttribute("CODE", dr[1].ToString()),
    new XAttribute("PRICE", dr[2].ToString()));
    xe.Value = dr[3].ToString();
    doc1.Descendants("DEAL").ToList().Where(x => x.Attribute("ID").Value == "22").FirstOrDefault().Add(xe);
    Fouad Roumieh

Maybe you are looking for