MASTER DATA LOADING  IN SAP BI7.0

how to load master data attribute, text , hierarchies in sapBI7.0

check these
http://help.sap.com/saphelp_nw2004s/helpdata/en/e8/2cd937651f1a13e10000009b38f8cf/content.htm
http://help.sap.com/saphelp_nw70/helpdata/en/c8/e92637c2cbf357e10000009b38f936/content.htm

Similar Messages

  • Master Data Load to APO from SAP MDM through SAP PI

    Hi SDners,
    This is a Parts Master Data Solution for one of the Largest Auto Manufacturer where SAP MDM will be the central hub of all the Global and Local attributes of Parts . The subscribing system is SAP SCM APO . Please advice if there is any possibility of direct integration between APO -MDM through PI. Does PI has the standard IDOC types to communicate with APO.
    Also does APO has some RFC/BAPI to do master data load for Product Master /n/sapapo/mat1
    Thanks,
    Prabuddha

    Hi,
    Check the LUWs in SM58 in source system and then execute it. Else check in BD87 and push the IDocs manually in ECC.
    Thanks
    Reddy
    Edited by: Surendra Reddy on Feb 24, 2009 10:59 AM

  • Sap bw 7.3 master data load error

    sap bw 7.3 master data load error
    erroe: Exception in Substep Rules

    Hi Jayram,
    I am assuming that you are getting this error when you are loading data from PSA to IO. If so, the error might be because of
    1. Duplicate records
    PSA might have duplicate records. But you will be able to load only one record to Master data IO. In DTP Update tab, there is an option for "Handle Duplicate Record Keys" . Just enable this & try to load again.
    2. Erroneous records.
    Some special reocrds are allowed in PSA where as thye might not be allowed in Master Data/Data Targets like Lower case letters/ some speacial characters etc. If error is because of this, then you might need to correct the data at PSA level & reload the data to Master data. Else get it corrected in source system itself & fetch the data to BW. Or write some code to take care of these special characters.
    Hope it helps!
    Regards,
    Pavan

  • Problem with Master Data Load

    Dear Experts,
    If somebody can help me by the following case, please give me some solution. Iu2019m working in a project BI 7.0 were needed to delete master data for an InfoObject material. The way that I took for this was through tcode u201CS14u201D. After that, I have tried to load again the master data, but the process was broken and the load done to half data.
    This it is the error:
    Second attempt to write record 'YY99993' to /BIC/PYYYY00006 failed
    Message no. RSDMD218
    Diagnosis
    During the master data update, the master data tables are read to determine which records of the data package that was passed have to be inserted, updated, or modified. Some records are inserted in the master data table by a concurrently running request between reading the tables at the start of package processing and the actual record insertion at the end of package processing.
    The master data update tries to overwrite the records inserted by the concurrently running process, but the database record modification returns an unexpected error.
    Procedure
    u2022     Check if the values of the master data record with the key specified in this message are updated correctly.
    u2022     Run the RSRV master data test "Time Overlaps of Load Requests" and enter the current request to analyze which requests are running concurrently and may have affected the master data update process.
    u2022     Re-schedule the master data load process to avoid such situations in future.
    u2022     Read SAP note 668466 to get more information about master data update scheduling.
    Other hand, the SID table in the master data product is empty.
    Thanks for you well!
    Luis

    Dear Daya,
    Thank for your help, but I was applied your suggesting. I sent to OSS with the following details:
    We are on BI 7.0 (system ID DXX)
    While loading Master Data for infoobject XXXX00001 (main characteristic in our system u2013 like material) we are facing the following error:
    Yellow warning u201CSecond attempt to write record u20182.347.263u2019 to BIC/ XXXX00001 was successfulu201D
    We are loading the Master data from data source ZD_BW_XXXXXXX (from APO system) through the DTP ZD_BW_XXXXX / XXX130 -> XXXX00001
    The Master Data tables (S, P, X) are not updated properly.
    The following reparing actions have been taken so far:
    1.     Delete all related transactional and master data, by checking all relation (tcode SLG1 à RSDMD, MD_DEL)
    2.     Follow instructions from OSS 632931 (tcode RSRV)
    3.     Run report RSDMD_CHECKPRG_ALL from tcode SE38 (using both check and repair options).
    After deleting all data, the previous tests were ok, but once we load new master data, the same problem appears again, and the report RSDMD_CHECKPRG_ALL gives the following error.
    u201CCharacteristic XXXX00001: error fund during this test.u201D
    The RSRV check for u201CCompare sizes of P and X and/or Q and Y tables for characteristic XXXX00001u201D is shown below:
    Characteristic XXXX00001: Table /BIC/ PXXXX00001, /BIC/ XXXXX00001 are not consistent 351.196 derivation.
    It seems that our problem is described in OSS 1143433 (SP13), even if we already are in SP16.
    Could somebody please help us, and let us know how to solve the problem?
    Thank for all,
    Luis

  • Error in Setting lock for Master Data Load

    Hi Team,
    I encounter the following error while uploading the Masterdata
    "Lock NOT set for: Loading master data attributes
    Attributes of characteristic 0BPARTNER are locked by terminated change run 465451241"
    I Performed the following steps
    1) Checked if any active jobs (change run jobs) running in the system. But no jobs running.
    2) RSATTR->infoobject list-> no infoobject found
    3)Tried after sometime ,but load continue to fail
    Could anyone help to solve the problem
    thanks
    Bala

    Hi Bala,
    Check in SM12 if there are any locks exist with the change run number?
    First check if there are any Master data loads running related to it?
    Check in the Attribute change run screen if there are any change run steps are running?
    Once after checking the above if no locks exist and no master data is running then check the below steps
    1. Run RSA1 transaction
    2. Tools -> Apply Hierarchy/Attibutes Change(or direct transaction RSATTR)... menu option
    3. Click at "Monitor and Start terminated Change Runs" button (at the bottom of the screen)
    4. Click "Reset Status" button at Run ID 465451241
    NOTE: this is not recommended, to unlock the locks better to wait till the locks get relased.
    If still the error message persists then we can use a FM in se37 in serious cases which was not recommended by SAP...
    Use RSDDS_AGGR_MOD_CLOSE in se37 and execute--->there give ur failed change run no(465451241) beside I_CNSID and execute...
    It releases all the locks by that change run....and continue with your further work/jobs...
    Regards
    KP

  • R/3 Master Data Load Error

    Hi Experts
       Our master data load from R/3 to BW was failed today.
    This is the error message on status Tab of failed request
    Diagnosis
    No IDocs could be sent to the SAP BW using RFC.
    System response
    There are IDocs in the source system ALE outbox that did not arrive in the ALE inbox of the SAP BW.
    Further analysis:
    Check the TRFC log.
    You can get to this log using the wizard or the menu path "Environment -> Transact. RFC -> In source system".
    Removing errors:
    If the TRFC is incorrect, check whether the source system is completely connected to the SAP BW. Check especially the authorizations of the background user in the source system.
    I navigated through Environment -> Transact. RFC -> In source system"
    In R/3 i got the following error
    timeout during allocate / CPIC-CALL: 'ThS
    #Timeout during connect                 
    How shall i proceed now.
    Thanks

    Hi,
    Check the RFC connection between R3 and BW. Use transaction SM59 for the same or take help from the BASIS team.
    If that is correct check the EDI partner profiles, if they are active and functioning properly. You can take BASIS help here.
    Cheers,
    Kedar

  • About Master data loading

    How the Master data will be loaded from R/3  to BW?
    Just u give detailed answer and if possible navigation. and if any important points are related with it, just u tell me.

    Hi
    Go through these links..........
    http://help.sap.com/saphelp_nw04/helpdata/en/58/a4513ce4af0a40e10000000a114084/frameset.htm
    http://help.sap.com/saphelp_nw04/helpdata/en/c8/e92637c2cbf357e10000009b38f936/frameset.htm
    http://help.sap.com/saphelp_nw04/helpdata/en/c4/b007720ae4c248b945bb16f24bba31/frameset.htm
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/help/sdn_nw04/sapdocu/netweaver/webas/630/helpdata/en/fa/9a2096e32f11d1a5370060087a79ea/content.htm
    http://help.sap.com/saphelp_nw04/helpdata/en/80/1a634be07211d2acb80000e829fbfe/content.htm
    Re: Master data identification
    Master Data Loading Steps?
    Master Data Extraction
    hope it is helpful..................
    assign points...
    GTR

  • Blank Row in table during Master Data Load

    I am having some success with my master data loads, but when I maintain the master data I have noticed that every table has had a blank row inserted.
    Does anybody know why and what I should do with the row (i.e. delete it)?

    This blank row is by default created and you can no way delete it.  Though you delete it, a new row with blank values will be appended.  This is required for some technical reason while reading the table within ABAP programs. 
    This is applicable only for SAP tables and may not be required for custom developed ones unless you want to use this in screen programs.
    Regards,
    Raj

  • Data load into SAP ECC from Non SAP system

    Hi Experts,
    I am very new to BODS and I have want to load historical data from non SAP source system  into SAP R/3 tables like VBAK,VBAP using BODS, Can you please provide steps/documents or guidelines on how to achieve this.
    Regards,
    Monil

    Hi
    In order to load into SAP you have the following options
    1. Use IDocs. There are several standard IDocs in ECC for specific objects (MATMAS for materials, DEBMAS for customers, etc., ) You can generate and send IDocs as messages to the SAP Target using BODS.
    2. Use LSMW programs to load into SAP Target. These programs will require input files generated in specific layouts generated using BODS.
    3. Direct Input - The direct input method is to write ABAP programs targetting on specific tables. This approach is very complex and hence a lot of thought process needs to be applied.
    The OSS Notes supplied in previous messages are all excellent guidance to steer you in the right direction on the choice of load, etc.,
    However, the data load into SAP needs to be object specific. So targetting merely the sales tables will not help as the sales document data held in VBAK and VBAP tables you mentioned are related to Articles. These tables will hold sales document data for already created articles. So if you want to specifically target these tables, then you may need to prepare an LSMW program for the purpose.
    To answer your question on whether it is possible to load objects like Materials, customers, vendors etc using BODS, it is yes you can.
    Below is a standard list of IDocs that you can use for this purpose to load into SAP ECC system from a non SAP system.
    Customer Master - DEBMAS
    Article Master - ARTMAS
    Material Master - MATMAS
    Vendor Master - CREMAS
    Purchase Info Records (PIR) - INFREC
    The list is endless.........
    In order to achieve this, you will need to get the functional design consultants to provide ETL mapping for the legacy data to IDoc target schema and fields (better to ahve sa tech table names and fields too). You should then prepare the data after putting it through the standard check table validations for each object along with any business specific conversion rules and validations applied. Having prepared this data, you can either generate flat file output for load into SAP using LSMW programs or generate IDoc messages to the target SAPsystem.
    If you are going to post IDocs directly into SAP target using BODS, you will need to create a partner profile for BODS to send IDocs and define the IDocs you need as inbound IDocs. There are few more setings like RFC connectivity, authorizations etc, in order for BODS to successfully send IDocs into the SAP Target.
    Do let me know if you need more info on any specific queries or issues you may encounter.
    kind regards
    Raghu

  • Automate Master data loads in BPC7(MS)

    I need to update dimension members in SAP BPC7 (MS) from an external source.
    The source system for my current project is BW, but if you have done any automated master data loads from any other system or even from a "temp" table please post your solution or tips.
    Thanks!
    Mez

    Hi Mezert,
    Would like to chime in with some of the things I've learned about this process over 3-4 years I've spent with BPC/Outlooksoft:
    -- Automating maintenance of the dimension/master data might be the biggest single weakness of the BPC solution on the MS platform.  It's one of the big reasons that the Netweaver platform basically replaces DTS/SSIS and has more direct linkage to BI/BW cubes and master data (InfoCubes).
    -- It is also one a very common issue and in fact, you will have to deal with what the old-timers in BI call "slowly-changing dimensions" at almost every client eventually.  For instance, you have a list of proucts/materials that every so often someone adds an item to.  ...Well, before you can enter anything for it in BPC, you have to add the dimension member, of course.
    -- So far, this challenge has been met by bringing in specialist DTS/SSIS experts in most cases who can employ the special "Custom Tasks" and build one of these packages to run in an automated way to populate the tables inside SQL Server and then execute the Admin MakeDim task to process the cube/application.
    -- Usually the development/testing/deployment of these solutions is a non-trivial undertaking.  It can be done, but there are peculiar challenges.
    -- One example of a challenge is what happens if you need a dimension with > 65,000 members?  Well, it can be created and populated by a DTS package, however, when you check the package and the custom task from the old Outlooksoft stuff out, you see that it CLEARS the entire dimension table before re-filling it.  ...Yep, it wouldn't allow an "Append" function to run. 
    -- A second example of a peculiar challenge is that unless you update the Excel Dimension Sheet, you will have to do all the updates in SQL Server Mgmt. Studio and really won't be able to touch the BPC Admin again for that dimension.
    Apologies for such a long post, but this is a fairly significant issue and I have to do a lot of research to piece together some of these things.  ...If anyone else has something to add as a best-practice, please post.

  • How many tables are created in Master data loading ( Including hierchy )

    hi ,
    Please anybody , can you tell us how many tables are created in Master data loading ( Including hierchy )
    thanks
    Phani

    Hi PP,
    Depending upon the definition of an infoobject, the tables are created when the objects are activated.
    /BI<C OR DIGIT>/<TABLE CODE><INFOOBJECT>
    <C or digit>: C = Customer-defined InfoObjects
    digit = SAP-defined InfoObjects
    <table code>: S = SID table
    T = Text table
    P = Time-independent master data attributes
    Q = Time-dependent master data attributes
    M = Union of time-dependent and time-independent master
    data attributes
    X = SID table for time-independent navigational attributes
    Y = SID table for time-dependent navigational attributes
    H = Hierarchy table
    K = Hierarchy SID table
    I = Hierarchy SID structure table
    J = Hierarchy interval table
    Thnaks...
    Shambhu

  • Master data Load - Strategy

    Hi All,
    I would like to know the better master data strategy.
    When using 3.x master data load strategy I can load directly into the info object without going thru additional DTP process. Is there any specific advantages of doing this, besides that it is 7.X.
    I have been using 2004s from 2005 but most of the implementations we used 3.x methodology for master data load and  DTP for transaction load. i would like to know whether SAP recommands new methodology using DTP for master data loads? If I load my master data using 3.x I can avoid one extra step, but will it be discontinued in the future? and have to use DTP even for this?
    Please advice if you know what is the best way forward strategically an dtechnically?
    Thanks,
    Alex.

    Alex,
    Please read my answer...
    The new data flow designed by SAP is using the DTP, even for Master Data... Right now you can use the "3.x" style which they maintain for backward-compatibility but down the road, eventually, it will be dropped, so looking ahead, technically and strategically, the right way to go is by using DTP...
    You can go here http://help.sap.com/saphelp_nw70/helpdata/en/e3/e60138fede083de10000009b38f8cf/frameset.htm and check under Data Warehousing, Data Distribution, Data Transfer Process...
    You could also open an OSS note to SAP and ask them directly.
    Thanks,
    Luis

  • BI Statistics Master Data Load

    Hi,
    We are implementing BI Statistics and BI Administration Cockpit.  One of the OSS Note 934848 mentions that I have to install/activate all the datasources that are under the Application Component BW_TECH --> TCT.
    My question is do I have to consider loading the master data for all those datasources/infoobjects ?  There are nearly 80 - 100 datasources.
    The note 934848 mentions to install the Business Content Process Chains.  And I see that those process chains are hardly loading 10 - 12 master data loads.  Is it OK if I just make use of those business content process chains to load master data ?
    Initially I have created DTP's and Transformations to load master data for approx 50 objects.  Later I came to know that BI statistics Cubes etc follow the 3.x format and I should use those as it is delivered.  So, can't I still make use of the 7.0 converted datasourse objects with DTP to load the data ?
    -or- do I have to delete all the transformations and DTP's and migrate all those DataSources back to 3.x ?
    Thanks,

    Hi Sesh,
    1. Yes you need to activate all those Datasources. Its better to activate all of them as in BI 7.0 TCT infoproviders are used as basis for Workload monitor ST03 also in addition to Admin cockpit.
    2. Its sufficient to just load the master data using the standard process chains (Init chains once followed by delta chains.
    3. Its recommended to use the TCT content as 3.x only.
    4. No need to migrate the 3.x content, just use as it is.
    for more information you can refer to this <a href="https://forums.sdn.sap.com/thread.jspa?threadID=279998&tstart=0">thread</a> and related posts by <a href="https://forums.sdn.sap.com/profile.jspa?userID=2035002&start=0">Rajani Saralaya K</a>
    Assign points if helpful,
    Regards, Uday

  • Master Data Loading for Prices and Conditions in CRM - "/SAPCND/GCM"

    Hi,
    Could anyone give me some inputs on Master Data Loading for Prices and Conditions in CRM.
    T. Code is:  /SAPCND/GCM
    I need to load data on a file (extracted from 4.6) for service contracts.
    I tried LSMW : for this transaction, recording does not work.
    I am trying loading thru Idocs (LSMW). But that too is note really working.
    Do we require some custom development for this , or is some SAP standard funcntionality available ??
    Can anyone provide some valuable inputs one this.
    Would appreciate your responses.

    Hi Tiest,
    Thanx for responding.
    U r right, our clint is upgrading from 4.6 to ECC.
    So as per the clients requirements, we are maintaining all the configs for Services in CRM.
    Services Data which was in 4.6 is being pulled put on flat files which needs to be loaded in CRM. So middleware would not be able to do this.
    What I am looking os some standard upload program.
    LSMW recording does not work.
    This I-Doc "CRMXIF_COND_REC_SLIM_SAVE_M", i am able to load a single record. But I am not able to find , how to make this function for multiple entries.
    IN standard we for loading master data thru I-docs, we map the values to the standard fields which are available in that I-Doc.
    But in this particular i-doc, there is a common field for which I need to define the field name and a field value..
    Till now, I am only able to define just one field name and a field value.
    I want this to word for mutliple entries.
    Hope u get my point.
    Thanx

  • Master data Load for Substance Creation :  Using BAPI in LSMW  ?

    Hi ,
    Had any one did a Master data load for creating a Substance in SAP RM. If so please let me know the best possible option. I am beeing trying to figure out the option to load Substances and its IDENTIFIER / MATERIAL ASSINGMNET from my legacy system. But I am not geeting a clue to use LSMW. ( Direct or BatchInput )
    I am not seeing any BAPI for uploading it through the LSMW except the one BUS1077 ( method SAVEREPMUL ) and I guess this is used along with IDOC.
    Appreciate your help and suggestion.
    David

    Thanks John.
    I will try this option what you have suggested. Even I told by business team to look for CG33. But they were been telling me some format issue. Probably they would meant the same what you have said. But they were not aware of work around for this. your information is intresting.
    The other option given to me was to create a recording for each and every Tab like Substance header / Identifier / Material assignment etc. seperately. Technically I don't feel good to create in that way. Do you think that is alernate approach.?
    With Respect & Regards
    David

Maybe you are looking for

  • Deleting from the loop browser

    Is there a way to delete loop names from the loop browser drop down list and the actual loops themselves? I inadvertently (sp?) added loop folders that I had bought and downloaded to the loop browser without first naming the loop folders being added.

  • How is a file attachment stored in database?

    Hi all~~ I am developping an email application(web based) using Javamail. I need to store the attachment together with the relevant email messages in the database. Is that correct that I add an attribute called ATTACHMENT in the entity which stores t

  • Can LR Web module converting file name hyphens into underscores be disabled

    Hi, I have problems when creating webpages from within LR. First I develop the RAW files and converts them to jpegs. Under Lightroom 3 I use the following settings under "Export location": Export to: Specific Folder Add to This Catalog: ticked Existi

  • Transferring files from windows media player

    how do i transfer all my library from windows media player and napster into itunes? when i first hooked up my ipod to my computer it prompted me to convert my whole library but now i can't figure out how to do it again. thanks

  • Backup project to DVDR (4.7GB)

    Hi, I've Done my job.and now i wish to backup all my projects (an album about 10songs).Some project/songs are below 4GB (i'm able to burn it into DVDR) but some more than 4GB.What s'ld i do to make the file size small and be able to burn into 4.7GB D