Master Data loading is very slow.

Hi Experts,
I have scheduled Master Data Attribute Process Chain daily. 0employee_attr, infoObject is having only 487315 & it is taking more than 12 hrs. There are number of infoobjects are much bigger is taking 10 minutes only. 0Employee attr in increasing 5-10 records daily only. Earlier it was taking 4-5 hours.
Regards,
Anand Mehrotra.

Hi,
You must have the following profiles to BWREMOTE or ALEREMOTE users.So add it. Bcoz either of these two users will use in background to get extract teh data from ECC, so add tehse profiels in BW.
S_BI-WHM_RFC, S_BI-WHM_SPC, S_BI-WX_RFC
And also check the following things.
1.Connections from BW to ECC and ECC to BW in SM59
2.Check Port,Partner Profiles,and Message Types in WE20 in ECC & BW.
3.Check Dumps in ST22, and SM21.
4.If Idocs are stuck i.e see the OLTP Idoc numbers in RSMO Screen in (BW) detials tab see in bottom, you can see OLTP Idoc number and take the Idoc numbers and then goto to ECC see the status in WE05 or WE02, if error then check the log else goto to BD87 in ECC and give the Idoc numbers and execute manually and see in RSMO and refresh.
5.Check the LUWs struck in SM58,User Name = * (star) and run it and see Strucked LUWs and select our LUW and execute manually and see in RSMO in BW.
See in SDN
Re: Loading error in the production  system
Thanks
Reddy

Similar Messages

  • Data load is very slow

    Hi Experts,
    I am working on CRM Analytic. I am loading address data from extractor 0BP_DEF_ADDRESS_ATTR to Business partner, with  19 lakhs records. When I execute the DTP it is taking 3 to 4 days to complete the load.
    Please provide me solution so that my data load will become fast.
    With Regards,
    Avenai

    Hi,
    Increase the Number of parallel processes.
    in order to increase parallel processes --> from menu of DTP -> goto -> "settings for batch manager" -> increase the Number of parallel processes (by default it will be 3 increase it to 6).
    Increase the Datapacket size in the DTp -extraction tab.
    Do you have any routines used in the transfromations? if yes try to debug the code where its taking time, find n try to fine tune the code with the help of ABAP person.
    Below option may also be one of the reasons while using CRM data sources...
    The data source consists of lots of fields which are not used or mapped in the transformation ... try to hide those fields or you can create a copy of your data source using BWA1 transaction in CRM system.
    Regards
    KP

  • Data load becomes very slow

    Hi,after a migration from Version 5 to 6.5 the dataload becomes very slow.With V5 the dataload takes 1 hour, with 6.5 it takes about 3 hours.the calculation takes the same time.Any idea?

    To many sub VIs could not be found so I can not give you more than some advises. But I see that you run all your loops at full speed. I do not think it is very wise. Insert the "wait (ms)" functions in all while loops, but not the loops handling the daq functions. Since they are controlled by a occurrence. In Loops handling user input only you may set the wait time as high as 500ms. In more important loops use shorter time.
    Besides which, my opinion is that Express VIs Carthage must be destroyed deleted
    (Sorry no Labview "brag list" so far)

  • Automate Master data loads in BPC7(MS)

    I need to update dimension members in SAP BPC7 (MS) from an external source.
    The source system for my current project is BW, but if you have done any automated master data loads from any other system or even from a "temp" table please post your solution or tips.
    Thanks!
    Mez

    Hi Mezert,
    Would like to chime in with some of the things I've learned about this process over 3-4 years I've spent with BPC/Outlooksoft:
    -- Automating maintenance of the dimension/master data might be the biggest single weakness of the BPC solution on the MS platform.  It's one of the big reasons that the Netweaver platform basically replaces DTS/SSIS and has more direct linkage to BI/BW cubes and master data (InfoCubes).
    -- It is also one a very common issue and in fact, you will have to deal with what the old-timers in BI call "slowly-changing dimensions" at almost every client eventually.  For instance, you have a list of proucts/materials that every so often someone adds an item to.  ...Well, before you can enter anything for it in BPC, you have to add the dimension member, of course.
    -- So far, this challenge has been met by bringing in specialist DTS/SSIS experts in most cases who can employ the special "Custom Tasks" and build one of these packages to run in an automated way to populate the tables inside SQL Server and then execute the Admin MakeDim task to process the cube/application.
    -- Usually the development/testing/deployment of these solutions is a non-trivial undertaking.  It can be done, but there are peculiar challenges.
    -- One example of a challenge is what happens if you need a dimension with > 65,000 members?  Well, it can be created and populated by a DTS package, however, when you check the package and the custom task from the old Outlooksoft stuff out, you see that it CLEARS the entire dimension table before re-filling it.  ...Yep, it wouldn't allow an "Append" function to run. 
    -- A second example of a peculiar challenge is that unless you update the Excel Dimension Sheet, you will have to do all the updates in SQL Server Mgmt. Studio and really won't be able to touch the BPC Admin again for that dimension.
    Apologies for such a long post, but this is a fairly significant issue and I have to do a lot of research to piece together some of these things.  ...If anyone else has something to add as a best-practice, please post.

  • Master Data load via DTP (Updating attribute section taking long time)

    Hi all,
    Iam loading to a Z infoobject. Its a master data load for attributes.Surprisingly, i could find that PSA  pulls records very fastly( 2 minutes) but the DTP which updates the infoobject takes a lot of time. It runs into hours.
    When observed the DTP execution monitor, which shows the breakup of time between extraction,filter,transformation,updation of attributes
    i could observe that the last step "updation of attributes for infoobject" is taking lots of time.
    The masterdata infoobject has got also two infoobjects compounded.
    In transformation ,even they are mapped.
    No of parallel processes for the DTP was set to 3 in our system.
    Job Class being "C".
    Can anyone think of what could be the reason.

    Hi,
    Check the T code ST22 for any short dump while loading this master data. There must be some short dump occured.
    There is also a chance that you are trying to load some invalid data (like ! character as a first character in the field) into the master.
    Regards,
    Yogesh.

  • Error is Master Data load.

    Hi Experts,
    From last 1 month the ( 0MAT_UNIT ) master data load is getting failed. Its a full load. So the the target is not updated from last one month( August)
    Now the problem is , when I am checking a report. The materials which are new or changed are not getting the data properly. It is because of the master data got stucked at PSA level.And I cant rectify at PSA level because, There is 673 errorneous record in 1 package.
    I tried to find the invalid character , but didn't get anything.
    Please suggest.
    Thanks,
    TapuKumar

    Hi Anand,
            This is the long text msg
    Data records in the PSA were marked for data package 1 . Here there were
    673 errors. The system wrote information or warnings for the remaining 
    data records.
                                                                                    System response                                                                               
    The data package was not updated.                                                                               
    Procedure                                                                               
    Correct the incorrect data records of the data package (for example by   
         manually editing them in the PSA maintenance). You can find the error    
         message for each of the records in the PSA by double-clicking on the     
         status of the record.                                                                               
    Note                                                                               
    Note that you are only allowed to edit and manually update the request   
         if it is contained in the request list in the data target. Where         
         necessary, first delete it from the data target.                         
    And it is a master data load( Full ), The data target is a INFO OBJECT.
    Please suggest, I can't rectify each and every reecord. Because the number of record id very high..
    Thanks,
    Tapu Kumar.
    Edited by: NewToBW on Sep 2, 2010 1:42 PM

  • Master Data loading got failed: error "Update mode R is not supported by th

    Hello Experts,
    I use to load master data for 0Customer_Attr though daily process chain, it was running successfully.
    For last 2 days master data loading for 0Customer_Attr got failed and it gives following error message:
    "Update mode R is not supported by the extraction API"
    Can anyone tell me what is that error for? how to resolve this issue?
    Regards,
    Nirav

    Hi
    Update mode R error will come in the below case
    You are running a delta (for master data) which afils due to some error. to resolve that error, you make the load red and try to repeat the load.
    This time the load will fail with update mode R.
    As repeat delta is not supported.
    So, now, the only thing you can do is to reinit the delta(as told in above posts) and then you can proceed. The earlier problem has nothing to do with update mode R.
    example your fiorst delta failed with replication issue.
    only replicating and repeaing will not solve the update mode R.
    you will have to do both replication of the data source and re-int for the update mode R.
    One more thing I would like to add is.
    If the the delat which failed with error the first time(not update mode R), then
    you have to do init with data transfer
    if it failed without picking any records,
    then do init without data transfer.
    Hope this helps
    Regards
    Shilpa
    Edited by: Shilpa Vinayak on Oct 14, 2008 12:48 PM

  • Problem with Master Data Load

    Dear Experts,
    If somebody can help me by the following case, please give me some solution. Iu2019m working in a project BI 7.0 were needed to delete master data for an InfoObject material. The way that I took for this was through tcode u201CS14u201D. After that, I have tried to load again the master data, but the process was broken and the load done to half data.
    This it is the error:
    Second attempt to write record 'YY99993' to /BIC/PYYYY00006 failed
    Message no. RSDMD218
    Diagnosis
    During the master data update, the master data tables are read to determine which records of the data package that was passed have to be inserted, updated, or modified. Some records are inserted in the master data table by a concurrently running request between reading the tables at the start of package processing and the actual record insertion at the end of package processing.
    The master data update tries to overwrite the records inserted by the concurrently running process, but the database record modification returns an unexpected error.
    Procedure
    u2022     Check if the values of the master data record with the key specified in this message are updated correctly.
    u2022     Run the RSRV master data test "Time Overlaps of Load Requests" and enter the current request to analyze which requests are running concurrently and may have affected the master data update process.
    u2022     Re-schedule the master data load process to avoid such situations in future.
    u2022     Read SAP note 668466 to get more information about master data update scheduling.
    Other hand, the SID table in the master data product is empty.
    Thanks for you well!
    Luis

    Dear Daya,
    Thank for your help, but I was applied your suggesting. I sent to OSS with the following details:
    We are on BI 7.0 (system ID DXX)
    While loading Master Data for infoobject XXXX00001 (main characteristic in our system u2013 like material) we are facing the following error:
    Yellow warning u201CSecond attempt to write record u20182.347.263u2019 to BIC/ XXXX00001 was successfulu201D
    We are loading the Master data from data source ZD_BW_XXXXXXX (from APO system) through the DTP ZD_BW_XXXXX / XXX130 -> XXXX00001
    The Master Data tables (S, P, X) are not updated properly.
    The following reparing actions have been taken so far:
    1.     Delete all related transactional and master data, by checking all relation (tcode SLG1 à RSDMD, MD_DEL)
    2.     Follow instructions from OSS 632931 (tcode RSRV)
    3.     Run report RSDMD_CHECKPRG_ALL from tcode SE38 (using both check and repair options).
    After deleting all data, the previous tests were ok, but once we load new master data, the same problem appears again, and the report RSDMD_CHECKPRG_ALL gives the following error.
    u201CCharacteristic XXXX00001: error fund during this test.u201D
    The RSRV check for u201CCompare sizes of P and X and/or Q and Y tables for characteristic XXXX00001u201D is shown below:
    Characteristic XXXX00001: Table /BIC/ PXXXX00001, /BIC/ XXXXX00001 are not consistent 351.196 derivation.
    It seems that our problem is described in OSS 1143433 (SP13), even if we already are in SP16.
    Could somebody please help us, and let us know how to solve the problem?
    Thank for all,
    Luis

  • Error in Setting lock for Master Data Load

    Hi Team,
    I encounter the following error while uploading the Masterdata
    "Lock NOT set for: Loading master data attributes
    Attributes of characteristic 0BPARTNER are locked by terminated change run 465451241"
    I Performed the following steps
    1) Checked if any active jobs (change run jobs) running in the system. But no jobs running.
    2) RSATTR->infoobject list-> no infoobject found
    3)Tried after sometime ,but load continue to fail
    Could anyone help to solve the problem
    thanks
    Bala

    Hi Bala,
    Check in SM12 if there are any locks exist with the change run number?
    First check if there are any Master data loads running related to it?
    Check in the Attribute change run screen if there are any change run steps are running?
    Once after checking the above if no locks exist and no master data is running then check the below steps
    1. Run RSA1 transaction
    2. Tools -> Apply Hierarchy/Attibutes Change(or direct transaction RSATTR)... menu option
    3. Click at "Monitor and Start terminated Change Runs" button (at the bottom of the screen)
    4. Click "Reset Status" button at Run ID 465451241
    NOTE: this is not recommended, to unlock the locks better to wait till the locks get relased.
    If still the error message persists then we can use a FM in se37 in serious cases which was not recommended by SAP...
    Use RSDDS_AGGR_MOD_CLOSE in se37 and execute--->there give ur failed change run no(465451241) beside I_CNSID and execute...
    It releases all the locks by that change run....and continue with your further work/jobs...
    Regards
    KP

  • How to rectify the errors in master data loads & transactional data loads?

    hy,
    please any one tell me
    How to rectify the errors in master data loads & transactional data loads?
    thnQ
    Ravi

    Hi,
    Please post specific questions in the forum.
    Please explain the error you are getting.
    -Vikram

  • R/3 Master Data Load Error

    Hi Experts
       Our master data load from R/3 to BW was failed today.
    This is the error message on status Tab of failed request
    Diagnosis
    No IDocs could be sent to the SAP BW using RFC.
    System response
    There are IDocs in the source system ALE outbox that did not arrive in the ALE inbox of the SAP BW.
    Further analysis:
    Check the TRFC log.
    You can get to this log using the wizard or the menu path "Environment -> Transact. RFC -> In source system".
    Removing errors:
    If the TRFC is incorrect, check whether the source system is completely connected to the SAP BW. Check especially the authorizations of the background user in the source system.
    I navigated through Environment -> Transact. RFC -> In source system"
    In R/3 i got the following error
    timeout during allocate / CPIC-CALL: 'ThS
    #Timeout during connect                 
    How shall i proceed now.
    Thanks

    Hi,
    Check the RFC connection between R3 and BW. Use transaction SM59 for the same or take help from the BASIS team.
    If that is correct check the EDI partner profiles, if they are active and functioning properly. You can take BASIS help here.
    Cheers,
    Kedar

  • Master data Load Failure

    Hi All,
    Could any body clearly explain the following options in the Info Package of master data Info Object with full update
    Update data
    1. PSA and then into Info Obejcts
    2  . Only PSA   ( Update Subsequently in data targets )
          Data source trasfers no duplicate records
         Igonore double data records
    While I am doing master data load with the 1 and 3 options I am getting same no of records updated in PSA but loadng is getting failured with the 1st option (Caller 01 02 .. 20 Messages ) .and request with 3rd option is success..
    Thanks in Advance...
    Regards,
    Nagamani

    Hi
    U r master data might have the Dublicate records
    select the PSA
    and select the option
    Ignore dublicate data records
    IF  you have the Deblicates
    EX
    CNO
    1010
    1020
    1010
    ITs not alowed the 1010 record
    Regards

  • Master Data loading in BPC 7.5 with prefix

    Hi Gurus,
    I need a help in master loading in BPC 7.5
    Currently we are loading GL account with prefix of A and controlling area(COA1)
    BI:0000600000
    BPC: ACOA10000600000
    I have maintained logic in the transformation file as below. Z001 represents the hierarchy Node.
    ID=IF(ID(1:4)=STR(Z001) then STR(A)+ID;STR(A)0CO_AREAID)
    Now client has came with new requirement as below
                               BI                   BPC
    GL Acc               0000600000     A_600000
    GL Node             Z00113            Z001_13
    GL Node             Z00213            Z002_13 
    For the above requirement i have maintained the logic in the transformation file as below. when running master data load package i got the error as
    Line 755 :Command failed ; end position is out of record index
    ID=IF(ID(1:4)=STR(Z001) then STR(Z001_)+ID(5:12);STR(A_)+ID(5:12))
    I need your help in modifying the logic in the transformation file for handling the above requirement.
    And also for handling two hierarchy Z001 and Z002.
    Thanks
    Mahesh

    Hi,
    Your command should like:
    ID=*IF(ID(1:4)=*STR(Z001) then *STR(Z001_)+ID(5:6);*STR(A_)+ID(5:10))
    Hope this helps.

  • Master Data load does not extract Hierarchy nodes in BPC Dimension ACCOUNT

    Hi Experts,
    I am performing master data load through standard DM package with Filter selection as:
    1. Chart of Accounts
    2. Hieararchy selection has 4 hierarchy names
    3. Selected Import Text nodes
    4. Selected Set Filters by Attribute OR Hierarchies
    I have run this DM package for a set of data and selections a week ago and it worked fine.
    However when i run it now, it is giving issues,
    It extracts any new GL maintained in the BI system however it does not extract any hierarchy nodes at all! (Have tested this by deleting the hierarchy nodes and tried to run the master data load)
    I am running the DM package in Update and have selection as External.
    Any sugestions for checks / has anyone encountered this issue earlier?
    Regards,
    Shweta Salpe

    Hi Guyz,
    Thanks.
    I found that the issue was with the transformation file where i was maintaining the RATETYPE.
    When i removed the mapping of RATETYPE this works fine. (Pulls the nodes of hierarchies)
    however now i do not have Ratetype populated in the system.
    my rate type mapping is:
    RATETYPE=*IF(ID(1:1)=*STR(C) then *STR(TOSKIP);ID(1:1)=*STR(H) then *STR(TOSKIP);ID)
    and in conversion file i have TOSKIP     *skip
    I have to skip the ratetypes for the hierarchy nodes and my hierarchy nodes start with C and H.
    So now that i have removed the mapping for RATETYPE can anyone suggest me a correct way to achieve this? (Note the above mapping formula was skipping all of the hierarchy nodes starting with C and H)
    Regards,
    Shweta Salpe

  • About Master data loading

    How the Master data will be loaded from R/3  to BW?
    Just u give detailed answer and if possible navigation. and if any important points are related with it, just u tell me.

    Hi
    Go through these links..........
    http://help.sap.com/saphelp_nw04/helpdata/en/58/a4513ce4af0a40e10000000a114084/frameset.htm
    http://help.sap.com/saphelp_nw04/helpdata/en/c8/e92637c2cbf357e10000009b38f936/frameset.htm
    http://help.sap.com/saphelp_nw04/helpdata/en/c4/b007720ae4c248b945bb16f24bba31/frameset.htm
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/help/sdn_nw04/sapdocu/netweaver/webas/630/helpdata/en/fa/9a2096e32f11d1a5370060087a79ea/content.htm
    http://help.sap.com/saphelp_nw04/helpdata/en/80/1a634be07211d2acb80000e829fbfe/content.htm
    Re: Master data identification
    Master Data Loading Steps?
    Master Data Extraction
    hope it is helpful..................
    assign points...
    GTR

Maybe you are looking for

  • Issues with Flash Pro CS5 on Mac

    Hi, I recently bought a MacBookPro 15", 2.4 GHz Intel Core i5, 4GB 1067 MHz DDR3 and CS5 Creative Suite. Built in Intel HD Graphics and NVIDIA GeForce GT 330M. Hi have some issues with Flash Pro: - when objects on the stage are quite big (for example

  • Mac Pro 2009 Raid Set Lost after Hard Reboot

    We have a newer Nehalem based Mac Pro with a 2009 Black Apple Hardware Raid card.  The Hard Drive Configuration is: 4 x 2TB Western Digital drives setup as a Raid 1+0 array providing 4TB of space with what we believed to be optimal performance and re

  • How to download Pages 4.3 to replace the flawed 5.2

    Pages 5.2 may well have a lot of nice new features but given that it is now 30% fatter why has Apple removed essential features without publishing any warning about such excisions. 5.2 is very likely to ruin any 4.3 file you open, worse it stops 4.3

  • Potential memory leak in 11.2.0.1.0 cluster stack components

    Hi All, We are running 11.2.0.1.0 RAC on OEL 5.5 and OEL 5.5. On all of our boxes, we have observed a gradual increase and filling of swap space. Swap usage slowly but steadily increase to a point where either the Node boots automatically or we have

  • Media Sync and Windows 2000

    Hello, I'm using for few days a BlackBerry Curve 8900 with a PC under Windows 2000. Did anybody kept the installation package for Media Sync V2, which seems to be compatible with Windows 2000 ? Only the V3 is available on the BlackBerry site. Thanks