Data loading from source system takes long time.

Hi,
     I am loading data from R/3 to BW. I am getting following message in the monitor.
Request still running
Diagnosis
No errors could be found. The current process has probably not finished yet.
System response
The ALE inbox of the SAP BW is identical to the ALE outbox of the source system
and/or
the maximum wait time for this request has not yet run out
and/or
the batch job in the source system has not yet ended.
Current status
in the source system
Is there anything wrong with partner profile maintanance in the source system.
Cheers
Senthil

Hi,
I will suggest you to check a few places where you can see the status
1) SM37 job log (In source system if load is from R/3 or in BW if its a datamart load) (give request name) and it should give you the details about the request. If its active make sure that the job log is getting updated at frequent intervals.
Also see if there is any 'sysfail' for any datapacket in SM37.
2) SM66 get the job details (server name PID etc from SM37) and see in SM66 if the job is running or not. (In source system if load is from R/3 or in BW if its a datamart load). See if its accessing/updating some tables or is not doing anything at all.
3) RSMO see what is available in details tab. It may be in update rules.
4) ST22 check if any short dump has occured.(In source system if load is from R/3 or in BW if its a datamart load)
5) SM58 and BD87 for pending tRFCs and IDOCS.
Once you identify you can rectify the error.
If all the records are in PSA you can pull it from the PSA to target. Else you may have to pull it again from source infoprovider.
If its running and if you are able to see it active in SM66 you can wait for some time to let it finish. You can also try SM50 / SM51 to see what is happening in the system level like reading/inserting tables etc.
If you feel its active and running you can verify by checking if the number of records has increased in the data tables.
SM21 - System log can also be helpful.
Thanks,
JituK

Similar Messages

  • Data load in the Production takes long time

    Hi All,
    We have load which is extracting data from 0HR_PA_0 Data Source. It is loading daily some 2 lacs 30thousand records.But it was taking lots of time to load around 3 hrs.So if we check in the joblog of the source system the job is active for a longer time and i can see there are no enhancements done for the data source.So is there any body who can throw light on this.The processing type is psa and data targets in serial. If we change to PSa and datatargets parallel is there any improvement in the load?
    Thanks

    You have told that you have enhancement in the data source.. in ABAP make sure below optimizations are met
    Below things we need to keep in our mind:
    1)       Try not use SELECT statement inside WHILE loop
    2)       Use WHILE <logic> ASSIGNING <FS> instead of Use WHILE <logic> INTO WA.
    3)       We should not take all table in work area instead we will use proper where clause in select statement to minimize load in memory.
    4)       While reading Internal table use BINARY SEARCH as much as possible.
    Above guideline also applies for Start routines.
    Cheers,
    DK
    (Giving points is another way to say thanks:))

  • Doesn't load from source system to PSA and DSO

    Hi Gurus,
    We have launched a Delta load from Source System (POSDM) to DSO. In the source system there are 5000000 rows but none was loaded into DSO or PSA. The delta load was ok, but any data was loaded.
    If the delta  load was lunch again, would be loaded those 5000000 rows. Or those 5000000 rows wouldn't be able to load again.
    Any idea abut this issue? Any feedback will be really appreciated.
    Thanks in advance.

    Hi David,
    Are you sure this 5 million records are delta records and should be pulled into BW as delta??
    did you count the number of records in the underlying tables??
    which data source you are using??
    Delta loads are suppose to bring the new records or changed records since the last data load and not all the records.
    Since the request is green as you said and still it shows 0 records then it is possible that nothing has changed since last delta.try to see the details of the data load... if it failed...may be thats the reason you are not able to see any records.
    If you schedule the delta again it will bring the records changed or created after last delta.
    If the delta was unsuccessfull then turn the QM status and overall status of request to red manually in the monitor of the request and delete it from PSA and all the targets in which its loaded and schedule the delta again and it should bring the delta.
    Thanks
    Ajeet
    Thanks
    Ajeet

  • Data load from Legacy system to BW Server through BAPI

    Requirements: We have different kind of legacy systems and SAP BW server. We want to load all legacy system data into SAP BW server using BAPI. Before loading we have to validate all data. If there are bad data, data missing we have to let the legacy system user/ operator knows to fix the data into their system with detail explanation. When it is fixed, we have to load the data again.
    Load Scenario:  We have two options to load data from legacy systems to BW Server.
    1.     We need to load data directly from legacy system to BW Server using BAPI program.
    2.     Legacy Systems data would be in workstations or flash drive as .txt (one line separated by comma) or .csv file. Need to load from .txt /.csv file to BW Server using BAPI program.
    What we want in the BAPI program code?
    It will Read / Retrieve data from text / csv file and will put into the Internal table. Internal table structure would be based on BAPI InfoObject structure.
    Call BAPI InfoObject function module ‘BAPI_IOBJ_CREATE’ to create InfoObject, include all necessary / default components, do the error check, load the data and return the status.
    Could some one help me with the sample code please? I am new in ABAP / BAPI coding.
    Is there any other better idea to load data from legacy system to BW server? BTW we are using BW 3.10. Is there any better option with BI 7.0 to resolve the issue? I appreciate your help.

    my answers:
    1. this is a scendario for a data push into SAP BW. You can only use SOAP-Based Transfer of Data.
    http://help.sap.com/saphelp_nw04/helpdata/en/fd/8012403dbedd5fe10000000a155106/frameset.htm
    (here for BW 3.5, but you'll find similar for 7.0)
    In this scenario you'll have an RFC dinamically created for every Infosource you need to transfer data.
    2. You can make a chain for each data load, so you can call the RFC "RSPC_API_CHAIN_START" to start the chain from external.
    the second solution is more simply and available on every release.
    Regards,
    Sergio

  • Data Load from Source to Destination

    Hello Forum Members,
    I have to databases on different solaris servers.
    I have to load data from Source Table[Server 1] to Target Current Table[truncate existing data and load from Source Table] [Server 2]and on every Wednesday from Target Current Table into Target History Table[Data always appended][Server 2].
    Any advice how to solve the issue is highly appreciated.
    Thanks and Regards,
    Suresh

    Unless there is a history table on the remote server, the history table would not appear to be an appropriate candidate for a materialized view.
    You could, in theory, create the live table as a materialized view rather than doing a TRUNCATE (or DELETE) and INSERT INTO SELECT in your stored procedure. Since it sounds like you need to have some coordination between the refreshing of the materialized view and the population of the history table, it's not obvious that creating the live table as a materialized view buys you anything.
    Justin

  • How to restrict number of Data Records from Source system?

    Hi,
    How can I restrict the number of Data records from R3 source system that are being loaded into BI. For example I have 1000 source data records, but only wish to transfer the first 100. How can I achieve this? Is there some option in the DataSource definition or InfoPackage definition?
    Pls help,
    SD

    Hi SD,
    You can surely restrict the number of records, best and simplest way is, check which characteristics are present in selection screen of InfoPackage and check in R3, which characteristics if given a secection could fetch you the desired number of records. Use it as selection in InfoPackage.
    Regards,
    Pankaj

  • How to restrict no. of data packets from source system into BW

    Hi,
    when i try loading data from the IDES system to the PSA in BI 7.0. I get the following error from Data packet 41 onwards...
    Data Package 41 : arrived in BW ; Processing : Selected number does not agree with transferred n
    I have already read the various suggestions on the forum. Since I am loading data for testing purpose, i would like to restrict the data load until the 40th data packet. Is this possible & how?
    Pls help. Thanks
    SD

    Hi,
    I don't think there's a parameter you can set for the number of data packages.
    There's a parameter for the size of the datapackage. But that's different.
    Of course you can always try and restrict the data you load with the selection of your infopackage.
    Thats a more sensible way of doing it rather than restricting the number of data packages.
    Because even somehow if you managed to get them to 40, you won't know what data you will be missing that the datapackages from 41 onwards would have brought.  And this will lead to reconcilliation issues with R/3 data in your testing.
    -RMP

  • SMD - Diagnostic Administration -- Managed Systems takes long time

    Hi All,
    On SMD administration URL , http://system:50000/smd ---> Diagnostics administration ---> Agent Administration --> managed systems  takes very long time.
    Sometimes it never ends and keep on going...
    CPU utilization is also very high...
    Have anyone encountered this issue ?
    We are using Solman 7.00 on Solaris, DB is oracle.
    Thanks!

    Reply with ur System Technical Note,
    find below DOCS for Solman System Technical Config,
    https://websmp103.sap-ag.de/~sapdownload/011000358700000255562007E/Sizing_Guide_SM_70_SP15.pdf
    https://websmp103.sap-ag.de/~form/sapnet?_FRAME=OBJECT&_HIER_KEY=501100035870000015092&_HIER_KEY=601100035870000179416&_HIER_KEY=601100035870000179443&_HIER_KEY=701100035871000457214&#HOME
    Basically SMD is mostly dependant on J2ee Heap and Abap Configuration. Set up the required Heap and T-Code ST02 paramaters properly and call the SMD again.
    Regards
    Srikanth Mandalapu
    Edited by: srikanth mandalapu on Oct 10, 2008 12:11 AM

  • System takes long time to Save/Activate using SFP Form Builder

    We are planning to customize the Travel Expense PDF Form – PTRV_EXPENSE_FORM in ECC 6.0 using transaction SFP – FORM BUILDER (ABAP Workbench). As the initial step we copied it into Z form and started the changes.  Here is the problem: Even for a small change on the Layout, system is taking long time to complete the process.
    For example:
    For Saving the changes: 6 to 7 minutes
    For Activating the changes: 12 to 13 minutes.
    On the server side :
    we are on ECC 6.0, SAP NW04s SP 9 for EA-HR
    ADOBE LiveCycle Designer 7.1 (Patch 3 applied).
    On client machine :
    we have SAP GUI 640 Patch Level 21,
    ADOBE Reader 7.0.9.
    Client machine has 1 GB RAM & 40 GB Hard Drive.
    Thanks in Advance for the quick reply.

    Same problem for us

  • Understanding the data transfer from source system to PSA

    hi experts,
    i an new to BI. i have created a transactional data source for VBAK table and replicated that to BI. Created transformations and info package for uploading the data into PSA. i just want to know the sequence of steps that the system executes when we run the infopackage. i know that it makes a rfc call to source and brings the data back to BI. i want to where that RFC call happens and what is the RFC FM name etc.,. I tried to debug the process using /h, but as you know, it is very difficult to debug the whole standard code. it is very complex. i got lost some where in the middle. if anybody has any idea / or done any research on this ,please share your findings.
    thanks

    Hi,
    Once you click on Start buton in Infopackage,
    1. BW system wil send on erequest to ECC along with DataSource details.
    2. Based on BW request ECC will respond and it will send message OK and starts extraction.
    3. Based on InfoPackage selection the Data will executes in ECC and picks the data in Packets.
    4. And ECC will send the data in teh form of Datapackets to BW (Idocs)
    5. Once the selection creatiro is over the Job will finish.
    6. You can see the data in BW
    Thanks
    Reddy

  • Master Data Extraction from source system

    Hi BW Gurus,
    I need to extract the "<b>Purchase Document Type</b>" and "<b>Purchase Document Status</b>" master data from R3.
    Can anybody shed some light how can i find out which table in R3 host this data.
    Thank you in advance and i promise to assign point for your answers
    Thank you

    Hi,
    I feel the base table for Purchasing Document Type is T161.
    For Status of purchasing document there is no base table .the avilable value are comming from Available Domain values.
    With rgds,
    Anil Kumar Sharma .P

  • Downloading video from sdhc card takes long time in imovie

    I have a 3 month old 1mac running mountain Lion...When I try to import a 3 minit long video useing a sdhc card it take about 20 to 30 minits.. Sometimes it just quits... as if its timed out  what am I doing wrong?

    See if Image Capture will download it & how long it takes.

  • DataSource extraction very slow ( from Source System to PSA it takes 23 hrs

    Friends,
    We have enhanced the datasource 0CRM_SALES_ORDER_I with the user exit....after the enhancement i.e (adding the new fields and wrote some coding to enhance ...) the data extraction takes place for around 23 hours. there is approximately 2,50,000 records.
    Can you please suggest any steps to tune up the performance of the datasource.
    NOTE: Data Extraction from source system to PSA alone takes 23 hrs.once the data is arrived in PSA then the loading of data to cube is fast.
    PLZ help me to solve this issue.
    BASKAR

    Hi Friends,
    This is the code used for the datasource enhancement.(EXIT_SAPLRSAP_001)
    DATA : IS_CRMT_BW_SALES_ORDER_I LIKE CRMT_BW_SALES_ORDER_I.
    DATA:  MKT_ATTR TYPE STANDARD TABLE OF  CRMT_BW_SALES_ORDER_I.
    DATA: L_TABIX TYPE I.
    DATA: LT_LINK TYPE STANDARD TABLE OF CRMD_LINK,
          LS_LINK TYPE CRMD_LINK.
    DATA: LT_PARTNER TYPE STANDARD TABLE OF CRMD_PARTNER,
          LS_PARTNER TYPE CRMD_PARTNER.
    DATA: LT_BUT000 TYPE STANDARD TABLE OF BUT000,
          LS_BUT000 TYPE BUT000.
    DATA: GUID TYPE CRMT_OBJECT_GUID.
    DATA: GUID1 TYPE CRMT_OBJECT_GUID_TAB.
    DATA: ET_PARTNER TYPE CRMT_PARTNER_EXTERNAL_WRKT,
          ES_PARTNER TYPE CRMT_PARTNER_EXTERNAL_WRK.
    TYPES: BEGIN OF M_BINARY,
           OBJGUID_A_SEL TYPE CRMT_OBJECT_GUID,
            END OF M_BINARY.
    DATA: IT_BINARY TYPE STANDARD TABLE OF M_BINARY,
          WA_BINARY TYPE M_BINARY.
    TYPES : BEGIN OF M_COUPON,
             OFRCODE TYPE CRM_MKTPL_OFRCODE,
             END OF M_COUPON.
    DATA: IT_COUPON TYPE STANDARD TABLE OF M_COUPON,
          WA_COUPON TYPE M_COUPON.
    DATA: CAMPAIGN_ID TYPE CGPL_EXTID.
    TYPES : BEGIN OF M_ITEM,
             GUID TYPE CRMT_OBJECT_GUID,
            END OF M_ITEM.
    DATA: IT_ITEM TYPE STANDARD TABLE OF M_ITEM,
          WA_ITEM TYPE M_ITEM.
    TYPES : BEGIN OF M_PRICE,
                  KSCHL TYPE PRCT_COND_TYPE,
                  KWERT  TYPE PRCT_COND_VALUE,
                  KBETR   TYPE PRCT_COND_RATE,
            END OF M_PRICE.
    DATA: IT_PRICE TYPE STANDARD TABLE OF M_PRICE,
          WA_PRICE TYPE M_PRICE.
    DATA: PRODUCT_GUID TYPE COMT_PRODUCT_GUID.
    TYPES : BEGIN OF M_FRAGMENT,
             PRODUCT_GUID TYPE COMT_PRODUCT_GUID,
             FRAGMENT_GUID TYPE COMT_FRG_GUID,
             FRAGMENT_TYPE TYPE COMT_FRGTYPE_GUID,
           END OF M_FRAGMENT.
    DATA: IT_FRAGMENT TYPE STANDARD TABLE OF M_FRAGMENT,
          WA_FRAGMENT TYPE M_FRAGMENT.
    TYPES : BEGIN OF M_UCORD,
          PRODUCT_GUID TYPE     COMT_PRODUCT_GUID,
          FRAGMENT_TYPE     TYPE COMT_FRGTYPE_GUID,
          ZZ0010 TYPE     Z1YEARPLAN,
            ZZ0011 TYPE Z6YAERPLAN_1,
            ZZ0012 TYPE Z11YEARPLAN,
            ZZ0013 TYPE Z16YEARPLAN,
            ZZ0014 TYPE Z21YEARPLAN,
         END OF M_UCORD.
    DATA: IT_UCORD TYPE STANDARD TABLE OF M_UCORD,
          WA_UCORD TYPE M_UCORD.
    DATA: IT_CATEGORY TYPE STANDARD TABLE OF COMM_PRPRDCATR,
          WA_CATEGORY TYPE COMM_PRPRDCATR.
    DATA: IT_CATEGORY_MASTER TYPE STANDARD TABLE OF ZPROD_CATEGORY ,
          WA_CATEGORY_MASTER TYPE ZPROD_CATEGORY .
    types : begin of st_final,
               OBJGUID_B_SEL  TYPE CRMT_OBJECT_GUID,
               OFRCODE TYPE CRM_MKTPL_OFRCODE,
               PRODJ_ID TYPE CGPL_GUID16,
               OBJGUID_A_SEL type     CRMT_OBJECT_GUID,
              end of st_final.
    data : t_final1 type  standard table of st_final.
    data : w_final1 type  st_final.
    SELECT  bOBJGUID_B_SEL aOFRCODE  aPROJECT_GUID bOBJGUID_A_SEL  INTO table t_final1 FROM
       CRMD_MKTPL_COUP as a  inner join CRMD_BRELVONAE as b on  bOBJGUID_A_SEL = aPROJECT_GUID .

  • Data Load from a DSO to a Cube

    Hello,
    I am facing a problem
    Case:
    There is a DSO with following Char/Key Figure:
    Characteristics     ( Value-1 )     ( Value-2 )     ( Value-3 )
    A(Key)          ( A1 )     ( A2 )     ( A3 )
    B(Key)          ( B1 )     ( B2 )     ( B3 )
    C          ( C1 )     ( C1 )     ( C1 )
    D          ( D1 )     ( D1 )     ( D1 )
    E          ( E1 )     ( E1 )     ( E1 )
    F          ( F1 )     ( F1 )     ( F1 )
    KF1(Key Figure)     ( 10 )     ( 20 )     ( 30 )
    I got below data from the DSO to a Cube(structure is below with data) details are below:
    Dimention          ( Values )     
    C          ( C1 )          
    D          ( D1 )          
    E          ( E1 )          
    F          ( F1 )          
    KF1(Key Figure)     ( 50 )
    My Question is why i am not getting KF1 value as 60.
    When i tried with including Key Fields of the DSO, i got all the records in the Cubes.
    Where i am mistaking?
    Can some one correct me please?
    Thanks.
    Avinash.

    Below are the basic steps which we follow in any BI 2004S system:
    1)Create datasource. Here u can set/check the Soucre System fields.
    2)Create Transformation for that datasource. (no more update rules/transfer rules)
    2.1) While creating transformation for DS it will ask you for data target name, so just assign where u want to update ur data.
    DataSource -> Transformation -> Data Target
    Now if you want to load data into data target from Source System Datasource:
    1) Create infopackage for that data source. If you are creating infopackage for new datasources, it will only allow you update upto PSA, all other options u can see as disabled.
    2)Now Create DTP (Data Transfer Process) for that data source.
    3) NOw schdule the Infopackage, once the data is loaded to PSA, you can execute your DTP which will load data to data target.
    If you are loading data from one one data target to other, no need to use PSA, you can directly execute DTP in that case.
    Data Source -> Transformation (IP/DTP) -> Data Target1 -> DTP ->Data Target 2
    Use the below link for detailed example:
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/fc61e12d-0a01-0010-2883-e2fc63ef729b
    Infosources are no more mandatory with BI 7.0, below is the link to scenarios where we use infosources:
    http://help.sap.com/saphelp_nw04s/helpdata/en/44/0243dd8ae1603ae10000000a1553f6/content.htm
    Full or delta depends on your requirement...
    chk the below thread to know better
    difference between the various loads
    hope it helps
    Message was edited by:
            sriram viswanathan

  • MIRO take long time when enter invoice for PO with GR-based invoice

    Hi,
    In my client system, system take long time for extracting PO data while booking invoice via. MIRO for the purchase order which have GR-based IV is marked. However system take few second for extracting PO data when enter invoice via. MIRO for the PO which don't have GR-based IV. Please note following point while providing the solution:
    - this problem is exist only for the purchase order related to one company code. However system working perfectly for other company code in the same client. Hence we assuming that some company code level cofiguration is missing.
    - the problem is exist for po with account assignment K.
    - we have one to one mapping for purchase organization to company code to plant.
    Appreciate for you quick respond. Thanks in advance.
    Regards,
    sp sahu

    Hi,
    Please check with FI guy for GL A/c and Cost centers which you are using to create the PO with item category K.
    Still problem permits check with your ABAP person.
    Regards,
    Mohd Ali.

Maybe you are looking for

  • Spin the wheel of weird boot problems!

    Hey! First time Mac owner here. So, I bought this poor little 13" white Macbook off a guy who had clearly abused it and I'm trying to nurse it back to health. I'm trying to fix it, and then wipe the hard drive and install a fresh copy of the best OS

  • Generate INSERT Statements question

    Is there a built-in PL/SQL procedure that SQL Developer utilizes to generate the INSERT statements when one exports data from a worksheet in 'insert' format? Is it part of the DBMS_METADATA package? Mike

  • Resolution resetting on start up

    Hey everyone, I am running into an issue that the screen resolution of my imac keeps resetting to scaled 1600*900. I can use system preferences to change this to 1080p but every time I restart the computer it is back to 1600*900. I have tried clearin

  • Using root bridge as a fallback radius server for WPA and EAP

    From reading the different documentation out there, it seems that one should be able to configure a root bridge as a fallback radius server in case a primary radius server were to be unreachable. Has anyone encountered this situation? And could they

  • Manage Access to Named Credentials via EMCLI

    Hi Colleagues, does anyone know how to manage access to Named Credentials via EMCLI or does anyone know if this function exist in EMCLI. We want to configure the access via scripts, so that we can for example grant access for all database administrat