Doesn't load from source system to PSA and DSO

Hi Gurus,
We have launched a Delta load from Source System (POSDM) to DSO. In the source system there are 5000000 rows but none was loaded into DSO or PSA. The delta load was ok, but any data was loaded.
If the delta  load was lunch again, would be loaded those 5000000 rows. Or those 5000000 rows wouldn't be able to load again.
Any idea abut this issue? Any feedback will be really appreciated.
Thanks in advance.

Hi David,
Are you sure this 5 million records are delta records and should be pulled into BW as delta??
did you count the number of records in the underlying tables??
which data source you are using??
Delta loads are suppose to bring the new records or changed records since the last data load and not all the records.
Since the request is green as you said and still it shows 0 records then it is possible that nothing has changed since last delta.try to see the details of the data load... if it failed...may be thats the reason you are not able to see any records.
If you schedule the delta again it will bring the records changed or created after last delta.
If the delta was unsuccessfull then turn the QM status and overall status of request to red manually in the monitor of the request and delete it from PSA and all the targets in which its loaded and schedule the delta again and it should bring the delta.
Thanks
Ajeet
Thanks
Ajeet

Similar Messages

  • DataSource extraction very slow ( from Source System to PSA it takes 23 hrs

    Friends,
    We have enhanced the datasource 0CRM_SALES_ORDER_I with the user exit....after the enhancement i.e (adding the new fields and wrote some coding to enhance ...) the data extraction takes place for around 23 hours. there is approximately 2,50,000 records.
    Can you please suggest any steps to tune up the performance of the datasource.
    NOTE: Data Extraction from source system to PSA alone takes 23 hrs.once the data is arrived in PSA then the loading of data to cube is fast.
    PLZ help me to solve this issue.
    BASKAR

    Hi Friends,
    This is the code used for the datasource enhancement.(EXIT_SAPLRSAP_001)
    DATA : IS_CRMT_BW_SALES_ORDER_I LIKE CRMT_BW_SALES_ORDER_I.
    DATA:  MKT_ATTR TYPE STANDARD TABLE OF  CRMT_BW_SALES_ORDER_I.
    DATA: L_TABIX TYPE I.
    DATA: LT_LINK TYPE STANDARD TABLE OF CRMD_LINK,
          LS_LINK TYPE CRMD_LINK.
    DATA: LT_PARTNER TYPE STANDARD TABLE OF CRMD_PARTNER,
          LS_PARTNER TYPE CRMD_PARTNER.
    DATA: LT_BUT000 TYPE STANDARD TABLE OF BUT000,
          LS_BUT000 TYPE BUT000.
    DATA: GUID TYPE CRMT_OBJECT_GUID.
    DATA: GUID1 TYPE CRMT_OBJECT_GUID_TAB.
    DATA: ET_PARTNER TYPE CRMT_PARTNER_EXTERNAL_WRKT,
          ES_PARTNER TYPE CRMT_PARTNER_EXTERNAL_WRK.
    TYPES: BEGIN OF M_BINARY,
           OBJGUID_A_SEL TYPE CRMT_OBJECT_GUID,
            END OF M_BINARY.
    DATA: IT_BINARY TYPE STANDARD TABLE OF M_BINARY,
          WA_BINARY TYPE M_BINARY.
    TYPES : BEGIN OF M_COUPON,
             OFRCODE TYPE CRM_MKTPL_OFRCODE,
             END OF M_COUPON.
    DATA: IT_COUPON TYPE STANDARD TABLE OF M_COUPON,
          WA_COUPON TYPE M_COUPON.
    DATA: CAMPAIGN_ID TYPE CGPL_EXTID.
    TYPES : BEGIN OF M_ITEM,
             GUID TYPE CRMT_OBJECT_GUID,
            END OF M_ITEM.
    DATA: IT_ITEM TYPE STANDARD TABLE OF M_ITEM,
          WA_ITEM TYPE M_ITEM.
    TYPES : BEGIN OF M_PRICE,
                  KSCHL TYPE PRCT_COND_TYPE,
                  KWERT  TYPE PRCT_COND_VALUE,
                  KBETR   TYPE PRCT_COND_RATE,
            END OF M_PRICE.
    DATA: IT_PRICE TYPE STANDARD TABLE OF M_PRICE,
          WA_PRICE TYPE M_PRICE.
    DATA: PRODUCT_GUID TYPE COMT_PRODUCT_GUID.
    TYPES : BEGIN OF M_FRAGMENT,
             PRODUCT_GUID TYPE COMT_PRODUCT_GUID,
             FRAGMENT_GUID TYPE COMT_FRG_GUID,
             FRAGMENT_TYPE TYPE COMT_FRGTYPE_GUID,
           END OF M_FRAGMENT.
    DATA: IT_FRAGMENT TYPE STANDARD TABLE OF M_FRAGMENT,
          WA_FRAGMENT TYPE M_FRAGMENT.
    TYPES : BEGIN OF M_UCORD,
          PRODUCT_GUID TYPE     COMT_PRODUCT_GUID,
          FRAGMENT_TYPE     TYPE COMT_FRGTYPE_GUID,
          ZZ0010 TYPE     Z1YEARPLAN,
            ZZ0011 TYPE Z6YAERPLAN_1,
            ZZ0012 TYPE Z11YEARPLAN,
            ZZ0013 TYPE Z16YEARPLAN,
            ZZ0014 TYPE Z21YEARPLAN,
         END OF M_UCORD.
    DATA: IT_UCORD TYPE STANDARD TABLE OF M_UCORD,
          WA_UCORD TYPE M_UCORD.
    DATA: IT_CATEGORY TYPE STANDARD TABLE OF COMM_PRPRDCATR,
          WA_CATEGORY TYPE COMM_PRPRDCATR.
    DATA: IT_CATEGORY_MASTER TYPE STANDARD TABLE OF ZPROD_CATEGORY ,
          WA_CATEGORY_MASTER TYPE ZPROD_CATEGORY .
    types : begin of st_final,
               OBJGUID_B_SEL  TYPE CRMT_OBJECT_GUID,
               OFRCODE TYPE CRM_MKTPL_OFRCODE,
               PRODJ_ID TYPE CGPL_GUID16,
               OBJGUID_A_SEL type     CRMT_OBJECT_GUID,
              end of st_final.
    data : t_final1 type  standard table of st_final.
    data : w_final1 type  st_final.
    SELECT  bOBJGUID_B_SEL aOFRCODE  aPROJECT_GUID bOBJGUID_A_SEL  INTO table t_final1 FROM
       CRMD_MKTPL_COUP as a  inner join CRMD_BRELVONAE as b on  bOBJGUID_A_SEL = aPROJECT_GUID .

  • Data loading from source system takes long time.

    Hi,
         I am loading data from R/3 to BW. I am getting following message in the monitor.
    Request still running
    Diagnosis
    No errors could be found. The current process has probably not finished yet.
    System response
    The ALE inbox of the SAP BW is identical to the ALE outbox of the source system
    and/or
    the maximum wait time for this request has not yet run out
    and/or
    the batch job in the source system has not yet ended.
    Current status
    in the source system
    Is there anything wrong with partner profile maintanance in the source system.
    Cheers
    Senthil

    Hi,
    I will suggest you to check a few places where you can see the status
    1) SM37 job log (In source system if load is from R/3 or in BW if its a datamart load) (give request name) and it should give you the details about the request. If its active make sure that the job log is getting updated at frequent intervals.
    Also see if there is any 'sysfail' for any datapacket in SM37.
    2) SM66 get the job details (server name PID etc from SM37) and see in SM66 if the job is running or not. (In source system if load is from R/3 or in BW if its a datamart load). See if its accessing/updating some tables or is not doing anything at all.
    3) RSMO see what is available in details tab. It may be in update rules.
    4) ST22 check if any short dump has occured.(In source system if load is from R/3 or in BW if its a datamart load)
    5) SM58 and BD87 for pending tRFCs and IDOCS.
    Once you identify you can rectify the error.
    If all the records are in PSA you can pull it from the PSA to target. Else you may have to pull it again from source infoprovider.
    If its running and if you are able to see it active in SM66 you can wait for some time to let it finish. You can also try SM50 / SM51 to see what is happening in the system level like reading/inserting tables etc.
    If you feel its active and running you can verify by checking if the number of records has increased in the data tables.
    SM21 - System log can also be helpful.
    Thanks,
    JituK

  • Understanding the data transfer from source system to PSA

    hi experts,
    i an new to BI. i have created a transactional data source for VBAK table and replicated that to BI. Created transformations and info package for uploading the data into PSA. i just want to know the sequence of steps that the system executes when we run the infopackage. i know that it makes a rfc call to source and brings the data back to BI. i want to where that RFC call happens and what is the RFC FM name etc.,. I tried to debug the process using /h, but as you know, it is very difficult to debug the whole standard code. it is very complex. i got lost some where in the middle. if anybody has any idea / or done any research on this ,please share your findings.
    thanks

    Hi,
    Once you click on Start buton in Infopackage,
    1. BW system wil send on erequest to ECC along with DataSource details.
    2. Based on BW request ECC will respond and it will send message OK and starts extraction.
    3. Based on InfoPackage selection the Data will executes in ECC and picks the data in Packets.
    4. And ECC will send the data in teh form of Datapackets to BW (Idocs)
    5. Once the selection creatiro is over the Job will finish.
    6. You can see the data in BW
    Thanks
    Reddy

  • Data Load from Source to Destination

    Hello Forum Members,
    I have to databases on different solaris servers.
    I have to load data from Source Table[Server 1] to Target Current Table[truncate existing data and load from Source Table] [Server 2]and on every Wednesday from Target Current Table into Target History Table[Data always appended][Server 2].
    Any advice how to solve the issue is highly appreciated.
    Thanks and Regards,
    Suresh

    Unless there is a history table on the remote server, the history table would not appear to be an appropriate candidate for a materialized view.
    You could, in theory, create the live table as a materialized view rather than doing a TRUNCATE (or DELETE) and INSERT INTO SELECT in your stored procedure. Since it sounds like you need to have some coordination between the refreshing of the materialized view and the population of the history table, it's not obvious that creating the live table as a materialized view buys you anything.
    Justin

  • Error in loading data into PSA from source system.

    Hi Experts !!!
    Good morning .
    I am trying to load data from SRM source system into BI system.When I execute infopackage,the data is not loaded into PSA and status of the process immediately becomes yellow.After sometime it gets timed out and converts to RED.When I checked the error documentation I found that the iDOCS response was not received from the source system.Detailed error is as below.
    System Response
    There are idocs in the source system ALE outbox that did not arrive in the ALE inbox of BI.
    Further Analysis
    Check the TRFC log
    you can access this log using the wizard OR menu path "Environment > Transact. RFC> In the source system
    However,I was not able to navigate to this path.I think if I can navigate to this path,I can manually push IDOCS from source system to BI system.
    Regards,
    Mandar.

    Hi,
    Check the DataSource in RSA3, if it is working fine and able to see the data in RSA3, there is no problem in DS level, then check the Mappings and any routines in BW for that DS, if this is also fine then check the below options.
    See Dumps in ST22, SM21 also.
    Check RFC Connection between ECC and BW systems, i.e. RSA1-->Source System->Right Click on Source system and Check.
    You must have the following profiles to BWREMOTE or ALEREMOTE users.So add it. Bcoz either of these two users will use in background to get extract the data from ECC, so add these profiels in BW.
    S_BI-WHM_RFC, S_BI-WHM_SPC, S_BI-WX_RFC
    And also check the following things.
    1.Connections from BW to ECC and ECC to BW in SM59
    2.Check Port,Partner Profiles,and Message Types in WE20 in ECC & BW.
    3.Check Dumps in ST22, and SM21.
    4.If Idocs are stuck i.e see the OLTP Idoc numbers in RSMO Screen in (BW) detials tab see in bottom, you can see OLTP Idoc number and take the Idoc numbers and then goto to ECC see the status in WE05 or WE02, if error then check the log else goto to BD87 in ECC and give the Idoc numbers and execute manually and see in RSMO and refresh.
    5.Check the LUWs struck in SM58,User Name = * (star) and run it and see Strucked LUWs and select our LUW and execute manually and see in RSMO in BW.
    See in SDN
    Re: Loading error in the production  system
    IDOC getting struck at TRFC (SM58)
    1.Reorg of table (ARFCSSTATE )
    2.Incraesing the resources in the system (no of processes and memory ) , as this issue happens owing to this.
    Source system tries to send TRFC to target and if there are no WP's available it will come to Transaction Recorded state, and form here it will not try to send this TRFC.So you have to execute this manually.
    Also we can increase the timeout parameter so that it can try few more times to send before actually it comes to recorded state.
    Regards,
    Suman

  • Error Occures while loading data from Source system to Target ODS

    Hi..
    I started loading Records From source system to target ODS.while i running the job i got the following errors.
    Record 18211 :ERROR IN HOLIDAY_GET 20011114 00000000
    Record 18212 :ERROR IN HOLIDAY_GET 20011114 00000000
    sp Please help me in these following Errors..
    Thanks in advance,

    Hello
    How r u ?
    I think this problem is at the ODS level, ZCAM_O04 is ur ODS Name.
    Could u check the ODS Settings, and the Unique Data Records is Checked or Not ?
    Best Regards....
    Sankar Kumar
    +91 98403 47141

  • Psa data is deviated from source system data...

    Hi all,
    My source data was manipulated when it reaches to PSA. We r extracting the data from Flatfiles. And v have one keyfigure as no.of months (say 12,13,14). When am extracting this data to my Infocube this particular fields has changed as 12,000 and 13,000 and 14,000. Even though it is appearing in Simulation under Schduler tab in the same way(12,000 and 13,000 and 14,000).Am not multiplying this field with any value under updaterules.
    And the datatype for keyfigure is Number.
    And PSA data should always similar to source system data.
    But here also am getting deviations. Let me inform me what are the routines and conversion takes place, before the data reaching to PSA from source system.

    Hi Suresh,
    Don't worry. The field presentation in the cube depends on the field type.
    And I guess that comma stands for a decimal delimiter.
    And if so, there is no difference at all!
    Check your settings.
    Best regards,
    Eugene
    Message was edited by: Eugene Khusainov

  • How to  upload Transaction data from source system to BI 7.0

    Dear friends,
    I want to know how to  upload Transaction data from source system to BI 7.0.
    if anybody having step by step material for this process ,please send it to my
    mail id.
    [email protected]
    [email protected]
    Thanks,
    D.prabhu

    Hi,
    it works in the same way as before in case you are not using the new features like the dtp and the transformations. Using the new features, you schedule a infopackage to load the data up to psa and schedule a dtp to post the data to the targets.
    regards
    Siggi
    PS: We want to share knowledge in sdn and not keeping it private by having the information sent to our email adresses.

  • Error message from source system

    Hi All,
    We have an infocube and ods which is being loaded from another ods. This is included in a process chain.
    In the process chain the Infopackage execution step has failed. When I right click and go for display messages, I found the following message in Messages tab:
    Error message from the source system
    Diagnosis
    An error occurred in the source system.
    System response
    Caller 09 contains an error message.
    Further analysis:
    The error occurred in Extractor .
    Refer to the error message.
    Procedure
    How you remove the error depends on the error message.
    Note
    If the source system is a Client Workstation, then it is possible that the file that you wanted to load was being edited at the time of the data request. Make sure that the file is in the specified directory, that it is not being processed at the
    And in Indiv Messages tab,
    Delta upload is only possible after successful initialization.
    Errors in source system.
    I tried to update the data targets from the initial ODS manualy but in vain.
    Any Help is awarded with points.
    Thanks
    AP

    Hi,
      Goto Info pack -> on menu bar select scheduler -> select init options for source system -> one request with green status -> select it and  click on delete symbol , it will delete the init. you can cross verify this in the base ODS where you dint find data mart status(after refresh). now  select init delta process under that select init without data transfer and schedule the load it will update one record and init flag to your source ODS.
    Note: if the last request is not updated from soruce ods to targets for that also you will find data mart status for this select the particular data mart -> click on delete symbol it will delete for the particular request (if it is updated to target it will not allow you to delete) then do delta load it will update the request which is not loaded from source ods to next targets.
    Regards
    Sankar

  • Error in Process Chain while extracting data from source system

    Hi All,
    Dail we are facing problem while retrieving the data from source system for the data source 0TB_AMOUNT1 in the process chain, it is giving error "Function module BANK_TMC_API_SIM_GET does not exist Furnction module". when I repeat the step than it is going fine.
    Please let me know why this error is coming for first why it is not coming after repeating the step?
    Thanks & Regards,
    Murali.

    Hi Murali,
    Hope you are doing good..
    Is it a full or Delta load on a daily basis? Did you check if the FM exists in your system?
    Did you try debugging the load in source system and check if the extractor is trying to call FM you mention.
    If possible please post the exact error message you are receiving so that we could get an idea whats the error and why you are encountering it.
    As of now what i can suggest is, go to RSA2 in source system, give the datasource name and check if the extractor code has FM in it.
    Regards,
    Anil Pragada.

  • DataSource for 0SD_C02 doesn't exist in Source System

    Hi Friends,
    I think the title explain all the problem.
    I want to load the 0SD_C02 infocube and I know that I have to do it with the infosource 2LIS_01_S005, but when I go to R/3 to search for it, the system tells me that it doesn't exist.
    I've proved with the Tcode RSA3 to verify the extractor and it gives me the same answer "Infosource 2LIS_01_S005 doesn't exist in Source System".
    Does anybody know how to load this infocube?
    Thanks in advance.
    Francis.
    Following with the same problem.
    I want to load data on 0SD_C02, 0SD_C03 & 0SD_C05. C03 & C05 has datasources named like 2LIS_11/2/3... I have these ones in R/3, but the datasources named like 2LIS_01_* aren't exist in R/3,
    How can I load data in these infocubes?
    Thanks in advance,
    Francis.
    Message was edited by:
            Francis Estevez

    I think you have to install it via RSA5 transaction in your source system...
    Hope it helps!
    Bye,
    Roberto

  • To retain the same file name in target(File) that coming from source system(AS2).

    Hi All,
    I have a requirement to send the two file from the third party system(GXS) to SAP ECC via SAP PI(version - 7.4). We are using AS2 adapter at the sender side to pick the file and using File Adapter to place the file in ECC.
    Its a pass through scenarion(No ESR part is used).
    Requirement:
    We are going to get 2 files ->Example:  one from Axis and another one from ICICI -> each bank file will have a file name ->
    We have to post the file at the target system with the same name as source file name, the source file name could come dynamically(with diff time stamp).
    Please let me know, Is it possible to achieve the above requirement(we have to retain the file name in target that coming from source system) with Sender AS2 adapter, If yes then please suggest the best approach.
    Regards,
    Pradeep Bawlia

    Hi Guys,
    The above requirement is solved now.
    Solution - I used java mapping for extracting the same source file name and assignning the same to the target system.
    Code:
    public class ExtractFileName_ASMA extends AbstractTransformation{
    public void transform(TransformationInput in, TransformationOutput out) throws StreamTransformationException  
    DynamicConfiguration conf=in.getDynamicConfiguration();
    DynamicConfigurationKey key_AS2FileNmae = DynamicConfigurationKey.create("http://sap.com/xi/XI/AS2/AS2", "AS2Filename");
    DynamicConfigurationKey key_FileNmae = DynamicConfigurationKey.create("http:/"+"/sap.com/xi/XI/System/File", "FileName");
    String As2FileName=conf.get(key_AS2FileNmae);
    conf.put(key_FileNmae, As2FileName);
    try{
    byte b[]=new byte[in.getInputPayload().getInputStream().available()];
    in.getInputPayload().getInputStream().read(b);
    out.getOutputPayload().getOutputStream().write(b);
    catch(Exception e)
    2. Another solution we can use the graphical mapping and need to use the dynamic config UDF at mapping level to acheive this requirement but here is one drawback - we need to do content conversion is this case.
    In case of Java mapping, no need to play with source file.
    WITHOUT USING ESR, WE CAN NOT ACHEIVE THIS REQUIREMENT.
    Regards,
    Pradeep Bawlia

  • Source system in PSA deletion variant

    Dear Experts
    In process chain PSA deletion variant,in source system column I will give the Development source system name(C6D),and after transport to the Quality styem(C6Q) Should I change this source system manually in qualtiy or system automatically will pick up the corresponding source system name?If not,what is the possible solution since we dont have any access in quality system.
    Cheers
    Rajesh

    Thanks for the reply.I got doubt because in DEV we have to enter manually both DS  and SOURCE SYSTEM manually...after importing to Quality will it change the source system name?...Can i take for grant...Pls confirm....
    Rajesh

  • Mapping new product code from source system to old product code in Oracle COA

    Mapping new product code from source system to old product code in Oracle COA
    Here is the situation:
    One of our legal entities is changing their product codes in their COA segment. We are not changing the code structure in the global SOB. Is there a workaround to pull the source system information into ADI (or any other tool) and map it to the current Oracle product code?
    Thanks and regards,
    [email protected]

    Only The  Data of Your Code will Change
    then if you are trying to copy a standarad sap-code , you need to copy
    every include, FM
    there is a Copy option over there in program
    you can do tat
    thnkx bhanu

Maybe you are looking for

  • Merging Calendars (on my mac with Mobile me)

    I have most of my calendars set up with my Mobile Me account to push to the cloud and to my iphone.  I also have some old calendars that are grouped under "On My Mac" and they do not sync.  Is there a way I can group the "On My Mac" calenars into the

  • Out of control volume control

    My HP Pavilion dv7=-3057 laptop has an out of control volume icon and control. I believe that it is possessed . It comes on all by itself.  The blue line runs back and forth eventually reaching the highest volume point. It sometimes goes off by itsel

  • Quicktime Pro and getting videos off the web

    I am bound and determined to get videos off the web and use them to make my own videos and what not. I was wondering if quicktime pro is the best program to use for my project. I have downloaded other programs that don't work and I am sick of the has

  • OLE DB Destination" failed validation and returned validation status "VS_ISBROKEN.

    Hi I have a package which has 10 sequence containers. Have 10 Flat File Source files and all is working perfect on BI Debug mode. When I deploy this package to server I am getting the following error. I have changed the property Retain the same conne

  • Downloading iWork - purchased '09 but won't show up in App Store

    Hi, I downloaded iWork for $79 back on August 10th, 2010.  I recently fresh installed Lion and of course lost all data.  I assumed everything I bought from Apple whould show up on the App Store, but not my iWork.  Now, too make things worse, there is