Field mismatched from Source system

Hi ,
        In BI Data target,for ZZAFLD000012 field  , data is not coming from CRM source system , when checking in CRM side ,it doesn't contain any data & mapped  in the Standard Data source 0CRM_SRV_PROCESS_H  of  BI . In CRM .
  In CRM source system ZZAFLD00003P  fields  contains the REQUIRED  Data ,which is  not mapped in the data source : 0CRM_SRV_PROCESS_H of   BI .
How can i bring the data to BI ?what are the authorization is required in CRM-user ?
Thanks in Advance,
Regards,
kalim

Hi,
I will suggest you to check a few places where you can see the status
1) SM37 job log (In source system if load is from R/3 or in BW if its a datamart load) (give request name) and it should give you the details about the request. If its active make sure that the job log is getting updated at frequent intervals.
Also see if there is any 'sysfail' for any datapacket in SM37.
2) SM66 get the job details (server name PID etc from SM37) and see in SM66 if the job is running or not. (In source system if load is from R/3 or in BW if its a datamart load). See if its accessing/updating some tables or is not doing anything at all.
3) RSMO see what is available in details tab. It may be in update rules.
4) ST22 check if any short dump has occured.(In source system if load is from R/3 or in BW if its a datamart load)
5) SM58 and BD87 for pending tRFCs.
Once you identify you can rectify the error.
If all the records are in PSA you can pull it from the PSA to target. Else you may have to pull it again from source infoprovider.
If its running and if you are able to see it active in SM66 you can wait for some time to let it finish. You can also try SM50 / SM51 to see what is happening in the system level like reading/inserting tables etc.
If you feel its active and running you can verify by checking if the number of records has increased in the data tables.
Thanks,
JituK

Similar Messages

  • Question to Transferring Global Settings from Source Systems

    1) When we do this steps, we found the following information as follows, so the question is that which fields that we should maintain?
    BW User ALE
    Debugging User
    BEx order number
    Releasest. hier
    Monitor initial
    Last IDoc no.
    Last IS-Nr.
    XPRA executed
    Lock Mangr Logs
    MD: Sing. Rec. Inst
    RFC BI Accelerator
    2) If we maintain some of fields should we do something before, and how? e.g. we know BW user ALE = aleremote, so should we maintain aleremote user using SU01 both in R/3 and BW before we do Transferring Global Settings from Source Systems
    Thanks for the answer.

    Is anyone know how to setting Transferring Global Settings from Source Systems, your answer will be high appreciate.

  • DataSource extraction very slow ( from Source System to PSA it takes 23 hrs

    Friends,
    We have enhanced the datasource 0CRM_SALES_ORDER_I with the user exit....after the enhancement i.e (adding the new fields and wrote some coding to enhance ...) the data extraction takes place for around 23 hours. there is approximately 2,50,000 records.
    Can you please suggest any steps to tune up the performance of the datasource.
    NOTE: Data Extraction from source system to PSA alone takes 23 hrs.once the data is arrived in PSA then the loading of data to cube is fast.
    PLZ help me to solve this issue.
    BASKAR

    Hi Friends,
    This is the code used for the datasource enhancement.(EXIT_SAPLRSAP_001)
    DATA : IS_CRMT_BW_SALES_ORDER_I LIKE CRMT_BW_SALES_ORDER_I.
    DATA:  MKT_ATTR TYPE STANDARD TABLE OF  CRMT_BW_SALES_ORDER_I.
    DATA: L_TABIX TYPE I.
    DATA: LT_LINK TYPE STANDARD TABLE OF CRMD_LINK,
          LS_LINK TYPE CRMD_LINK.
    DATA: LT_PARTNER TYPE STANDARD TABLE OF CRMD_PARTNER,
          LS_PARTNER TYPE CRMD_PARTNER.
    DATA: LT_BUT000 TYPE STANDARD TABLE OF BUT000,
          LS_BUT000 TYPE BUT000.
    DATA: GUID TYPE CRMT_OBJECT_GUID.
    DATA: GUID1 TYPE CRMT_OBJECT_GUID_TAB.
    DATA: ET_PARTNER TYPE CRMT_PARTNER_EXTERNAL_WRKT,
          ES_PARTNER TYPE CRMT_PARTNER_EXTERNAL_WRK.
    TYPES: BEGIN OF M_BINARY,
           OBJGUID_A_SEL TYPE CRMT_OBJECT_GUID,
            END OF M_BINARY.
    DATA: IT_BINARY TYPE STANDARD TABLE OF M_BINARY,
          WA_BINARY TYPE M_BINARY.
    TYPES : BEGIN OF M_COUPON,
             OFRCODE TYPE CRM_MKTPL_OFRCODE,
             END OF M_COUPON.
    DATA: IT_COUPON TYPE STANDARD TABLE OF M_COUPON,
          WA_COUPON TYPE M_COUPON.
    DATA: CAMPAIGN_ID TYPE CGPL_EXTID.
    TYPES : BEGIN OF M_ITEM,
             GUID TYPE CRMT_OBJECT_GUID,
            END OF M_ITEM.
    DATA: IT_ITEM TYPE STANDARD TABLE OF M_ITEM,
          WA_ITEM TYPE M_ITEM.
    TYPES : BEGIN OF M_PRICE,
                  KSCHL TYPE PRCT_COND_TYPE,
                  KWERT  TYPE PRCT_COND_VALUE,
                  KBETR   TYPE PRCT_COND_RATE,
            END OF M_PRICE.
    DATA: IT_PRICE TYPE STANDARD TABLE OF M_PRICE,
          WA_PRICE TYPE M_PRICE.
    DATA: PRODUCT_GUID TYPE COMT_PRODUCT_GUID.
    TYPES : BEGIN OF M_FRAGMENT,
             PRODUCT_GUID TYPE COMT_PRODUCT_GUID,
             FRAGMENT_GUID TYPE COMT_FRG_GUID,
             FRAGMENT_TYPE TYPE COMT_FRGTYPE_GUID,
           END OF M_FRAGMENT.
    DATA: IT_FRAGMENT TYPE STANDARD TABLE OF M_FRAGMENT,
          WA_FRAGMENT TYPE M_FRAGMENT.
    TYPES : BEGIN OF M_UCORD,
          PRODUCT_GUID TYPE     COMT_PRODUCT_GUID,
          FRAGMENT_TYPE     TYPE COMT_FRGTYPE_GUID,
          ZZ0010 TYPE     Z1YEARPLAN,
            ZZ0011 TYPE Z6YAERPLAN_1,
            ZZ0012 TYPE Z11YEARPLAN,
            ZZ0013 TYPE Z16YEARPLAN,
            ZZ0014 TYPE Z21YEARPLAN,
         END OF M_UCORD.
    DATA: IT_UCORD TYPE STANDARD TABLE OF M_UCORD,
          WA_UCORD TYPE M_UCORD.
    DATA: IT_CATEGORY TYPE STANDARD TABLE OF COMM_PRPRDCATR,
          WA_CATEGORY TYPE COMM_PRPRDCATR.
    DATA: IT_CATEGORY_MASTER TYPE STANDARD TABLE OF ZPROD_CATEGORY ,
          WA_CATEGORY_MASTER TYPE ZPROD_CATEGORY .
    types : begin of st_final,
               OBJGUID_B_SEL  TYPE CRMT_OBJECT_GUID,
               OFRCODE TYPE CRM_MKTPL_OFRCODE,
               PRODJ_ID TYPE CGPL_GUID16,
               OBJGUID_A_SEL type     CRMT_OBJECT_GUID,
              end of st_final.
    data : t_final1 type  standard table of st_final.
    data : w_final1 type  st_final.
    SELECT  bOBJGUID_B_SEL aOFRCODE  aPROJECT_GUID bOBJGUID_A_SEL  INTO table t_final1 FROM
       CRMD_MKTPL_COUP as a  inner join CRMD_BRELVONAE as b on  bOBJGUID_A_SEL = aPROJECT_GUID .

  • Problem in activating datasource after replication from source system

    Hi Experts,
    After replication of data sources from source system ,I get all data sources inactive. I tried to make it active manually but every time I get
    Syntax error in GP_ERR_RSAPTD1, row 27 (-> long text)
        Message no. RG102
    Diagnosis
        Field "/BIC/CCZVBAK2031" is unknown. It is neither in one of the
        specified tables nor defi...
    How do I go about solving this problem....Any advise please...

    Hello,
    You do not mention your source system release but can you execute the check report
    described in the SAP note 493422 , if this does not resolve the problem can you goto
    RSA1, right click on the relevant "source system" and choose the "restore" option,
    then please try and replicate and activate the datasource again.
    If you  have BW release 7.0 please check that the note 929751is applied in your system
    or that you are on a higher support package level.
    If there is still a problem please le me know your BW Release and Support package level.
    Best Regards,
    Des

  • Psa data is deviated from source system data...

    Hi all,
    My source data was manipulated when it reaches to PSA. We r extracting the data from Flatfiles. And v have one keyfigure as no.of months (say 12,13,14). When am extracting this data to my Infocube this particular fields has changed as 12,000 and 13,000 and 14,000. Even though it is appearing in Simulation under Schduler tab in the same way(12,000 and 13,000 and 14,000).Am not multiplying this field with any value under updaterules.
    And the datatype for keyfigure is Number.
    And PSA data should always similar to source system data.
    But here also am getting deviations. Let me inform me what are the routines and conversion takes place, before the data reaching to PSA from source system.

    Hi Suresh,
    Don't worry. The field presentation in the cube depends on the field type.
    And I guess that comma stands for a decimal delimiter.
    And if so, there is no difference at all!
    Check your settings.
    Best regards,
    Eugene
    Message was edited by: Eugene Khusainov

  • Error in Process Chain while extracting data from source system

    Hi All,
    Dail we are facing problem while retrieving the data from source system for the data source 0TB_AMOUNT1 in the process chain, it is giving error "Function module BANK_TMC_API_SIM_GET does not exist Furnction module". when I repeat the step than it is going fine.
    Please let me know why this error is coming for first why it is not coming after repeating the step?
    Thanks & Regards,
    Murali.

    Hi Murali,
    Hope you are doing good..
    Is it a full or Delta load on a daily basis? Did you check if the FM exists in your system?
    Did you try debugging the load in source system and check if the extractor is trying to call FM you mention.
    If possible please post the exact error message you are receiving so that we could get an idea whats the error and why you are encountering it.
    As of now what i can suggest is, go to RSA2 in source system, give the datasource name and check if the extractor code has FM in it.
    Regards,
    Anil Pragada.

  • To retain the same file name in target(File) that coming from source system(AS2).

    Hi All,
    I have a requirement to send the two file from the third party system(GXS) to SAP ECC via SAP PI(version - 7.4). We are using AS2 adapter at the sender side to pick the file and using File Adapter to place the file in ECC.
    Its a pass through scenarion(No ESR part is used).
    Requirement:
    We are going to get 2 files ->Example:  one from Axis and another one from ICICI -> each bank file will have a file name ->
    We have to post the file at the target system with the same name as source file name, the source file name could come dynamically(with diff time stamp).
    Please let me know, Is it possible to achieve the above requirement(we have to retain the file name in target that coming from source system) with Sender AS2 adapter, If yes then please suggest the best approach.
    Regards,
    Pradeep Bawlia

    Hi Guys,
    The above requirement is solved now.
    Solution - I used java mapping for extracting the same source file name and assignning the same to the target system.
    Code:
    public class ExtractFileName_ASMA extends AbstractTransformation{
    public void transform(TransformationInput in, TransformationOutput out) throws StreamTransformationException  
    DynamicConfiguration conf=in.getDynamicConfiguration();
    DynamicConfigurationKey key_AS2FileNmae = DynamicConfigurationKey.create("http://sap.com/xi/XI/AS2/AS2", "AS2Filename");
    DynamicConfigurationKey key_FileNmae = DynamicConfigurationKey.create("http:/"+"/sap.com/xi/XI/System/File", "FileName");
    String As2FileName=conf.get(key_AS2FileNmae);
    conf.put(key_FileNmae, As2FileName);
    try{
    byte b[]=new byte[in.getInputPayload().getInputStream().available()];
    in.getInputPayload().getInputStream().read(b);
    out.getOutputPayload().getOutputStream().write(b);
    catch(Exception e)
    2. Another solution we can use the graphical mapping and need to use the dynamic config UDF at mapping level to acheive this requirement but here is one drawback - we need to do content conversion is this case.
    In case of Java mapping, no need to play with source file.
    WITHOUT USING ESR, WE CAN NOT ACHEIVE THIS REQUIREMENT.
    Regards,
    Pradeep Bawlia

  • Error in loading data into PSA from source system.

    Hi Experts !!!
    Good morning .
    I am trying to load data from SRM source system into BI system.When I execute infopackage,the data is not loaded into PSA and status of the process immediately becomes yellow.After sometime it gets timed out and converts to RED.When I checked the error documentation I found that the iDOCS response was not received from the source system.Detailed error is as below.
    System Response
    There are idocs in the source system ALE outbox that did not arrive in the ALE inbox of BI.
    Further Analysis
    Check the TRFC log
    you can access this log using the wizard OR menu path "Environment > Transact. RFC> In the source system
    However,I was not able to navigate to this path.I think if I can navigate to this path,I can manually push IDOCS from source system to BI system.
    Regards,
    Mandar.

    Hi,
    Check the DataSource in RSA3, if it is working fine and able to see the data in RSA3, there is no problem in DS level, then check the Mappings and any routines in BW for that DS, if this is also fine then check the below options.
    See Dumps in ST22, SM21 also.
    Check RFC Connection between ECC and BW systems, i.e. RSA1-->Source System->Right Click on Source system and Check.
    You must have the following profiles to BWREMOTE or ALEREMOTE users.So add it. Bcoz either of these two users will use in background to get extract the data from ECC, so add these profiels in BW.
    S_BI-WHM_RFC, S_BI-WHM_SPC, S_BI-WX_RFC
    And also check the following things.
    1.Connections from BW to ECC and ECC to BW in SM59
    2.Check Port,Partner Profiles,and Message Types in WE20 in ECC & BW.
    3.Check Dumps in ST22, and SM21.
    4.If Idocs are stuck i.e see the OLTP Idoc numbers in RSMO Screen in (BW) detials tab see in bottom, you can see OLTP Idoc number and take the Idoc numbers and then goto to ECC see the status in WE05 or WE02, if error then check the log else goto to BD87 in ECC and give the Idoc numbers and execute manually and see in RSMO and refresh.
    5.Check the LUWs struck in SM58,User Name = * (star) and run it and see Strucked LUWs and select our LUW and execute manually and see in RSMO in BW.
    See in SDN
    Re: Loading error in the production  system
    IDOC getting struck at TRFC (SM58)
    1.Reorg of table (ARFCSSTATE )
    2.Incraesing the resources in the system (no of processes and memory ) , as this issue happens owing to this.
    Source system tries to send TRFC to target and if there are no WP's available it will come to Transaction Recorded state, and form here it will not try to send this TRFC.So you have to execute this manually.
    Also we can increase the timeout parameter so that it can try few more times to send before actually it comes to recorded state.
    Regards,
    Suman

  • Mapping new product code from source system to old product code in Oracle COA

    Mapping new product code from source system to old product code in Oracle COA
    Here is the situation:
    One of our legal entities is changing their product codes in their COA segment. We are not changing the code structure in the global SOB. Is there a workaround to pull the source system information into ADI (or any other tool) and map it to the current Oracle product code?
    Thanks and regards,
    [email protected]

    Only The  Data of Your Code will Change
    then if you are trying to copy a standarad sap-code , you need to copy
    every include, FM
    there is a Copy option over there in program
    you can do tat
    thnkx bhanu

  • Doesn't load from source system to PSA and DSO

    Hi Gurus,
    We have launched a Delta load from Source System (POSDM) to DSO. In the source system there are 5000000 rows but none was loaded into DSO or PSA. The delta load was ok, but any data was loaded.
    If the delta  load was lunch again, would be loaded those 5000000 rows. Or those 5000000 rows wouldn't be able to load again.
    Any idea abut this issue? Any feedback will be really appreciated.
    Thanks in advance.

    Hi David,
    Are you sure this 5 million records are delta records and should be pulled into BW as delta??
    did you count the number of records in the underlying tables??
    which data source you are using??
    Delta loads are suppose to bring the new records or changed records since the last data load and not all the records.
    Since the request is green as you said and still it shows 0 records then it is possible that nothing has changed since last delta.try to see the details of the data load... if it failed...may be thats the reason you are not able to see any records.
    If you schedule the delta again it will bring the records changed or created after last delta.
    If the delta was unsuccessfull then turn the QM status and overall status of request to red manually in the monitor of the request and delete it from PSA and all the targets in which its loaded and schedule the delta again and it should bring the delta.
    Thanks
    Ajeet
    Thanks
    Ajeet

  • How to  upload Transaction data from source system to BI 7.0

    Dear friends,
    I want to know how to  upload Transaction data from source system to BI 7.0.
    if anybody having step by step material for this process ,please send it to my
    mail id.
    [email protected]
    [email protected]
    Thanks,
    D.prabhu

    Hi,
    it works in the same way as before in case you are not using the new features like the dtp and the transformations. Using the new features, you schedule a infopackage to load the data up to psa and schedule a dtp to post the data to the targets.
    regards
    Siggi
    PS: We want to share knowledge in sdn and not keeping it private by having the information sent to our email adresses.

  • Select the language field in the source system ?

    Hi,
    when I schedule the load for 0cust_sales_text there popups the message 'Select the language field in the source system -> Long text'.
    Long text says: 'Either no language field is selected as a selection field in the source system KC1CL005 in DataSource 0CUST_SALES_TEXT, or the language field is not assigned to InfoObject 0LANGU in BW.'
    In the DataSource there is not an object for language? I don't know what to do.
    Tobias

    Hello,
    texts of 0cust_sales are not marked as language-dependent.
    the psa tells me, which records are marked as error, but i don't understand it.
    green     1     1     0009     03     MB     L605930     cust a
    green     1     2     0009     03     XX     L605930     cust a
    red     1     3     0009     04     MB     0000013081     cust b
    red     1     4     0009     04     XX     0000013081     cust b
    green     1     5     0020     01     MB     0000001570     cust c
    green     1     6     0020     01     XX     0000001570     cust c
    red     1     7     0020     01     MB     0000001571     cust c
    red     1     8     0020     01     XX     0000001571     cust c
    red     1     9     0020     01     MB     0000001572     cust c
    red     1     10     0020     01     XX     0000001572     cust c
    green     1     11     0020     01     MB     0000001578     cust d
    The error messages say there were double records
    Tobias

  • Can the Adobe PDF plugin carry exact phrase search criteria through from source system into Reader?

    In SharePoint 2013 if you search for an exact phrase and then view a PDF that is returned by the search (in Reader via the PDF plugin) then the Reader will treat the complete phrase as individual words.
    For example, search for "High Court" in SharePoint, get returned a set of PDF documents, and when you view one of them through the Reader the hit highlighting will be for each instance of "High" or "Court".
    What is wanted is hit highlighting only of instances of "High Court".
    I get the same behavior if the Adobe PDF plugin is installed with the Filesite DMS. So it seems to be standard behavior for the Adobe PDF plugin
    Is there a way to make the Adobe PDF plugin carry exact phrase search criteria through from source system into Reader? And then hit highlight the exact phrase only.

    Hi Sam,
    You can opt for 'Advance Search' option in Reader to match the whole words to get to the specific documents.
    Regards,
    Rave

  • Error Occures while loading data from Source system to Target ODS

    Hi..
    I started loading Records From source system to target ODS.while i running the job i got the following errors.
    Record 18211 :ERROR IN HOLIDAY_GET 20011114 00000000
    Record 18212 :ERROR IN HOLIDAY_GET 20011114 00000000
    sp Please help me in these following Errors..
    Thanks in advance,

    Hello
    How r u ?
    I think this problem is at the ODS level, ZCAM_O04 is ur ODS Name.
    Could u check the ODS Settings, and the Unique Data Records is Checked or Not ?
    Best Regards....
    Sankar Kumar
    +91 98403 47141

  • Slow extraction from Source System

    Experts,
    I'm on the basis team and we're trying to figure out what we can do to increase the performance of a BI extraction from an ECC source system.
    Our BI team is doing a Full extraction of Data Source 0UC_SALES_STATS_02 which has about 24 million records.  This has been running for about 2.5 days and has only extracted 12.5 million records according to RSMO
    I'm no expert on BI performance, but I'm trying to get there.  One thing we noticed is that we have over 1000 Data Packages and each has about 11K records per package.
    RSBIW "control params for data transfer" in the BI system is set as this:
    LOGSYS     --  Max KB.  --  MAX Lines  --  Freq  --  MAx PROC  -- Target Sys  -- Max DPs
    OURBWSYS --  20000 --          0             --  10    --        3            --     blank        --   blank
    OURECCSYS --  20000 --          0             --  10    --        3            --     blank        --   blank
    Also we only see one background process running on the ECC system which is the Zprogram that was written to perform the extract.
    We also checked IMG - sap netweaver - BI - links to other source systems - maintain contrl params for data xfer
    Freq/IDOC = 10
    pkg size = 50,000
    partition size = 1,000,000
    We are on NWEHP1 for our BI system
    We are on ECC6.0 EHP3sp5 for our ECC system
    We use Oracle 10.2.0.4.0 with it fully tuned
    We also think we have all the correct tuning parameters in our ECC/BI systems.  There are a few tweaks we could make but I don't see any glaring problems.  All our memory parameters seem to be in line with what is recommended.
    We do not see any kind of Memory/CPU bottlenecks.
    Do you think the many data packages and small records per package is a problem? 
    Any tips to increase performance would be helpful, I'm still looking.
    I saw this wiki thread and it mentions to look at tcode RSA3 in ECC system to monitor long extracts.   But I have looked at this and I have not figured out what all needs to go into the fields to correctly run this.
    Any help you can provide would be great.
    Thanks,
    NICK

    HI Nick,
    This problem is due to the huge volumes of data and the job is ruinng from long time onwards, normally we will kill the job if it exceeds more than 8 hours time while extracting the data.
    Please suggest your BI Team to pull data in small volumes by providing selections in info package. This will drastically redcues the time for extraction of huge data and will also not consume all the resources in system.
    Please pass below links to your BI team as this is completely releated to design issue.
    http://help.sap.com/saphelp_nw70/helpdata/en/39/b520d2ba024cfcb542ef455c76b250/frameset.htm
    Type of DSO should be write optimized to increase load performance.
    http://help.sap.com/saphelp_nw70/helpdata/en/34/1c74415cb1137de10000000a155106/frameset.htm
    2. And at your end look for the table spaces of the PSA tables, may be it is execeeding its capacity and make it as automatic. instead of manullay assigning the space at the regular intervals.
    Hope this helps.
    Regards,
    Reddy
    Edited by: Reddybl on Apr 2, 2010 12:23 AM

Maybe you are looking for