Initial load of custom table from r3

Hello all,
I am pretty much new to crm, with the documentation i had i was able to create a site for my r3 system in crm.
Initially i created a z table in my r3 system and crm system. i loaded some data into r3 ztable. i just want to replicate the data into my crm system.
If some one tell me if i need to create any custom function module to create adapter object in crm system. and also necessary steps from here on to perform the required task.
Any doucment regardig could be sent to: [email protected]
Thanks,
Krishnam Raju.

Hi,
  You need to perform a number of steps to result in the intial load of a z-R/3 table to a Z-CRM table. Here are some hints although I do not have a detailed document for this. May be you should try looking the CR 550 SAP course for <i>'Enhancing CRM middleware'</i> -
1. Create a Bdoc structure and model a mBdoc using the Bdoc Modeller in CRM.
2. Create a Adapter object in trx.
R3AC1
giving all details of the Bdoc type and tables used in source and destination systems.
3.  You need to create an extract function module for extracting the data from Z-r/3 table. This needs to be part of crmsubtab mw table.
4. You also need to create a mapping function module in CRM for mapping the data received from the R/3 plug-in to the mBdoc structure. Insert this in the appropriate tab of trx.
R3AC1
5. Create a validation module which validates the data before persisting into the CRM z-table. This should be part of MW table smw3bdocif.
6. Once everything is in place, you can start the load using trx.
R3AS
. Monitor using
SMW01
7. Use the trx.
SMQR
to deregister the inbound queues and debug the CRM inbound. You can debug the R/3 outbound using
SMQS
Hope this helps. Reward if helpful.
Regards,
Sudipta.

Similar Messages

  • Error in delta load of 'Customer Group' from R/3 to CRM

    Hi Experts,
    I added new 'Customer Groups' in R/3. But this is not being updated in CRM. Do I have to trigger initial load again through R3AS or should I create new Customer Groups in CRM?
    The BP delta load BDocs are stuck in SMW01 due to this change. Also, if I need to do the Initial load of Customer Group, which object do I need to do the load for? Is it dnl_cust_sales?
    Please help as it is a Production issue.
    -- Pat

    Hi Pat,
    Use object 'DNL_CUST_SALES' to download Customer Group from R/3 to CRM.
    Use R3AS4 transaction to execute the same.
    Best Regards,
    Pratik Patel
    <b>Reward with Points!</b>

  • How to MODIFY A CUSTOM TABLE  FROM A FLAT FILE

    Dear Friends,
                     I have a requirement where i have to upload data from excel file to my custom table  so i have used a FM
    'TEXT_CONVERT_XLS_TO_SAP' and i have collected data into a internal table , till here i am able to get data correctly , now i hae to upload this data into a custom table .
    the flat file is having  6 fields and the custom table is having
    8 fields , for uploading the data into this custom table from the internal table where i have collected above iam getting problem . Actually iam using a modify statement to update the custom table .
    the flat file which i have collected into the internal table is as below :
      IDNo.     Name     Date      Location   Designation  Dept
      101       Raja      4/12/2007  Delhi      Manager      HR
      102       James    4/12/2007  Delhi      Clerk          HR
    Custom table  is having the below fields
    IDNO.  Name  Date  Location Designation  Dept   Manager
    101                                                                   Raja
    now when i run the program iam getting the problem   while usign the modify statment is the ID no which is already having
    a record as  IDno = 101  and manger = Raja.......with the other fields   name , date, location,designation and dept as blank.
    if i want to fill this fields from my flat file the modify statment
    just filling all the fields  for the ID no = 101  and manager field which already having Raja as being overwritten by space .....
    becasue this field is not being there in the flat file.
    the code iam using as follows.
    The flat file is having with the below structure
    TYPES: BEGIN OF t_emp_data,
              IDNO(11) TYPE c,
             Name(13) TYPE c,
             Date(20) TYPE c,
              Location (40) TYPE c,
             Designation(40) TYPE c,
             Dept(40) TYPE c,
             end of t_emp_data.
    The Custom Table(ZEMP_DATA) is having with the below structure
    TYPES: BEGIN OF t_emp_data_table,
              IDNO(11) TYPE c,
             Name(13) TYPE c,
             Date(20) TYPE c,
              Location (40) TYPE c,
             Designation(40) TYPE c,
             Dept(40) TYPE c,
             Manager(20) type c,  -- this is the extra field in table
             end of t_emp_data_table.
    data :
    it_empdata TYPE STANDARD TABLE OF t_emp_data,
    it_empdata_tmp TYPE STANDARD TABLE OF t_empdata_tmp,
    wa_empdata_tmp  type t_empdata_tmp,
    wa_empdata type t_emp_data.
    loop at it_empdata into  wa_empdata.
      move-corresponding  wa_empdata to   wa_empdata_tmp.
      modify ZEMP_DATA  from  wa_empdata_tmp .
    endloop.
    could any one please let me know what i have to do inorder to not get the manager field data not being overwritten with the modify statment , for the IDNo. 101  . I want the data which is already ( manager = Raja) shouldnt not be get overwritten with Space.
    please help me in this regard
    Regards
    Madhuri.

    Hi,
    use a slect statement before
    "move-corresponding wa_empdata to wa_empdata_tmp."
    select manager
    from ztable
    into wa_empdata_tmp-manager
    where id = 100.
    regards,
    lavanya

  • Need help- need to read a customer table from the payroll schema

    Hi gurus,
    I need to read an amonunt from a customer table in the payroll schema. Is it possible to input TABLE XXXX in the schema, or is impossible to read from a customer table?Do I need to create a report to upload the amount in an infotype?
    Thanks in advanced for for support!

    You should create your own custom operation (based on operation TABLE) to read your custom table from a Payroll PCR.
    If your custom table's name is longer than 5 characters, you may be required to use operations VAOFF and/or VALEN prior to your custom operation in the PCR.

  • Loading  a database table from a UNIX file

    hi,
    Need A program to load a database table from a UNIX file
    thnks.

    HI,
    report zmjud001 no standard page heading.
    tables: z_mver.
    parameters: test(60) lower case default '/dir/judit.txt'.
    data: begin of unix_intab occurs 100,
    field(53),
    end of unix_intab.
    data: msg(60).
    ***open the unix file
    open dataset test for input in text mode message msg.
    if sy-subrc <> 0.
    write: / msg.
    exit.
    endif.
    ***load the unix file into an internal table
    do.
    read dataset test into unix_intab.
    if sy-subrc ne 0.
    exit.
    else.
    append unix_intab.
    endif.
    enddo.
    close dataset test.
    ***to process the data. load the database table
    loop at unix_intab.
    z_mver-mandt = sy-mandt.
    z_mver-matnr = unix_intab-field(10).
    translate z_mver-matnr to upper case.
    z_mver-werks = unix_intab-field+10(4).
    translate z_mver-werks to upper case.
    z_mver-gjahr = sy-datum(4).
    z_mver-perkz = 'M'.
    z_mver-mgv01 = unix_intab-field+14(13).
    z_mver-mgv02 = unix_intab-field+27(13).
    z_mver-mgv03 = unix_intab-field+40(13).
    to check the data on the screen (this is just for checking purpose)
    write: / z_mver-mandt, z_mver-matnr, z_mver-werks, z_mver-gjahr,
    z_mver-perkz, z_mver-mgv01,
    z_mver-mgv02, z_mver-mgv03.
    insert z_mver client specified.
    *if the data already had been in table z_mver then sy-subrc will not be
    *equal with zero. (this can be *interesting for you - (this list is
    *not necessary but it maybe useful for you)
    if sy-subrc ne 0.
    write:/ z_mver-matnr, z_mver-werks.
    endif.
    endloop.
    1. This solution is recommended only if the database table is NOT a standard SAP database table .
    Cheers,
    Chandra Sekhar.

  • Initial Load of DD* tables

    Hi everyone,
    Just a hopefully quick and simple question. We set up a SLT Replication Server recently and created our first SLT Configuration into HANA. We have no need for real time so we choose a scheduled time out of office hours. The configuration was successful however the DD* tables appeared in a scheduled state. From what I thought these tables should populate (load) initially regardless of the schedule. What appeared to happen was that they were waiting for the first schedule to run. Is this expected? Without these populating initially we could not choose any real ERP tables to replicate.
    We also tried with a non-SAP source (Oracle) and the DD* tables for that configuration were populated instantly even though that config was scheduled to run "off-peak" as well.
    Thanks,
    Marcus.

    Hi Marcus,
    As far as I understand your question, please find below my comments-
    (sap source system scenario)
    If the configuration is created with "schedule by time" option which I think is done in your case, then the  replication server would replicate database changes to target system at the time set by you. Here the metadata of tables would be copied from source system tables[DD002L and DD02T]
    yes, you are correct that ideally we should start with ERP tables replication after DD* tables are replicated successfully. This generally is faster but then depends upon the system.
    (non sap source system scenario)
    here the DD* tables are just initially loaded and not automatically replicated.
    So you would find a difference in how the replication takes place in both scenarios.
    Hope it answers to some extent for your query.
    Regards,
    Saritha K

  • Initial load of inventory level from csv - double datarows in query

    Hello everybody,
    a query result shown in a web browser seems strange to me and I would be very glad, if anyone can give me some advice how to solve the problem. As I do not think that it is related to the query, I posted it into this forum.
    The query refers to an InfoCube for inventory management with a single non-cumulative key figure and two other cumulative key figures for increase and decrease of inventory. The time reference characteristic is 0CALDAY. The initial load has been processed reading from a flat file (CSV), the structure looks like this:
    Product group     XXX
    Day               20040101
    Quantity          1000
    Increase          0
    Decrease          0
    Unit               ST
    The initial load runs fine, the system fills all the record sets into the InfoCube. Unfortunately I do not know how to look at the records written into the cube, because only the cumulative key figures are shown in InfoCube-> Manage-> Contents.
    Well, when executing the query, a really simple one, the result is just strange, because somehow there are now two rows for each product group with different dates, one with the 1st of January, 2004 and the other for the 31st of December, 2003 containing both 1000 units. The sum is 2000.
    It became more confusing, when I loaded the data for increase and decrease: now the quantities and sums      are correct, but the date of the initial load is a few days later than before, the data table in the query does not contain the 1st of January.
    Does anybody know, what I did wrong or where there is information about how to perform an initial load of inventory from csv in a better way?
    Kind regards
    Peter

    Peter,
    Inventory is not that straight forward to evaluate as it is non-cumulative. Basically it means that one KF is derived from one/two other KFs. You cannot see non-cumulative KFs in manage infocube.
    Have you uploaded opening balances separately? If so, your data for 31st of december is explained.
    In non-cumulative cubes, there need not be a posting for a particular day for a record to exist. For e.g. if you have stock as 10 units on 1st and then no posting for 2nd and 3rd and then increase 10 units on 4th, even for 2nd and 3rd, the non-cumulative KF will report as 10 units (stock on 1st rolled forward).
    There is a how to...inventory management document on service market place that explains this quite nicely.
    Cheers
    Aneesh

  • Delta load of customizing objects from R3 to CRM

    I need to set up a delta load for customizing objects for ex: DNL_CUST_BNKA from R3 to CRM.  Is there any documentation on this subject?  Please advise.  Thanks.

    This is what I got from SAP
    Bank Details information is a customizing object, and is not designed
    for frequent changes, because there's no delta download for customizing
    objects. If you create new Bank/change detail or delete an old one,
    you have to perform initial load manually for DNL_CUST_BNKA again.
    You can also do a request download of the replication object
    'DNL_CUST_BNKA'to get some bank data added later on.
    (Transaction \R3AR2 ).
    You will need to do a request download of bank data from time to time
    whenever you change the bank details. This is currently how it can be
    done in SAP Standard.
    If you change regularly bank master data:
    The data has to be in sync between R/3 and CRM. If you make changes
    often to the bank master data, you might consider scheduling a regular
    job. See below for steps.
    1. Transaction R3AS
    2. Load Object = DNL_CUST_BNKA
    Source Site = your R/3 system
    Destination Site = your CRM system
    3. Menu Goto > Variants > Save Variant
    4. Enter a Variant Name and Description
    5. Start Transaction SM36 to define a background job
    6. Select the Job Wizard button:
    In step Defining a Job Step, select ABAP program step
    In step Define an ABAP Program Job Step, enter ABAP program name
    SMOF_DOWNLOAD and the variant you created in step 4
    7. Click on Job Wizard and follow the steps, choose ABAP Program
    SMOF_DOWNLOAD and the variant you created in step 4
    8. Select start time Immediately
    9. Press Complete to finish
    10.Choose Job Selection (SM37)
    11.Enter the Job Name -> should have status finished
    Schedule the job (SM36 > start condition) according to how often you
    change your bank

  • Initial load of sales orders from R3 to CRM without statuses

    1) Some sales orders were uploaded into CRM without statuses in the headers or line items. 
    2) Some sales orders were uploaded without status, ship-to, sold-to, payer.....If I deleted them and use R3AR2, R3AR4 to upload each individual then no problem.
    Any ideas or suggestions?
    Thanks.

    Hi,
       Request load of adapter objects uses different extractor modules for extracting the data from external system to CRM. While your initial load of sales docs. will use a different extraction logic bases on the filter conditions specfied on trx.
    R3AC1
       There may be a problem in the extraction of data from the source system (don't know if you are using a R/3). Can you please de-register the R/3 (i suppose) outbound queue using trx.
    SMQS
    , and then debug the extraction (R/3 outbound) before the data is sent to CRM using FM
    CRS_SEND_TO_SERVER
       If this goes well, you may try debugging the mapper in CRM inbound and the validation module in CRM as a last resort. Also, please refer to trx.
    SMW01
    to see if the Bdocs are fully processed.
    Hope this helps...Reward if helpful.
    Regards,
    Sudipta.

  • Insert delete or update entry in Custom table from KONV entry changed

    Hi All,
    I have custom table ZKONV with only few required columns and should have same number of records as KONV has at any point in time.
    KONV is a cluster table so its not readable from ORACLE level. So ZKONV is created. But I dont know how to keep these both tables in sync.
    I need to perform insert delete or update entry in Custom table if insert delete or update happens on cluster table KONV from any transactions.
    As KONV is a cluster table and does not have changed time stamp I am not able to know the number of records changed in perticulat time period.
    Thanks,

    Thanks for reply,
    There is a Outside SAP system which needs to read KONV data to feed into their system, but as KONV is cluster table they are not able to read it from ORACLE level.
    To solve this we are thinking to create a transparent Z-table and will fill it with KONV and catch Update, delete or Insert statement and do same on ZKONv.
    Is this possible some how? by some database event or something....

  • Update custom table from qa_Results table

    Dear Experts,
    The question is related with Oracle quality collection plan. Here is the question.
    After entering the result in collection plan we all know that result is store in qa_Results_V view and also in respective view for the collection plan.
    In the same way , the same result I want to store in one custom table. Is it possible?
    Thanks,
    Atul Ramteke

    <u>Syntax:</u>
    EXPORT param_1 TO MEMORY ID 'EXIT_PCCD'.
    IMPORT param_1 FROM MEMORY ID 'EXIT_PCCD'.
    FREE MEMORY ID 'EXIT_PCCD'.
    Please look into the F1-help on Export/Import statements for more details.
    Message was edited by: Sam

  • How can I update a customized table from a text document?

    Hi,
    I am one sample file. The file is an text document. From this .txt file how can I update a customized table in SAP.
    How can I Update the fields in the Customized table.
    Thanks & Regards,
    NManohar.

    use the Function module GUI_UPLOAD or WS_UPLOAD and give the path and the file name from where the data has to be uploaded. in table parameter give the iternal table
    CALL FUNCTION 'GUI_UPLOAD'
      EXPORTING
        filename                      =
      FILETYPE                      = 'ASC'
      HAS_FIELD_SEPARATOR           = ' '
      HEADER_LENGTH                 = 0
      READ_BY_LINE                  = 'X'
      DAT_MODE                      = ' '
      CODEPAGE                      = ' '
      IGNORE_CERR                   = ABAP_TRUE
      REPLACEMENT                   = '#'
      CHECK_BOM                     = ' '
      VIRUS_SCAN_PROFILE            =
      NO_AUTH_CHECK                 = ' '
    IMPORTING
      FILELENGTH                    =
      HEADER                        =
      tables
        data_tab                      =
    EXCEPTIONS
      FILE_OPEN_ERROR               = 1
      FILE_READ_ERROR               = 2
      NO_BATCH                      = 3
      GUI_REFUSE_FILETRANSFER       = 4
      INVALID_TYPE                  = 5
      NO_AUTHORITY                  = 6
      UNKNOWN_ERROR                 = 7
      BAD_DATA_FORMAT               = 8
      HEADER_NOT_ALLOWED            = 9
      SEPARATOR_NOT_ALLOWED         = 10
      HEADER_TOO_LONG               = 11
      UNKNOWN_DP_ERROR              = 12
      ACCESS_DENIED                 = 13
      DP_OUT_OF_MEMORY              = 14
      DISK_FULL                     = 15
      DP_TIMEOUT                    = 16
      OTHERS                        = 17
    IF sy-subrc <> 0.
    MESSAGE ID SY-MSGID TYPE SY-MSGTY NUMBER SY-MSGNO
            WITH SY-MSGV1 SY-MSGV2 SY-MSGV3 SY-MSGV4.
    ENDIF.
    and finally use insert command to update the database table with the data

  • Initial Load for Master Data from R/3 to CRM

    Hi Middleware Experts,
        I have an ambiguity on initial load of enhanced master data from R/3 to CRM. The current situation is
    1)I have already enhanced master data on R/3 side with some additional fields. Now shall i go ahead with standard intial load of these Master data to CRM without creating middleware interfaces to map those enhanced fields on CRM side?
    2)Also after doing initial load from R/3 to CRM, if I develop middleware interfaces for mapping these new fields of master data and then again do the initial load, so that new fields also get mapped, will the initial load be done properly as i have already done it once? will it overwrite the previous initial load data on CRM side properly?
    Please suggest whether it can be done or not and which is the better option.
    Regards,
    Rahul

    Hi Rahul,
    Since you have not done the mapping yet the middleware will not bring those fields to CRM it will work just like a standard initial load without considering the enhanced master data field.
    When you develop the middleware interface and enhance the CRM data structure also and then you can see the exchange of those fields through MW. You can start the intial load at that time also its not that you can do the initial load just once. But the better option would be that you avoid doing initial load several times because its time consuming activity. You do all the enhancement and then finally perform the initaial load.
    <b>Allot points if my post helps!!</b>
    Best regards,
    Vikash.

  • Synonym required to query custom tables from Oracle Apps database

    When creating an Answers report that queries from a table owned by a custom schema (not APPS), I cannot query the table and get error. Once I add a Synonym for APPS on this table, then OBIEE works normally.
    Since OBIEE is logging in as the APPS user, are Synonyms required for all non-APPS tables/views or is there another feature of the Administrator where it can store the table alias (our typical owner of custom tables is XXFI)?

    OK, well "compare" can mean one of two things: Compare structure or compare contents. Here is a quick script to compare column structures of two tables on one database:
    (select COLUMN_NAME,
    DATA_TYPE,
    DATA_LENGTH,
    DATA_PRECISION,
    DATA_SCALE,
    NULLABLE
    from dba_Tab_columns
    where owner=:OWNR1
    and table_name = :tablename
    minus
    select COLUMN_NAME,
    DATA_TYPE,
    DATA_LENGTH,
    DATA_PRECISION,
    DATA_SCALE,
    NULLABLE
    from dba_Tab_columns
    where owner=:ownr2
    and table_name = :tablename)
    union all
    (select COLUMN_NAME,
    DATA_TYPE,
    DATA_LENGTH,
    DATA_PRECISION,
    DATA_SCALE,
    NULLABLE
    from dba_Tab_columns
    where owner=:ownr2
    and table_name = :tablename
    minus
    select COLUMN_NAME,
    DATA_TYPE,
    DATA_LENGTH,
    DATA_PRECISION,
    DATA_SCALE,
    NULLABLE
    from dba_Tab_columns
    where owner=:ownr1
    and table_name = :tablename)
    If this query returns any rows, then these indicate that there are structural differences between the tables. We do a minus in both directions to ensure that an additional column in either schema will be returned in the query.
    If you are going across dblinks to remote tables then you have to amend the "dba_tab_columns" to "sys.dba_tab_columns@yourdblink"
    If you also want to compare indexes, triggers, etc then do the same sort of thing for the associated dba_ views for those objects.
    And if you want to compare table contents, then often the fastest way is also to check minuses in both directions, if the tables are not too big:
    e.g.
    (select * from schema1.table@dblink1
    minus
    select * from schema2.table@dblink2)
    union all
    (select * from schema2.table@dblink2
    minus
    select * from schema1.table@dblink1)
    local tables, of course, simply omit the "@dblink" issues.
    There are tools to help in such things. TOAD, for example, has a pretty good schema comparison tool, and there are plenty of other options out there. But if you need to script this yourself then the logic I've shown is a good starting point.
    Cheers,
    Mike

  • Replicate Custom Table from CRM to ECC

    I have a custom table in CRM and I have to transfer the values of the custom table to ECC.I have BADI  which inserts the values into the custom table in CRM. What should i do to replicate the values in a Custom table in ECC.The tables are of same structure.

    Hi
    Please reffer the C71 : CRM Connectivity document in Best Practices. It will gives you the perfect information for you.
    Please follow the following link to find the same.
    http://help.sap.com/bp_crmv250/CRM_DE/index.htm
    Technical Information>Building Block Library> C71 : CRM Connectivity>Configuration guide.
    Generally: You have to enhance the BAPIMTCS structure which is exchanged between CRM and R/3, and the functionality mapping the BDoc to this structure in CRM and in R/3 the functionality of saving the BAPIMTCS structure to the DB.
    You should use the Transaction Business Events, see also tables TBE24, TPS34 and you can search for the SMOUTIL2 or SMOUTIL3 functionality esp in the CRM system.
    Have a look at note 417906 with some information on used functions.
    Reward with points if helpful
    Cheers
    Manohar

Maybe you are looking for

  • Itunes video playback started  crashing and no longer able to view on apple tv ideas?

    hi since yesterday after everything working ok all videos played fine Itunes has started crashing on my pc ( windows 7 ) when playing some videos I have deleted and re installed itunes no help and apple tv no longer able to share videos from itunes m

  • Apple releasing iPod Touch in "october" when can will it be available in stores.

    I want to buy the new iPod Touch as soon as it comes out, but I can't find where it says anything about store release date in stores. does anybody know if when it says october it will be available octber the 1st?

  • On-line Number in Canada

    Hi I am in Victoria BC Canada and would like to get a on-line number when I click on Canada I get a message that says un avaiable is this just temporay ? Thanks Coachrobwille

  • Workgroup Client.InDesignPlugIn Conflicts

    Hi, Yesterday I updated In Design and after that it is not launching. It gives message "Workgroup Client.InDesignPlugIn Conflicts Workgroup Client Ul.InDesignPlugIn" 1) I have removed from PlugIn folder and check did not solve the problem. 2) I selec

  • Help with looped equation

    I am really struggling here with this (what seems) fairly simple labview problem.  I have a set of equations that rely on the answer of the previous other formula and so forth.  I can do this easily enough in excel.  I tried to recreate this function