R/3 tables for a data source

HI all,
How to find the R/3 tables for a perticular data sources..
like 2lis_13_vditm etc...
Thanks,
Pooja

Hi Pooja,
U can get the same from LBWE t-Code.. where for ur Extractor.. look at the Maintain Extractor Structure.
If u click that u wil get the Extract Structure.. where u can see the Communication Filed names which starts with MC<Table Name> - <Field Name>...... for every field in ur data source u can see the corresponding the table entry.
Or u can use the below link for more options:
How to find tables related to data sources
Thanks
Hope this helps

Similar Messages

  • Error while filling set up tables for 2LIS_11  data sources

    hello All ,
    I tried to fill setup tables for 2LIS_11 data sources in OLI7BW , but the job RMCVENUA is getting cancelled. below is the Job Log ,
    Job started
    Step 001 started (program RMCVNEUA, variant &0000000000005, )
    Start of order processing
    Error determining rate: foreign curr. local curr. INR date 10.09.2008 (doc. 1)
    Job cancelled
    Is there any setting required in R3 side before filling setup tables of SD ?
    I am able to fill the setup tables for 2LIS_05 ( Notifications) successfully.
    Regards ,
    Amol Kulkarni
    Edited by: Amol Kulkarni on Dec 4, 2008 3:59 AM
    Edited by: Amol Kulkarni on Dec 4, 2008 4:00 AM

    Hi,
    one of the sales document currencies is not maintained properly in the system and therefore its failing.
    you need to correct those documents before filling the set up tables.
    Either you can chose the option fault document toleration limit to 1000...so that 1000 records wrong records can be tolerated or you have to crrect it first before filling up the set up tables.
    This will generally happen in test systems.... but very rare in production...so try to maintain currrency again in the system.
    generally you will see the sales doc number in the error for which its failing.
    May be at the time of filling some óne was working on some sales document and tried it save it.
    correct that doc and fill the set up tables again.
    Thanks
    Ajeet

  • SAP GRC 10.1 AMF No data selected when adding duplicate fields from separate tables for configurable data sources

    Hi There,
    In SAP GRC 10.0, our team had an issue where we could not add duplicate fields from separate table (see ERROR: Select Currency/UoM field for the selected analyzed fields). This was resolved by the SAP Note 1904313/ 1904314 (http://service.sap.com/sap/support/notes/1904313).
    We upgraded our system to SAP GRC 10.1 SP05 and could now add the duplicate fields from separate tables. SAP Note 1904313/ 1904314 was part of SAP GRC 10.1 SP03 so it makes sense that we, in a higher version (SP05), would be able to do this.
    The issue now is when we add the duplicate fields from different tables and run the Ad-hoc Query to test if the data source works correctly, the No Data Selected warning persists. This means that the data source provides no data for analysis, which is required to write our business rules.
    Below is an example:
    Basic data source with just one currency reference field EBAN-WAERS.
    When you run the Ad-Hoc Query you receive data.
    Basic data source with second currency reference field EKKO-WAERS.
    When you run the Ad-Hoc Query no data is found.
    Please also make reference to the following thread logged by my colleague (ERROR: Select Currency/UoM field for the selected analyzed fields)
    Any assistance to receive data with duplicate fields from separate tables will be highly appreciated.
    Thanking you in advance.
    Regards
    Gary Khan

    Hi
    following are the  error messages from dump
    hrtText
       There is already a line with the same key.
    hat happened?
       Error in ABAP application program.
       The current ABAP program "SAPLCKMS" had to be terminated because one of the
       statements could not be executed.
       This is probably due to an error in the ABAP program.
    rror analysis
       You wanted to add an entry to table "\FUNCTION-POOL=CKMS\DATA=T_DYN_CKMLCR",
        which you declared
       with a UNIQUE KEY. However, there was already an entry with the
       same key.
       This may have been in an INSERT or MOVE statement, or within a
       SELECT ... INTO statement.
       In particular, you cannot insert more than one initial line into a
       table with a unique key using the INSERT INITIAL LINE... statement.
    rigger Location of Runtime Error
       Program                                 SAPLCKMS
       Include                                 LCKMSF01
       Row                                     226
       Module type                             (FORM)
       Module Name                             DYNAMIC_PERIOD_CLOSING
    Source code where dump ocured
    222
    223           APPEND ht_ckmlpp TO t_add_ckmlpp.
    224           APPEND LINES OF ht_ckmlcr TO t_add_ckmlcr.
    225           INSERT ht_ckmlpp INTO TABLE t_dyn_ckmlpp.
    >>>>           INSERT LINES OF ht_ckmlcr INTO TABLE t_dyn_ckmlcr.
    227         ENDWHILE.
    Also I guess there is problem with material ledger in R/3 side
    I have never worked on material ledger before so dont hav idea of Tcode and tables in SAP R/3 for material ledger.
    Thanks
    Navneet

  • What are the steps need to be taken care  for MM data sources  while extrac

    what are the steps need to be taken care  for MM data sources  while extracting  from R/3 tables
    and please provide the Steps involved in R/3 side
    thnaks
    Racha

    Hi,
    For Inventory Management, you can have a look at the following link.
    [https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/f83be790-0201-0010-4fb0-98bd7c01e328]

  • Time series functions are not working in OBIEE for ESSBASE data source

    Hi All,
    I am facing a problem in OBIEE as I am getting error messages for measure columns with Time series functions(Ago,ToDate and PeriodRolling) in both RPD and Answers.
    Error is "Target database does not support Ago operation".
    But I am aware of OBIEE supports Time Series functions for Essbase data source.
    using Hyperion 9.3.1 as data source and obiee 11.1.1.5.0 as reporting tool.
    Appreciate your help.
    Thanks,
    Aravind

    Hi,
    is because the time series function are not supported for the framentation content, see the content of the oracle support:
    The error occurs due to the fact the fragmented data sources are used on some Time series measures. Time series measures (i.e. AGO) are not supported on fragmented data sources.
    Confirmation is documented in the following guide - Creating and Administering the Business Model and Mapping Layer in an Oracle BI Repository > Process of Creating and Administering Dimensions
    Ago or ToDate functionality is not supported on fragmented logical table sources. For more information, refer to “About Time Series Conversion Functions” on page 197.
    Regards,
    Gianluca

  • Web Analysis : populate the same table with multiple data sources

    Hi folks,
    I would like to know if it is possible to populate a table with multiple data sources.
    For instance, I'd like to create a table with 3 columns : Entity, Customer and AvgCostPerCust.
    Entity and Customer come from one Essbase, AvgCostPerCust comes from HFM.
    The objective is to get a calculated member which is Customer * AvgCostPerCust.
    Any ideas ?
    Once again, thanks for your help.

    I would like to have the following output:
    File 1 - Store 2 - Query A + Store 2 - Query B
    File 2 - Store 4 - Query A + Store 4 - Query B
    File 3 - Store 5 - Query A + Store 5 - Query B
    the bursting level should be give at
    File 1 - Store 2 - Query A + Store 2 - Query B
    so the tag in the xml has to be split by common to these three rows.
    since the data is coming from the diff query, and the data is not going to be under single tag.
    you cannot burst it using concatenated data source.
    But you can do it, using the datatemplate, and link the query and get the data for each file under a single query,
    select distinct store_name from all-stores
    select * from query1 where store name = :store_name === 1st query
    select * from query2 where store name = :store_name === 2nd query
    define the datastructure the way you wanted,
    the xml will contain something like this
    <stores>
    <store> </store> - for store 2
    <store> </store> - for store 3
    <store> </store> - for store 4
    <store> </store> - for store 5
    <stores>
    now you can burst it at store level.

  • Delta Activation for a data source

    Hi All,
    Thank You Very much for reading this query.
    In Source System, I have enabled Delta on Calday field for a generic data source by clicking on Generic Delta Button.
    For this data Source, extraction is from data base table.
    Now Do I need to write any function module for this data source, as I have enabled Delta.
    Would you please help me on this.
    Thanks and Regards,
    K.Krishna Chaitanya.

    Delete duplicate request is a process chain variant, which will allow you to set the duplicate request deletion options. You need to explore those options.
    Now say you have a routine in the DTPs.
    DTP1 = Current Month
    DTP2 = Prevoius Month
    In the month of July,
    DTP1 = 07/2009
    DTP2 = 06/2009
    On daily basis DTP1 will load data for July and DTP2 will load data for June. Now delete duplicate request will delete the previous requests loaded by each of them for the same month. Meaning delete duplicate request for DTP1, will delete the request loaded for 07/2009 and DTP2 will delete the duplicate request for 06/2009.
    Now when you will be in Aug, the routine will change the selection criteria
    DTP1 = 08/2009
    DTP2 = 07/2009, so now DTP2 will delete teh duplicate request for 07/2009 and load the new data. Now the data changed for 06/2009 is not possible to load as you have only 2 DTPs.
    For safer side, you can create 3 DTPs to take care of the last 3 months, but first test the logic for two DTPs and then go further for the third DTP.
    - Danny

  • How to enable the Delta for the data source 0CO_OM_OPA_7

    Hi Experts,
    I have a business content data source 0CO_OM_OPA_7 and I need to activate and enable Delta for this data source.
    This data source is activated (but not delta enabled) and used for full load in my system for the past three year. As this is a full load this data source is carrying more than 4 million records every day with adding 2 millions records every year. The load takes every day 4 to 5 hours for completion.
    Now I need to enable the delta for this data source. To enable the delta for this data source I used the transaction RSA5 and clicked on the 'Activate DataSources' but the data source is not enabled the 'Delta'
    I have also checked in the RSA6 transaction, the check box for 'Delta' is not checked for this data source.
    Can anybody help in enabling the 'Delta' for this data source.
    Thanks in Advance
    Narendra

    Hi,
    DS "0CO_OM_OPA_7" won't support deltas, so load to ODS then to Cube. Else delete the overlaping request then load this you can set in InfoPackage level.
    You check any Period/Month fields in 0CO_OM_OPA_7 then write simple routine in Infopackage and the set the overlaping deletion.
    Use the followng code and change it for your requirement in Infopackage level.
    I'm using 0CO_PC_01 DataSource, so it won't have delta so I'm using the following code in InfoPackageag and set delete overlaping req. So eveyday it will delete the old req for that period and then it will load data againg for that period (Full loads).
    DATA: l_idx LIKE sy-tabix,
            zzdate LIKE sy-datum,
            zzbuper LIKE t009b-poper,
            zzbdatj LIKE t009b-bdatj,
            zzperiod(7) TYPE c.
      READ TABLE l_t_range WITH KEY
           fieldname = 'FISCPER'.
      l_idx = sy-tabix.
      zzdate = sy-datum - 1.
      CALL FUNCTION 'DATE_TO_PERIOD_CONVERT'
        EXPORTING
          i_date               = zzdate
    *             I_MONMIT             = 00
          i_periv              = 'V3'
               IMPORTING
         e_buper              = zzbuper
         e_gjahr              = zzbdatj.
    *           EXCEPTIONS
    *             INPUT_FALSE          = 1
    *             T009_NOTFOUND        = 2
    *             T009B_NOTFOUND       = 3
    *             OTHERS               = 4
      IF sy-subrc <> 0.
    * MESSAGE ID SY-MSGID TYPE SY-MSGTY NUMBER SY-MSGNO
    *         WITH SY-MSGV1 SY-MSGV2 SY-MSGV3 SY-MSGV4.
      ENDIF.
      CONCATENATE zzbdatj zzbuper INTO zzperiod.
      l_t_range-low = zzperiod.
      l_t_range-option = 'EQ'.
      l_t_range-sign = 'I'.
      MODIFY l_t_range INDEX l_idx.
    Thanks
    Reddy

  • Power View in SharePoint Server - The data extension type for a data source is not valid

    Hi All,
    All of a sudden I am getting following error when trying to create Power View report using shared report data source (no error testing the connection):
    "The current action cannot be completed. The data extension type for a data source
    'http://dev/Shared Ducuments/Sales.rsds' is not valid for this operation"
    I already have a data source (I had created it after creating my site collection a week ago) and when I use this source to create Power View report then there is no error but I am getting above error when I create another similar data source and use it to create
    a Power View report.
    Please help me to resolve the error.
    Thanks

    I am going nuts! I had selected 'Analysis Services' instead of 'Microsoft BI Semantic Model for Power View'

  • Discrepency while REPLICATING META DATA (for new DATA SOURCE) in BI side.

    In R/3 I have created a simple TRANSACTIONAL data store based on an INFO QUERY.
    I even checked the veracity of this data store using RSA3 to see whether it extracts data. Works perfectly.
    I go to BW side. Click on the SAP module under which I have created the data source, right click, select 'REPLICATE' and click on it.
    ( I would love to post the screen shot here, but I think I may not be able to paste BMP files here).
    I will write the contents of the POP-UP that appears,
    Title:Data Source from Source System Unknown
    Pop-up contents:
    Data Source (OSOA) DS_01
    does not exist in BI system
    How do you want to create the object in BI?
    1. as DataSource (RSDS)
    2. as 3.x DataSource (ISFS)
    3. this and following 3 as DataSource (RSDS)
    4. this and following 3 as 3.x DataSource (ISFS).
    Well, I normally select option three as per my instructions (without knowing the real reason).
    But sometimes, either for the same data source or for another data sources that I created, the same pop up will appear like this
    Title:Data Source from Source System Unknown
    Pop-up contents:
    Data Source (OSOA) DS_01
    does not exist in BI system
    How do you want to create the object in BI?
    1. as DataSource (RSDS)
    2. as 3.x DataSource (ISFS)
    Just TWO options.
    And, if I select option 1, the data source does not work properly on BI side, though it worked perfectly on R/3 under TRANSACTION RSA3 and showed me data.
    For some unknown reasons, if I delete the erroneous datasource on BI side and sleep overnight and comeback in the morning and replicate, the POP-UP sometimes appears with FOUR options, (Notice the word 'SOMETIMES')
    Can someone explain the secret behind this?
    Thanks again in advance,
    Gold

    3. this and following 3 as DataSource (RSDS)
    That means there are total 3 new (not yet in BI) DataSources available, u wanted to replicate as 7.0 datasource (RSRS) or 3.x datasource (ISFS).
    (other 2 datasources activated from RSA5, or created by other users under that SAP module)
    If there is 1 new DataSource, u will  get just TWO options.
    1. as DataSource (RSDS)
    2. as 3.x DataSource (ISFS)
    After replication with option 1, u should activate datasource in BI, then create infopackages, transformation, DTP etc.

  • Error in Installing the Business Content for 2LIS_06_INV data source

    Hi Experts,
    We are trying to install the Business content for 2lis_06_inv. When we replicate the datasource, we are not seeing it in 3.x version. It is only vailable in 7 version. There is no transformation available in Business content for this data source. Please let me know how we need to proceed with this installation. Should we install the 7.0 flow or 3.x flow? Please provide th details if you have implemented th same in your system. Thanks is advance.
    Regards,
    Kavitha Jagannath

    Hi,
    Follow bellow steps:
    1. Goto RSA1.
    2. Select the BI content.
    3. Select the Grouping as' In Dataflow before and after'
    Then,you can get the data flow which are haveing the DS/TR/Data targets.
    4. Select the Install by setting the 'Install in background'.
    Regards.

  • Tables for Credit data

    Hi all,
    please tell me the tables  for Credit data.
    Thanks,
    kishore.

    1) Check to see if there are any entries in table S066 and S067. These SIS tables hold the credit data for sales order value and delivery order value.
    hope it helps

  • Tables for Contract data

    Hi all,
    if you know any tables for contract data(For Contract report),
    please tell me.....
    Thanks,
    kishore.

    VBAK-contract header
    VBAP-contract item
    VEDA-changed contracts
    KONV-condition records
    MARA-material related to contract

  • Tables for Budget data

    Hi,
    Please provide the base tables for Budget data
    thanks

    hi Arun,
    check these
    BPGE (overall budgets),
    BPJA (annual budgets) and
    BPPE (period budgets).

  • Data Not coming Properly for 0HR_PA_OS_1 Data Source?

    Hi Experts,
    I am trying to load data to 0PAOS_C01 using 0HR_PA_OS_1 data source. For field 0POS_OCCVAC key figure i am not getting any data from Data Source.
    I checked in RSA3 for 0HR_PA_OS_1 data source and  in that it is not populating data for filed OCC_VAC_PERC  field. I checked with HR functional team and they mentioned its a BI issue.
    Kindly tell me how can i populate values for the particular field. FYI i am getting correct data for other fields.
    Please help me.
    Thanks

    The possible reasons
    1.Your user id might not be having the proper authorization to pull the HR data.This you can confirm with your basis guy,if you are unable to find what are the authorization roles required for HR data, for the time being you can take SAP_ALL authorization in development box and try to pull.
                            OR
    2.Try to pull from BW side,you may be able to see the data there because this uses the background user id used in RFC connection ( provided if your background user id having proper authorization ).

Maybe you are looking for