Master data coming from which particular table of source system

Hi,
my requirement is i have a master data (ABC) coming from a source sytem XYZ
there is an attribute (L) in my master data. Now i want to know from which table in my source sytem the data pretaining to attribute L is coming up.
I m using BW 3-5 system
Is there a way through which i can directly find out abt the same.
thanks

Hi,
Is this standard Master data or Generic one, if it is customised one then you can find the same in customer exit of ECC system,
go to tcode CMOD in ECC system, there willbe an project defined for the BW system, in that get an exact component, and check for the custom code for your masterdat, there you will be in a position to find out from where the data is being extracted.
if you are good at abap you can debug it via RSA3 , else get an help of an abaper to find it out.
if it is standard , then you can search in SDN, but for for better reuslts, you can go to RSO2 and give the datasource name of that master data and clik on display you, will get an display in the bottom saying that this data source using FM 'XXXXXXX', copy that FM name and try to give it in SE37 and debug it via RSA3 for better results.
Regards,
Nanda.S

Similar Messages

  • Master Data coming from 1 backend and GOA distributed to many bakends

    Hello,
    In our case (using SRM 5.0) one backend system is used to replicate all material master data and GOA will be distributed to multiple backends including the one that is used for master data. As far as i know the logical backend system is determined through materials/material groups while GOA is distributed. In the above case how can we intervene with this logic to be able to distribute the GOA to whatever backend we have, even though if the MD is not coming from them?
    Thank you,

    Hi,
    You can use the BADI BBP_DETERMINE_LOGSYS for this purpose.
    The method :CONTRACT_LOGSYS_DETERMINE (Determine Target System of Contract) has been provided for the same purpose.
    Sample BADI Implementation at SDN Wiki:
    https://www.sdn.sap.com/irj/sdn/wiki?path=/display/srm/bbp_determine_logsys-contract_logsys_determine&
    Regards
    Kathirvel
    Edited by: Kathirvel Balakrishnan on Feb 25, 2008 1:00 PM

  • Master data Load from Oracle table

    Hi,
    My master data comes from an Oracle table. The table has both attributes and texts combined. How do I load the Master data into the text and attribute info objects thru Direct Upload of Master Data?
    Is it necessary to go in for Flexible Upload in this case?
    Regards

    Hi,
    you can create two views on the table. One for the text and the other one for the attributes. Create your datasources on these views and assign them to your infoobject.
    regards
    Siggi

  • Fill a table with data coming from an RFC

    Hello everyone:
    I've followed the Weblog "How many lines of java code did i write for a simple Web Dynpro?"
    /people/durairaj.athavanraja/blog/2004/10/17/how-many-lines-of-java-code-did-i-write-for-a-simple-web-dynpro
    I've called an RFC and created a table with data coming from it (which is also a table). My question is, if in this table there's a field named "UserType" there are two possible values for this field:
    "userA"
    "userB"
    How can I get the table only show me the "userA" registers? The RFC does return all of the users, but when filling the table, can I put an if-else somewhere on my code?
    Thanks a lot
    Alejandro

    Hi Alejandro,
    Referring to the link provided "The logic of the filter process is not implemented in Web Dynpro. The application developer must implement the action to be executed."
    We would have to implement the action onFilter in the controller implementation. Ideally, we fill the data retrieved from backend into a List (java.util.List) (this could be done on init of view) and then subset the list after meeting the criteria in the action handler(say
    onActionFilterData(com.sap.tc.webdynpro.progmodel.api.IWDCustomEvent wdEvent).
    Having done this, you may bind the output list back to the node (shown in table)
    Regards,
    Chaitanya

  • ATWRT field is from which database table.

    Hi Gurus,
        I need to see the field name ATWRT(Manufac Method) for the particular material number.from which database table can i access both the data.Please help me .
    Thanks in advance.
    Points will be rewarded.
    Thanks & Regards,
    A.Ashok kumar.

    Hi
    these are the tables where atwrt exists:
    ABAUSP
    AFFV
    AFFVB
    AFFVD
    APICAWN
    API_CAWN
    API_CCHAR
    API_VALI
    API_VALUE
    API_VAL_IN
    CCIHS_IHPRIOT
    CCRCS_EHS_SVT_MA
    CLGO_MULTIPLE
    CLRS_OBJ_KEYS
    CLRS_OBJ_VALUES
    CLSELSTATISTICS
    CLS_CHARAC
    COFIV
    COFV
    COFVP
    COME
    COMEP
    COMER
    COMSE
    COMW
    COVLP
    CRMT_IC_SCRIPT_M
    CSIM_ST_DETAIL
    CTBW_VALUE_TEXT
    CTIH_01
    CTMSEXIT001
    CTMS_02
    CUCD_CONF
    CUD0_02
    CUGEN_SIMPLE_LIT
    CUGEN_SIMPLE_SIN
    CUGEN_VALUE
    CURTO_CU_GEN_SIMPLE_L
    CURTO_CU_GEN_SIMPLE_S
    CUSL_01
    CUSL_02
    CUSL_05
    CUSL_06
    CUSTOMER_FILTER_CHAR
    CUXREF
    CUXR_OBJ
    CUX_EXP_VALUE
    DDB_AW
    DDB_C01
    DDB_C02
    DDB_C03
    DDB_C04
    DDB_C06
    DDB_C07
    DESCR
    DIWPS_ATINN
    DIWPS_CA
    DIWPS_CONF
    DIWPS_CONF_MARK
    DIWPS_RV_ATINN
    DIWPS_RV_CA
    DIWPS_RV_CONF
    DIWPS_RV_CONF_MARK
    DIWPS_RV_NAV
    DIWPS_WA_WORKPACKAGE
    DIWPS_WPID_CONF
    DIWPS_WP_NAV
    E1AUSPM
    E1CAWNM
    E1CUV1M
    E1EDL12
    E1EDL15
    E1ISU_SDORD_POS_CONF
    E1WBB17
    DIWPS_ATINN
    DIWPS_CA
    DIWPS_CONF
    DIWPS_CONF_MARK
    DIWPS_RV_ATINN
    DIWPS_RV_CA
    DIWPS_RV_CONF
    DIWPS_RV_CONF_MARK
    DIWPS_RV_NAV
    DIWPS_WA_WORKPACKAGE
    DIWPS_WPID_CONF
    DIWPS_WP_NAV
    E1AUSPM
    E1CAWNM
    E1CUV1M
    E1EDL12
    E1EDL15
    E1ISU_SDORD_POS_CONF
    E1WBB17
    E1WTADDI12
    E2AUSPM
    E2CAWNM
    E2CUV1M
    E2EDL12
    E2EDL15
    E2WBB17
    E3AUSPM
    E3CAWNM
    E3CUV1M
    E3EDL12
    E3EDL15
    E3WBB17
    ECONF_OUT
    EHQMSAPIM1
    Regards
    Preeti
    <b>
    Reward if useful</b>

  • My requirement is to update 3 valuesets daily based on data coming to my staging table. What is the API used for this and how to map any API to our staging table? I am totally new to oracle and apps. Please help. Thanks!

    My requirement is to update 3 valuesets daily based on data coming to my staging table. What is the API used for this and how to map any API to our staging table? I am totally new to oracle and apps. Please help. Thanks!

    Hi,
    You could use FND_FLEX_LOADER_APIS.UP_VALUE_SET_VALUE to upload them from staging table (I suppose you mean value set values...).
    You can find a sample scripts if you google around.
    What do you mean "how to map any API to our staging table" ?
    You should do at least the following mapping (which column(s) in the staging table will provide these information):
    - the 3 value sets name which you're going to update/upload (I suppose these are existing value sets or which have been already created)
    - the value set values and  description
    Try to start with something and if there is any issues the community could then help... but for the time being with the description of the problem you have provided, that's the best I can do...

  • How to  extract  master data  attribute from  r/3 to bw give steps details

    how to  extract  master data  attribute from  r/3 to bw give steps details screenshots

    Hi
    Go through the below process to extract Master Data Attribute from R/3
    Hi,
    Maintaining Generic DataSources
    Use
    Regardless of the application, you can create and maintain generic DataSources for transaction data, master data attributes or texts from any transparent table, database view or SAP Query InfoSet, or using a function module. This allows you to extract data generically.
    Procedure
    Creating Generic DataSources
    1. Select the DataSource type and specify a technical name.
    2. Choose Create.
    The screen for creating a generic DataSource appears.
    3. Choose the application component to which you want to assign the DataSource.
    4. Enter the descriptive texts. You can choose any text.
    5. Select the datasets from which you want to fill the generic DataSource.
    a. Choose Extraction from View if you want to extract data from a transparent table or a database view. Enter the name of the table or the database view.
    After you generate the DataSource, you have a DataSource with an extraction structure that corresponds to the database view or transparent table.
    For more information about creating and maintaining database views and tables, see the ABAP Dictionary Documentation.
    b. Choose Extraction from Query if you want to use a SAP Query InfoSet as the data source. Select the required InfoSet from the InfoSet catalog.
    Notes on Extraction Using SAP Query
    After you generate the DataSource, you have a DataSource with an extraction structure that corresponds to the InfoSet.
    For more information about maintaining the InfoSet, see the System Administration documentation.
    c. Choose Extraction Using FM if you want to extract data using a function module. Enter the function module and extraction structure.
    The data must be transferred by the function module in an interface table E_T_DATA.
    Interface Description and Extraction Process
    For information about the function library, see the ABAP Workbench: Tools documentation.
    d. With texts you also have the option of extracting from fixed values for domains.
    6. Maintain the settings for delta transfer, as required.
    7. Choose Save.
    When performing extraction, note SAP Query: Assigning to a User Group.
    Note when extracting from a transparent table or view:
    If the extraction structure contains a key figure field that references a unit of measure or a currency unit field, this unit field has to be included in the same extraction structure as the key figure field.
    A screen appears on which you can edit the fields of the extraction structure.
    8. Edit the DataSource:
    &#9675; Selection
    When you schedule a data request in the BI scheduler, you can enter the selection criteria for the data transfer. For example, you can determine that data requests are only to apply to data from the previous month.
    If you set the Selection indicator for a field within the extraction structure, the data for this field is transferred in correspondence with the selection criteria in the scheduler.
    &#9675; Hide field
    You set this indicator to exclude an extraction structure field from the data transfer. The field is no longer available in BI when you set the transfer rules or generate the transfer structure.
    &#9675; Inversion
    Reverse postings are possible for customer-defined key figures. Therefore inversion is only active for certain transaction data DataSources. These include DataSources that have a field that is marked as an inversion field, for example, the update mode field in DataSource 0FI_AP_3. If this field has a value, the data records are interpreted as reverse records in BI.
    If you want to carry out a reverse posting for a customer-defined field (key figure), set the Inversion indicator. The value of the key figure is transferred to BI in inverted form (multiplied by –1).
    &#9675; Field only known in exit
    You can enhance data by extending the extraction structure for a DataSource by adding fields in append structures.
    The Field Only Known in Exit indicator is set for the fields of an append structure; by default these fields are not passed to the extractor from the field list and selection table.
    Deselect the Field Only Known in Exit indicator to enable the Service API to pass on the append structure field to the extractor together with the fields of the delivered extract structures in the field list and in the selection table.
    9. Choose DataSource ® Generate.
    The DataSource is saved in the source system.
    Maintaining Generic DataSources
    &#9679; Change DataSource
    To change a generic DataSource, in the initial screen of DataSource maintenance, enter the name of the DataSource and choose Change.
    You can change the assignment of a DataSource to an application component or change the texts of a DataSource. Double-click on the name of the table, view, InfoSet or extraction structure to get to the appropriate maintenance screen. Here you make the changes to add new fields. You can also completely swap transparent tables and database views, though this is not possible with InfoSets. Return to DataSource maintenance and choose Create. The screen for editing a DataSource appears. To save the DataSource in the SAP source system, choose DataSource ® Generate.
    If you want to test extraction in the source system independently of a BI system, choose DataSource ® Test Extraction.
    &#9679; Delta DataSource
    On the Change Generic DataSource screen, you can delete any DataSources that are no longer relevant. If you are extracting data from an InfoSet, delete the corresponding query. If you want to delete a DataSource, make sure it is not connected to a BI system.
    NR

  • How to insert data coming from 2 different file adapters in to one DB adapt

    Hi
    i want insert data in to database containing two diffferent tables, so i imported tables in to DB adapter by creating relation ships.But, data for two tables are in xml format & two are in different locations.So, i used 2 file adapters to get data from 2 different & i used BPEL(Define service later) Service. now in bpel i used receive activity to receive one file adapter data ( i checked create instance in receive) then used transform activiy for tranformation & finally invoke activity to invoke DB adapteer........similarly i repeated sequetially to 2 file adapter, by keeping 2nd receive(no need to check create instance) next to invoke.*Problem is after deployment finished only data coming from 1st receive is inserting to the table...& 2 nd receive not working it showing as Pending & showing as Asynchronus Call back inte console*
    I configured all the adapters perfectly..........Can any one can help me how to Commit 2 nd receive to insert data in the 2nd table
    Regards,
    jay

    Thank u both for ur replay.........
    I am doing this in 11g there is no problem regarding transform activity.
    My requirement is
    two different files from two different folders in a drive & we can't use one file adapter bcoz both have different columns(only few are common columns) so we use two different xsd's .So, i am using two file adapters to insert in database having two different tables with respect to two different files data coming from two file adapters. i am using one DB adapter to insert bcoz both tables are in same database with relationships & i used BPEL(define service later) .
    NOW PLEASE SUGGEST ME THE FLOW IN BPEL TO INSERT BOTH FILES IN THERE RESPECTIVE TABLES IN DATABASE.
    The flow i did 1st file adater--->receive--->transform---->invoke----->DB adapter.....Then again repeating this as keeping 2nd receive below 1st invoke
    2nd file adapter-------->2nd receive---->2nd invoke------>same DB adapter
    MY problem is only data coming from 1st process is inserting & 2nd one is not working as i discussed earlier........I USED READ FILE OPTION, UNCHECKED DELETE FILES OPTION & SET DIFFERENT POLLING FREQUENCY FOR BOTH FILE ADAPTERS.
    I tried to set correlation but it is not working & later tried i kept non-blocking invoke as TRUE in DB adapter also didn't work...........also i tried this transaction property in bpel component _<property name="bpel.config.transaction"_
    many="false" type="xs:string">required / requiresNew</property>...............BUT NO CHANGE...........
    Regards,
    jay
    Edited by: 910162 on Apr 5, 2012 12:38 AM

  • I want to create generic master data attribute datasource on ANLA table .

    hI,
    I want to create generic master data attribute datasource on ANLA table .
    and have to activate delta load.
    Is it possible for delta upload on this table?
    if it is possible then on which field i can create genric delta?

    1.I want to fetch quantity ,inventory no , serial no.......
    I have to enhance that datasource.....
    2.else i can create master data for asset number using ANLA
    I have opted 2nd option

  • Data deleted in ODS, data coming from the data source : 2LIS_11_VASCL

    Friends,
    I have situation :
    Data deleted in ODS, data coming from the data source : 2LIS_11_VASCL.
    when for the above ods and the data source when trying to delete the request whole data got delted.
    All the data got deleted in the fore ground. no background job has been generated for this.
    I ma really worried abt this issues. can u please tell me what should be the possibilities for this issue.
    Many Thanks
    VSM

    Hi,
    I suppose you want to know the possibilitiy of getting the data.
    If the entire data is being deleted, you can reload the data from source system.
    Load the setup table for your application. Then carry out init request.
    Please note that you would have to take a transaction-free period for carrying out this activity so that no data is missed.
    Once this is done, delta queues will again start filling up.

  • Is it possible to update master data attributes from an ODS?

    HELLO ALL,
    we have records coming into our ods like the following:
    costcenter1, subcostcentera, subcostcenterb, manager responsible, costs (kf). 
    This is a custom flat file load from a legacy system.
    We would like to just create an update rule from the ODS to the cost center mater data characteristic attributes. 
    Is this possible?
    thank  you

    Yes its possible to update the master data attributes from an ODS.
    Define your master data characteristics as a infoprovider.
    Create a update rule on Characteristic with ODS as a source and do general mapping.
    For step by step, pls refer..
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sapportals.km.docs/library/business-intelligence/g-i/how%20to%20implement%20flexible%20master%20data%20staging

  • How to delete Master Data transferred from ECC to APO via CIF

    Hi, all
    I'm connecting SCM 5.1 with ECC 6.0 via CIF.
    When activating Integration Model in ECC, the following error occurred;
       "Location 0001 of type 1001 (BSG 592BUSGP) already exists in APO"
    This is because Location 0001 was already transferred from different ECC client via CIF, and not cleared.
    I'm now trying to initialise master data transferred from ECC to APO via CIF, but in vain.
    I suppose I have to first delete all master data CIFed from ECC in APO, or have to execute initialisation of APO master data with ECC data in ECC.
    Please tell me how to do this.
    For details, please read the followings;
    I connected APO client with ECC client 590.
    I transferred Plant and Material Master from ECC to APO via CIF.
    After that, I wanted to change ECC client from 590 to the other.
    I created the other client (592) by client-copying from 590 with SAP_ALL copy profile.
    I deactivated Integration Model in ECC 590, and activated the same in ECC 592.
    Here, I faced the problem.
    The initial data transfer failed due to the duplicate of data.
    Now I'm wondering how I can initialise master data in APO and transfer initial data from ECC.
    Due to testing purposes, I have to change ECC clients from one to the other this time, but I cannot initialise master data in APO and cannot tansfer initial data from different ECC client.
    thanks in advance,
    Hozy

    Hi all,
    Thank you very much indeed for lots of your replies.
    Thanks to all of your advices, I'm getting to understand what to do.
    Well, I'll say I have to create one more client for testing with other ECC clients or I have to rebuild the client by client-copying from 001 and implementing all transportation requests, for marking del flags to each master data and running del programme sound like tough work, and sound like odd ..
    Then, I've got next issue here.
    In SCM-APO, I use BI objects, and , if I have to create several clients for testing, I have to set up clients where you can use both BI objects and CIFed master data.
    I suppose I have to change 'BWMANDT' in RSADMINA and make RSA1 accessable in other SCM clients and implement tests.
    Also, I have to register BI System for newly created SCM clients in RSA1-SourceSystem.
    Well, if you have any experiences on this kind of use, and if you have any advice on it, please give me any tips.
    Particulary, I'd appreciate very much if you give me advice on which way you have chosen; creating several clients or rebuilding the one client in SCM-APO when you use both BI objects, CIFed master data and Demand Planning.
    Thanks in advance.
    Hozy

  • Customer master data updation from one system to another system

    HI  Experts,
    I have an issue with customer master data updation from one system to another system. And here only one field is not updating properly from one system to another system. and we found this is through idoc , but we don't find field in any segment.
    and the field is called ADDR1_DATA-HOME_CITY.
    and this will displayed as different city in address column.
    but as per the notes given in help , it is by program , form through only passing. but not able to trace and as per my analysis ADRC table , ADRNR  had this field HOME CITY.
    And this is for some customers updating wrong one from one system to another system , and for some customers field itself not updating in address column.
    please can you suggest any one. here are screenshot for your reference.

    that is not enough, it was enough in times before 40b release, but this was in the  last century
    Please read OSS note 384462 - Master data and addresses
    it explains what you have to do for ALE

  • How to store the data coming from network analyser into a text or excel file

    Hii everyone
    I'm using Agilent 8719ET network analyser and wish to store the data coming from netowrk analyser into a text file/ excel file.
    Presently I'm able to get the data on Labview graph using GPIB . Can anyone suggest how to go ahead after collect data sub vi. How can the data be stored into a file apart from showing on the graph?
    Attached is the vi for kind consideration...
    Looking for help
    Regards
    Rohit
    Attachments:
    Agilent 87XX Series Exceed Max Meas.vi ‏43 KB

    First let me say that your code really looks pretty good. The data handling could be made more efficient by calculating the number of datapoints that are going to be in the completed dataset and preallocating the entire array -- but depending upon your answer to my questions, the logic in the lower shift register may be going away - so we won't worry about that right now.
    The thing I need to know before addressing the data storage question is: Each time you call "Collect and Display Data.vi", how many element are in the array? Are you reading single data points, or a group of data? (BTW: if the answer to that question is obvious based on the way the other VIs are setup, I don't have the drivers so I can't tell what the setup values are.) Second, how fast does the loop iterate? Are we talking msec per loop?, seconds? fortnights?
    The issues here are two-fold: how much data? and how fast is it coming? The answer to these will tell you how to save the data.
    Mike...
    Certified Professional Instructor
    Certified LabVIEW Architect
    LabVIEW Champion
    "... after all, He's not a tame lion..."
    Be thinking ahead and mark your dance card for NI Week 2015 now: TS 6139 - Object Oriented First Steps

  • Master data extraction from SAP ECC

         Hi All,
    I am a newbie here and teaching myself SAP BI. I have looked through the forums regarding master data extraction from SAP ECC in all forums but could not answers for my question. Please help me out.
    I want to extract customer attributes from SAP ECC. i have identified the standard data source 0customer_attr and replicated in SAP BI. I have created infopackage for full update. I validated the extractor checker(RSA3) and found 2 records for 0customer_attr.
    When I run the info package, the info package remains in yellow state until it times out. Please let me know in case i am missing anything, Please let me know if there is any other process for master data extraction similar to transaction data extraction.

    Hi All,
    i did the below and afte clicking execute in the Simple job, it takes me back to the Monitor Infopackage screen.
    From your info pack monitor --> menu environment-->job overview--> job in the source system--> prompts for source system(ECC) user id and password, enter them, you will get your job at SM37(ECC), select job and click on log details icon. there you can see details about your error. please share that error screen shot.
    Please find the screenshots.
    I did an Environment -->Check connection and found the below,

Maybe you are looking for

  • How do I get the Time Capsule to automatly erase old back ups?

    I have a 1TB Master HD in my Mac Pro that I back up using a 1 TB Time Capsule. So far I've used just over 500GBs on the Hard Drive which has really stumped my Time Capsule because it comes up saying it needs 400 odd gbs to do a full back-up of my Mas

  • Ibooks battery  power from 22 % to 0%

    hi all , I was looking for this issue if any one has faced it but can't find any. I had an year old Ibook g4 and it was working fine untill some day it was low in battery. I saw it at a 33 % level at first and within seconds it responded that battery

  • Styling in INDESIGN

    Hello, This is quite a tough question to ask written. I am trying to find a quicker way to stylize a row of numbers in my inDesign file. Currently I am making each number (see Visual attached) a mustard yellow by using character styles, but this requ

  • Early 2008 Mac Pro-What type of HD?  Smart failing?

    I have an early 2008 Mac Pro and have two hard drives in it.  The main system drive is a 2TB Western Digital drive I installed about a year ago.  The 2nd drive is a 1TB Seagate drive which is the original drive that came with the Mac Pro. When lookin

  • [SOLVED] Flashplugin 11.2.202.236-1 flash crash on any browser

    Sadly, I found that flashplugin just plain crash in latest chrome (Chromium 20.0.1132.57) and firefox (Firefox 14.0.1), I tried to downgrade flashplugin (until 10.x)[1] but the problem persists. I also tried to disable hardware acceleration but this