Master Data R/3 Tables

What are the R/3 Master Data Tables for the below objects?
1)Accounts
2)Order
3)Cost Center
4)Profit Center
5)Company Code

Hi,
Tcodes
1)Accounts > <b>FS00</b>
2)Order > <b>KO02</b>
3)Cost Center > <b>KS02</b>
4)Profit Center > <b>KE52</b>
5)Company Code > <b>IMG > Enterprise Structure > Definition > Financial Accounting > Define company</b>
6)Cost element
<b> - table CSKA, CSKB, text table CSKU
- tcode KA02</b>
Hope this helps.
Cheers,
Gimmo

Similar Messages

  • QM action not allowed for DTP requests in master data and Text tables

    Hi,
    I have a master data object load request which has failed, returned error in st22 and job finished.
    BUT the request has not turned red in the monitor - it's still yellow. That means that I can not delete request nor start new load to infoprovider because it believes a request is still running. But it's not, I've checked both sm37, sm50 etc.
    When trying to manually change QM status to 'red' it says 'QM action not allowed for DTP requests in master data and Text tables'
    How can I force QM status to red so that I can move on?
    Running NW2004s BI (7.0) Patch 15.
    I searched for this question but there is no answer
    Thank you

    Folks
    I know how frustrating this problem is even in netweaver 7.0 environment. I found a solution for this problem atlast and it works but a not direct way of resolving this issue.
    When this request is in yellow status and not able to change / delete the request, its actually in a pseudo status and it is a bug in the program. This request sits in table RSBKREQUEST with processing type as 5 in data elements USTATE, TSTATE and this 5 is actually "Active" status which is obviously wrong.
    All we need is, ask your basis person to change the status to 3 in both these data elements and then it allows to reload the delta. Once the delta is successfully loaded, you can delete the previous bad request even though it is still in yellow. Once the request is deleted, the request status gets updated as "4" in table RSBKREQUEST.
    Hope this helps.
    Thanks

  • 'Master data type User table cannot add row'-DTW error

    Hi All,
          I am creating a template for user define Master data table from DTW, when i am trying to port data by using that template through DTW. it is giving error like--'Master data type User table cannot add row'
    So can any one have solution for this.
    Regard's
    Hari

    Hari,
    Please see SAP Note 1234690 on the SAP PartnerEdge Portal.  This seems like a similar problem although the example is using the Business One SDK which uses the DI API.  The DTW also uses the DI API ... so there may be a relation.
    You may want to check the latest patch level for SAP Business One 2007A as the note says it is a known issue.
    Eddy

  • Master data : T077Y (Polled Table extarction )

    Hi Guys,
    I want to extract the Master Data Attributes (Account group and Account Group Name) from T077Y Table in ECC and bring it to BI, I use generic datasource using RSO2 it is not allowing me to create it and saying u201CExtraction from pool tables and cluster tables is not permitted (Message no. RJ037)u201D  
    T077Y is a polled table, Extraction from pool tables and cluster tables is not permitted.
    Whatu2019s the best way the extract the Master Data attributes from this table T077Y,
         1. SAP Query
         2. Function Module
    Whatu2019s the step by step for SAP Query.
    Thanks and best regards
    Ilyaebi

    Hi,
    You can go for any of these methods,
    In case of FM based extraction you will have to write some ABAP code to fetch the data, but that kind of coding is not required in SAP query based extraction.
    Check the below given step by step guide,
    [http://www.sdn.sap.com/irj/scn/index?rid=/library/uuid/9018e4f3-5537-2e10-238e-ef06d9ab31e8]
    Regards,
    Durgesh.

  • Master data abt sid table

    hi friends,
    i ahve some doubts abt generating sid table with master data, according to external star schema when we will activate master data only , sid will be generate automatically, actually i created info object, i switch off with master data and text after activationg my info object its still showing sid table from char info object also.
    if i activate with master data then only its need to show, why its showing now.please mail me.
    Thanking u
    suneel.

    why if i will activate info object its showing master data?
    the system defaults all the Io's as msterdata objects, you need to go the masterdata/texts tab - uncheck the check box for both with masterdata and with texts.
    Even though you dont have any attributes available for a IO, but still the IO would be created to hold some values as a part of either transaction data or just masterdata values for that IO. If you load transaction data for the Zobject you created, the system would populate the corresponding fields in the Stable, by assigning SID;s to the values that have been populated thru transaction data. So every IO will have a Stable, eventhough it might or might not have corresponding master data.

  • Guidance for getting material master data from database table

    Hi,
    I need guidance to fetch following Material master data from the system data base.Please guide me for the same.
    BASIC DATA1
    BASIC DATA2
    MRP1,2,3,4
    WORK SCHEDULING
    QUALITY MANAGEMENT
    ACCOUNTING1
    ACCOUNTING2
    COSTING1,2
    PURCHASING
    PURCHASE ORDER TEXTSALES ORG1,2
    SALES GENERAL/PLANT
    PLANT DATA/STOR.1,2
    WAREHOUSE MGMNT1,2
    Also please tell me in general what is the reason for error while uploading the data into system using BAPi(AM talking about return message error).
    Thank you.
    Edited by: sanu debu on Feb 24, 2009 12:41 PM
    Edited by: sanu debu on Feb 24, 2009 12:42 PM

    Use BAPI's 
    BAPI_MATERIAL_GET_DETAIL
    BAPI_MATERIAL_GETALL
    BAPI_MATERIAL_GET_ALL

  • ADF Faces Query Panel: How to display master data in detail table?

    Before stating the question, here is the background:
    I've a Fusion ADF (model/viewController) app that includes 2 tables with a master-detail relationship. The master table stores error numbers and associated levels of logging severity; the detail table stores the actual message text. These need to be separate, related tables because the requirement is that message text for a given error number can vary by product release.
    Messages are added in a bounded task flow; data for the master table entered in 1 fragment, data for the detail table in a 2nd page fragment. In the app, the Entity Objects for these tables are related with a foreign key association and have a composition association. This was done because the error numbers are generated in an Oracle DB sequence, and from what i have read that is a way in a task flow to have ADF put the number generated for the master table into the related key of the child table.
    The question is this:
    I need to create an ADF query panel with a table that searches on the detail DB table. However, i also need to use a couple of fields from the master DB table in both the query and in the ADF table. In addition, I need to add the ability for the user to edit either the master or detail record for the row he/she selects in the ADF table. I know how to create a view that would display the info i need; however, i have not been able to figure out how to obtain the keys for the selected ADF table row and pass them to the task flows that would edit.
    The master table has a numeric primary key. The detail table primary key consists of a foreign key to the master key and also a release number.
    I'm new to ADF and have been struggling with how to do this: i've a workaround that is functional but too klunky to be a permanent solution. So, could someone point me to some information on how to accomplish this?
    Versions: JDeveloper 11.1.1.4, ADF Business Components 11.1.1.59.23
    Thanks for your help!

    Thanks. I tried this out, but ran into an issue in that the detail table has 2 keys - a foreign key to the master primary key, and a 2nd key that uniquely distinguishes the detail. So i tried various things centered around using the SetCurrentRowWithKey operation and trying to obtain the selected ADF table row using an EL OF #{data.<pageDef>.<Iterator>.currentRowKeyString}; however i could not get the detail record for the selected row to display.
    I did find another approach that worked - I am wondering, is there anything problematic with this approach?:
    1. Create a VO joining the Master and Detail tables, with the Detail table being updatable.
    2. Made sure the applicable attributes were defined as keys.
    3. Add a query panel w/table using this VO.
    4. Create an ADF form based on the VO, showing the updatable fields.
    5. Create a bounded task flow with the following set on the Behavior tab:
    a. Share data controls with calling task flow
    b. Always Begin Transaction
    I specifically am wondering whether "Share data controls with calling task flow" has any negative implications or gotchas to handle. The doc at http://download.oracle.com/docs/cd/E21764_01/web.1111/b31974/taskflows_parameters.htm#ADFFD1693 doesn't suggest so, but just wanted to double-check before I go down this path.

  • BAPI to upload master data in multiple tables

    Dear all,
    Is there any BAPI to update the multiple database tables at a time? I want to update the client data into MARA, A108, KNOP and KNOH tables.
    Is there any other procedure to do it? Pleae give me some suggestions to do it.
    Thank You In advance..!
    Regards,
    Praveen

    For custom defined field use EXTENSIIONIN table in Tables section of BAPI. Below is the description of that table. Check the long text (green button after short text) available to learn more about it.
    Short Text
    Reference Structure for BAPI Parameters EXTENSIONIN/EXTENSIONOUT
    Description
    You use this structure to transfer the material's customer-defined fields. For information on transferring these fields, see the function module documentation.
    I doubt if there is any BAPI for T023T. In that case you need to use BDC

  • Master Data Background Tables

    Hi Xpertz
    Could any one tell me the answer to the following sorry to asking this silly question in Sdn .After a long time just i am
    reviewing my basics i got alittle confussion in the following
    what r the master data back ground tables i know that P,Q,M,X.tell me what is the difference between thez tables.
    If any docs please send it to the following id : [email protected]
    Any answer will appreciate
    Thanx In advance
    Best regards
    R K
    **NO need to say good answers will get full pointz*****

    Hi,
    In fact, do you really need to know the names of these master data tables? The same for the tables of a cube...
    SAP BW will generate these tables for you, depending of your settings. An attribute is time dependant, then SAP BW creates a table to store at least that attribute with Dateto and Datefrom. One other is time dependant AND navigational. Then SAP BW will generate another table.
    Your texts are language and/or time dependant, then again SAP BW will do its job.
    Same for the cubes.
    In 5 years of BW consulting and training, I checked only once the technical settings of a cube, because a full extract of that cube (containing only one request) was generating less rows in a backup cube (completely similar).
    And I could learn that the key of the fact table of a cube is not unique when you run SAP BW under ORACLE.
    But, once again, I could have let it beside, because the results were good, the backup was correct, and I investigated only because I was curious.
    So, do you really need to know them by heart?
    Regards,
    Laurent

  • How to trace master data source (table-field)?

    Hello All,
    Please help, its Urgent!!
    I have the following 4 characteristic requirement in BW from PM/PS module.
    Master data source (format: table-field)
    AFVV-FSAVD - early start date
    AFVV-FSEDD - early finish date
    AFVV-SSAVD - late start date
    AFVV-SSEDD - late finish date
    can anyone tell me the procedure, how to find them in R/3 n if there r any standard business content for the same?
    Thanks in advance,
    JP

    Go to RSA5 Tcode in R/3, under SAP R/3 Application component, you can get PO/PS Components, and under that you can get PO-IO/PS-IO Components, here you can check all the vaialble master data datasources.
    KJ!!!

  • Master data tables and Transaction data Tables

    Hello Gurus,
    Please let me know how to know which table belongs to master data  and which table belongs to transaction data.
    for FICO module.
    Does any one  have specific material relating to master data table and transaction data tables.
    Thanks
    Edited by: Manu Rathore on Jan 18, 2012 4:38 AM

    Hi Manu,
    Find attached table relation diagram by Christopher Solomon. It is one of the very comprehensive chart on this topic.
    deleted
    Warm regards,
    Murukan Arunachalam

  • Performance: reading huge amount of master data in end routine

    In our 7.0 system, each day a full load runs from DSO X to DSO Y in which from six characteristics from DSO X master data is read to about 15 fields in DSO Y contains about 2mln. records, which are all transferred each day. The master data tables all contain between 2mln. and 4mln. records. Before this load starts, DSO Y is emptied. DSO Y is write optimized.
    At first, we designed this with the standard "master data reads", but this resulted in load times of 4 hours, because all master data is read with single lookups. We redesigned and fill all master data attributes in the end routine, after fillilng internal tables with the master data values corresponding to the data package:
    *   Read 0UCPREMISE into temp table
        SELECT ucpremise ucpremisty ucdele_ind
          FROM /BI0/PUCPREMISE
          INTO CORRESPONDING FIELDS OF TABLE lt_0ucpremise
          FOR ALL ENTRIES IN RESULT_PACKAGE
          WHERE ucpremise EQ RESULT_PACKAGE-ucpremise.
    And when we loop over the data package, we write someting like:
        LOOP AT RESULT_PACKAGE ASSIGNING <fs_rp>.
          READ TABLE lt_0ucpremise INTO ls_0ucpremise
            WITH KEY ucpremise = <fs_rp>-ucpremise
            BINARY SEARCH.
          IF sy-subrc EQ 0.
            <fs_rp>-ucpremisty = ls_0ucpremise-ucpremisty.
            <fs_rp>-ucdele_ind = ls_0ucpremise-ucdele_ind.
          ENDIF.
    *all other MD reads
    ENDLOOP.
    So the above statement is repeated for all master data we need to read from. Now this method is quite faster (1,5 hr). But we want to make it faster. We noticed that reading in the master data in the internal tables still takes a long time, and this has to be repeated for each data package. We want to change this. We have now tried a similar method, but now load all master data in internal tables, without filtering on the data package, and we do this only once.
    *   Read 0UCPREMISE into temp table
        SELECT ucpremise ucpremisty ucdele_ind
          FROM /BI0/PUCPREMISE
          INTO CORRESPONDING FIELDS OF TABLE lt_0ucpremise.
    So when the first data package starts, it fills all master data values, which 95% of them we would need anyway. To accomplish that the following data packages can use the same table and don't need to fill them again, we placed the definition of the internal tables in the global part of the end routine. In the global we also write:
    DATA: lv_data_loaded TYPE C LENGTH 1.
    And in the method we write:
    IF lv_data_loaded IS INITIAL.
      lv_0bpartner_loaded = 'X'.
    * load all internal tables
    lv_data_loaded = 'Y'.
    WHILE lv_0bpartner_loaded NE 'Y'.
      Call FUNCTION 'ENQUEUE_SLEEP'
      EXPORTING
         seconds = 1.
    ENDWHILE.
    LOOP AT RESULT_PACKAGE
    * assign all data
    ENDLOOP.
    This makes sure that another data package that already started, "sleeps" until the first data package is done with filling the internal tables.
    Well this all seems to work: it takes now 10 minutes to load everything to DSO Y. But I'm wondering if I'm missing anything. The system seems to work fine loading all these records in internal tables. But any improvements or critic remarks are very welcome.

    This is a great question, and you've clearly done a good job of investigating this, but there are some additional things you should look at and perhaps a few things you have missed.
    Zephania Wilder wrote:
    At first, we designed this with the standard "master data reads", but this resulted in load times of 4 hours, because all master data is read with single lookups.
    This is not accurate. After SP14, BW does a prefetch and buffers the master data values used in the lookup. Note [1092539|https://service.sap.com/sap/support/notes/1092539] discusses this in detail. The important thing, and most likely the reason you are probably seeing individual master data lookups on the DB, is that you must manually maintain the MD_LOOKUP_MAX_BUFFER_SIZE parameter to be larger than the number of lines of master data (from all characteristics used in lookups) that will be read. If you are seeing one select statement per line, then something is going wrong.
    You might want to go back and test with master data lookups using this setting and see how fast it goes. If memory serves, the BW master data lookup uses an approach very similar to your second example (1,5 hrs), though I think that it first loops through the source package and extracts the lists of required master data keys, which is probably faster than your statement "FOR ALL ENTRIES IN RESULT_PACKAGE" if RESULT_PACKAGE contains very many duplicate keys.
    I'm guessing you'll get down to at least the 1,5 hrs that you saw in your second example, but it is possible that it will get down quite a bit further.
    Zephania Wilder wrote:
    This makes sure that another data package that already started, "sleeps" until the first data package is done with filling the internal tables.
    This sleeping approach is not necessary as only one data package will be running at a time in any given process. I believe that the "global" internal table is not be shared between parallel processes, so if your DTP is running with three parallel processes, then this table will just get filled three times. Within a process, all data packages are processed serially, so all you need to do is check whether or not it has already been filled. Or are you are doing something additional to export the filled lookup table into a shared memory location?
    Actually, you have your global data defined with the statement "DATA: lv_data_loaded TYPE C LENGTH 1.". I'm not completely sure, but I don't think that this data will persist from one data package to the next. Data defined in the global section using "DATA" is global to the package start, end, and field routines, but I believe it is discarded between packages. I think you need to use "CLASS-DATA: lv_data_loaded TYPE C LENGTH 1." to get the variables to persist between packages. Have you checked in the debugger that you are really only filling the table once per request and not once per package in your current setup? << This is incorrect - see next posting for correction.
    Otherwise the third approach is fine as long as you are comfortable managing your process memory allocations and you know the maximum size that your master data tables can have. On the other hand, if your master data tables grow regularly, then you are eventually going to run out of memory and start seeing dumps.
    Hopefully that helps out a little bit. This was a great question. If I'm off-base with my assumptions above and you can provide more information, I would be really interested in looking at it further.
    Edited by: Ethan Jewett on Feb 13, 2011 1:47 PM

  • Error while deleting master data

    Hello all,
    i require some help.
    i am unable to delete some record from the master data it gives me a message "y table for the characteristic does not exist". the master data object does not contain any navigational and time depended attributes.
    the BW system is upgraded to 7.0.

    hi shakir,
    if u use this links, u can get some idea
    Q table - Time dependent master data attributes
    Y table - Interface between master data SIDs and time dependent navigational attributes SIDs (Q is linked to the Y table)
    http://help.sap.com/saphelp_nw04/helpdata/en/80/1a6365e07211d2acb80000e829fbfe/content.htm
    http://help.sap.com/saphelp_nw04/helpdata/en/2e/caceae8dd08e48a63c2da60c8ffa5e/content.htm
    http://help.sap.com/saphelp_nw04/helpdata/en/13/c3e7379899d06fe10000009b38f8cf/content.htm
    with hopes,
    Raja Singh

  • Duplicate records found while loading master data(very urgent)

    Hi all,
    One infopackage in the process chain failed while laoding the master data(full update).Its showing the following error-->duplicate record found ..1 record used in /BI0/PTCTQUERY and the same record occured in /BI0/PTCTQUERY tables.
    can anyone give me the solution...its very urgent...
    Thanks & Regards,
    Manjula

    Hi
    You can see the check box in the Processing tab page. Make a tick mark for the check box Ignore Duplicate Data Records indicator . When multiple data records that have the same key are transferred, the last data record in the request is updated to BI. Any other data records in the request with the same key are ignored.
    Help says that:
    To maintain consistency, ignoring duplicate data records is only possible if the data is updated serially. A serial update is when data is first updated into the PSA then, after it has been successfully written to the PSA, it is updated into the master data or text tables of the InfoObject.
    If a DataSource transfers potentially duplicate data records or if you manually set the Ignore Duplicate Data Records indicator, the PSA Only update type is automatically selected in the scheduler.
    hope it clears ur doubt, otherwise let me know.
    Regards
    Kiran

  • BI 7.0 : Error in loading text of a Master Data with a data source time dep

    I have a problem loading text from the data source(time dep) to the Master Data.
    For instance if the Data Source 0EVALGROUP1_TEXT is time dependent as there are the fields DATETO and DATEFROM, while i create the transformation these 2 fileds does not have any target fields. After i create the infopackage and execute it to load into the PSA, i have a error message : the message says "An error occurred in the source system."and "The error occurred in Extractor "; and the PSA is empty.
    Can anyone tell me if there is an OSS note to apply for this kind of problem, actually we are running under SAPKW70015 SAP_BW.
    Thanks for your help.

    Hi,
    Pl try the below
    The l_dta data target that you want to load is an InfoObject (master data or text
    table).
    You must declare this for the InfoProvider in Transaction RSD1.
    You do this on the 'Master Data/Text' tab by assigning any InfoArea to the InfoObject.
    The InfoProvider (that is, this InfoObject) is then displayed below this InfoArea in
    the InfoProvider tree.
    Rregards,
    Senthil

Maybe you are looking for