Relation between Company and Personnel Area

Hello friends,
Which table i can find the relation between companies and personnel areas ?
I found T001P, but this table has only the relations between personnel area and personeel subarea...
Thanks for your help.

Hi,
Personnel area - company code assignment : V_T500P
Company - company code assignment : V_001_Y
Regards,
Dilek

Similar Messages

  • Urgent :Relation between vendor and business area

    < MODERATOR:  Message locked.  Please read the [Rules of Engagement|https://www.sdn.sap.com/irj/sdn/wiki?path=/display/home/rulesofEngagement] before posting next time. >
    Hi ,
    Is there any table or transaction which will give an one-to-one relation between vendor and business area?
    Thnks and Regards,
    Anand

    hi
    in this situation u have to creat new  Z table and Z transaction with link of this both . it will solve ur problem . better to take abaper help other wise u can use T-CODE se11 to creat table.
    se93 for transaction code .

  • Relation between Responsibility and Business Area

    Hi,
    I have a requirement wherein i have to make a particular workbook available in specific responsibilities.
    Now when i have to create a folder then how do i know in which Business Area should i create the folder.
    Basically i need to know the relation between Responsibility and Business Area i.e which Business Area i should create the folder so that it is visible in desired responsibilities.
    Is there any way possible i can know this.
    Regards,
    Shruti

    Hi,
    You can do that in two ways:
    1. Go to the administrator and check the sharing BA for each responsibility you require.
    2. Query the database to retrieve the sharing.
    Do whatever you find easier.
    For the database option you can run the following (change it to your own schema name):
    select distinct
    t.ba_name,
    t.ba_id,
    t.ba_description,
    case when substr(t.ba_created_by,1,1)='#' then fu.user_name else t.ba_created_by end as CREATED_BY,
    disco_users.eu_username,
    case when instr(disco_users.eu_username,'#')=0 then disco_users.eu_username
    when instr(disco_users.eu_username,'#')>0 and instr(disco_users.eu_username,'#',2)=0 then (select fu.user_name from fnd_user fu where fu.user_id=substr(disco_users.eu_username,2,5))
    else (select resp.responsibility_name from fnd_responsibility_tl resp where resp.responsibility_id=substr(disco_users.eu_username,2,5))
    end as "Shared Name / Responsibility"
    from eul_us.eul5_bas t,
    apps.fnd_user fu,
    eul_us.eul5_access_privs disco_shares,
    eul_us.eul5_eul_users disco_users
    where substr(t.ba_created_by,2,10)=to_char(fu.user_id(+))
    and t.ba_id=disco_shares.gba_ba_id
    and disco_users.eu_username(+) NOT IN ('EUL5', 'PUBLIC')
    AND disco_users.eu_id(+) = disco_shares.ap_eu_id
    Hope it helps
    Tamir

  • Relation Between Company code Business Area,Cost centre etc

    Hi Experts,
    Can anybody tell me How the following are related to each other the same is mentioned below.
    1.Company Code
    2.Business Area
    3.Profit Centre
    4.Cost Centre
    5.Cost elements
    AR

    Hi,
    Follow the Link, I think it may help you
    [http://wiki.sdn.sap.com/wiki/display/ERPFI/BuldingBlocks(Masterdata)ReleventtoControlling]

  • Database table/FM to find relation between plant and company code in SRM

    Hi,
    I have a requirement where I need to determine the relation between the plant and company code in SRM.
    The plant and company code both belong to R/3. We are replicating the plant by using the report BBP_LOCATIONS_GET_ALL.
    This plant is now available as a BP in SRM. But from which SRM table or using which SRM FM can I get the relation between this and the company code?
    Thanks,
    Srivatsan

    Hi
    <b>Check table BBP_LOCMAP. (Plants) -  Manage Table Business Partner   > System   > Location
    BBP_COMPCODE_FAV               User-Specific Favorites for Permitted Company code.</b>
    Hope this will help.
    Please reward suitable points, incase it suits your requirements.
    Regards
    - Atul

  • Where can I find the relation between company code and plant?

    where can I find the relation between company code and plant?
    I need to the list of plants under a company code.
    Which table?

    yes,wayne weng   .What you said is right.
    Thank you very much!
    My MSN:[email protected]

  • Table that gives the relation between plant and company code

    Hi gurus,
    I actually have plant number and using this plant value i need to get the company code.
    Is there any table that gives the relation between plant and company code. So,that i can get the company code details.
    Thanks in advance.

    Hi Bhanuphani,
    Use  T001K  where BWKEY is the plant
    Reward if useful
    Thanks Arjun

  • What are the Relations between Journalizing and IKM?

    What is the best method to use in the following scenario:
    I have about 20 source tables with large amount of data.
    I need to create interfaces that join the source tables into target tables.
    The source tables are inserted every few secondes with about hundreds to thousands rows.
    There can be a gap of few seconds between the insert of different tables that sould be joined.
    The source and target tables are on the same Oracle instance and schema.
    I want to understand the role of: 'Journalizing CDC' and 'IKM - Incremental Update' and
    how can i use it in my scenario?
    In general What are the relations between 'Journalizing' and 'IKM'?
    Should i use both of them? Or maybe it is better to deelte and insert to the target tables?
    I want to understand what is the role of 'Journalizing CDC'?
    Can 'IKM - Incremental Update' work without 'Journalizing'?
    Does 'Journalizing' need to have PK on the tables?
    What should i do if i can't put PK (there can be multiple identical rows)?
    Thanks in advance Yael

    Hi Yael,
    I will try and answer as many of your points as I can in one post :-)
    Journalizing is way of tracking only changed data in your source system, if your source tables had a date_modified you could always use this as a filter when scanning for changes rather than CDC, Log based CDC (Asynchronous in ODI, Logminer/Streams or Goldengate for example) removes the overhead of of placing a trigger on the source table to track changes but be aware that it doesnt fully remove the need to scan the source tables, in answer to you question about Primary keys, Oracle CDC with ODI will create an unconditional log group on the columns that you have defined in ODI as your PK, the PK columns are tracked by the database and presented in a Journal table (J$<source_table_name>) this Journal table is joined back to source table via a journalizing view (JV$<source_table_name>) to get the rest of the row (ie none PK columns) - So be aware that when ODI comes around to get all data in the Journalizing view (ie Inserts, Updates and Deletes) the source database performs a join back to the source table. You can negate this by specifying ALL source table columns in your PK in ODI - This forces all columns into the unconditional log group, the journal table etc. - You will need to tweak the JKM to then change the syntax sent to the database when starting the journal - I have done this in the past, using a flexfield in the datastore to toggle 'Full Column' / 'Primary Key Cols' to go into the JKM set up (there are a few Ebusiness suite tables with no primary key so we had to do this) - The only problem with this approach is that with no PK , you need to make sure you only get the 'last' update and in the right order to apply to your target tables, without so , you might process the update before the insert for example, and be out of sync.
    So JKM's provide a mechanism for 'Change data only' to be provided to ODI, if you want to handle deletes in your source table CDC is usefull (otherwise you dont capture the delete with a normal LKM / IKM set up)
    IKM Incremental update can be used with or without JKM's, its for integrating data into your target table, typically it will do a NOT EXISTS or a Minus when loading the integration table (I$<target_table_name>) to ensure you only get 'Changed' rows on the load into the target.
    user604062 wrote:
    I want to understand the role of: 'Journalizing CDC' and 'IKM - Incremental Update' and
    how can i use it in my scenario?Hopefully I have explained it above, its the type of thing you really need to play around with, and throroughly review the operator logs to see what is actually going on (I think this is a very good guide to setting it up : http://soainfrastructure.blogspot.ie/2009/02/setting-up-oracle-data-integrator-odi.html)
    In general What are the relations between 'Journalizing' and 'IKM'?JKM simply presents (only) changed data to ODI, it removes the need for you to decide 'how' to get the updates and removes the need for costly scans on the source table (full source to target table comparisons, scanning for updates based on last update date etc)
    Should i use both of them? Or maybe it is better to deelte and insert to the target tables?Delete and insert into target is fine , but ask yourself how do you identify which rows to process, inserts and updates are generally OK , to spot a delete you need to compare the table in full, target table minus source table = deleted rows , do you want to copy the whole source table every time to perform this ? Are they in the same database ?
    I want to understand what is the role of 'Journalizing CDC'?Its the ODI mechanism for configuring, starting, stopping the change data capture process in the source systems , there are different KM's for seperate technologies and a few to choose for Oracle (Triggers (Synchronous), Streams / Logminer (Asynchronous), Goldengate etc)
    Can 'IKM - Incremental Update' work without 'Journalizing'?Yes of course, Without CDC your process would look something like :
    Source target ----< LKM >---- Collection table (C$) ----<IKM>---- Integration table (I$) -----< IKM >---- Target table
    With CDC your process looks like :
    Source Journal (J$ table with JV$ view) ----< LKM >---- Collection table (C$) ----<IKM>---- Integration table (I$) -----< IKM >---- Target table
    as you can see its the same process after the source table (there is an option in the interface to enable the J$ source , the IKM step changes with CDC as you can use 'Synchronise Journal Deletes'
    Does 'Journalizing' need to have PK on the tables?Yes - at least a logical PK in the datastore, see my reply at the top for reasons why (Log Groups, joining back the J$ table to the source table etc)
    What should i do if i can't put PK (there can be multiple identical rows)? Either talk to the source system people about adding one, or be prepared to change the JKM (and maybe LKM, IKM's) , you can try putting all columns in the PK in ODI. Ask yourself this , if you have 10 identical rows in your source and target tables, and one row gets updated - how can you identify which row in the target table to update ?
    >
    Thanks in advance YaelA lot to take in, as I advised I would reccomend you get a little test area set up and also read the Oracle database documentation on CDC as it covers a lot of the theory that ODI is simply implementing.
    Hope this helps!
    Alastair

  • What are the relation between JPA and Hibernate, JPA and TopLink?

    What are the relation between JPA and Hibernate, JPA and TopLink?
    Can JPA instead of Hibernate and TopLink?

    The Java Persistence API (JPA) is the object relational mapping persistence
    standard for Java. Hibernate and TopLink provide an Open source Object-relational mapping framework for Java.
    They provide an implementation for the Java Persistence API. In my opinion, both Hibernate and TopLink provide support to JPA
    and they can also be regarded as the complementary to JPA.
    Let's wait to see other person's opinions.

  • Relation between PS and FI func areas

    Hi
    I am doing now as a BI consultant,I am planning to learn PS func area.
    Can any one say the relation between PS and FI func areas
    Cheers
    Sunil Reddy LCP

    Hi,
    Please follow the details which depicts about integration of PS module with FI/CO in nutshell.
    1. Costing and Budgeting of the Project
    • Cost planning in the work breakdown structure • Network costing • Budgeting and availability control
    • Commitment and actual costs through account assignment and order assignment
    2. Revenue and Payments
    • Revenue and payment planning in the project
    3. Period-End Closing
    • Interest calculation
    • Incoming orders
    • Settlement
    • Result Analysis
    To know more details of the above, please refer the below URL
    http://help.sap.com/erp2005_ehp_02/helpdata/en/86/98853478616434e10000009b38f83b/frameset.htm
    Hope this will help you.
    Regards,
    Rakesh Pradhan

  • What's the relation between Gain and transmitted power with coverage area?

    what's the relation between Gain and transmitted power with coverage area?
    if we want add coverage area, could we just add antenna's gain only or add transmitted power only?

    That information can be found at: http://www.cisco.com/warp/customer/cc/pd/witc/ao350ap/prodlit/agder_rg.htm

  • How to make relation between  gl_je_lines  and wip_transaction_accounts

    I try to write script to make relation between GL and WIP, wip_transaction_accounts and mtl_material_transactions
    1.
    SELECT
    (SELECT meaning
    FROM mfg_lookups
    WHERE lookup_type = 'WIP_TRANSACTION_TYPE'
    AND lookup_code =
    (SELECT TRANSACTION_TYPE
    FROM wip_transactions
    WHERE transaction_id = wta.transaction_id
    ) "Transaction Type" ,
    gjb.NAME "Journal Batch Name" ,
    gjh.NAME "Journal Name" ,
    gjh.je_source "JE Source" ,
    gjh.je_category "JE Category" ,
    glp.period_num "GL Month" ,
    glp.period_year "GL Year" ,
    gjh.default_effective_date "GL Date" ,
    NVL(gje.accounted_dr,0) -NVL( gje.accounted_cr,0) "GL Line Amount" ,
    gje.description "GL Line Description" ,
    (SELECT organization_code
    FROM mtl_parameters
    WHERE organization_id = wta.organization_id
    ) "ORG Name" ,
    NULL "Vendor/Customer Name" ,
    NULL "SO/PO Number" ,
    NULL "Reference Number" ,
    NULL "AP/AR Invoice Number" ,
    NULL "Doc Sequence Value" ,
    NULL "Invoice Type Lookup Code" ,
    NULL "Check Number" ,
    NULL "Line Type" ,
    NULL "Category" ,
    NVL (wta.base_transaction_value, 0) "Transaction Amt" ,
    NULL "AR/PO Receipt Number" ,
    NULL "Applied Invoice Number" ,
    (SELECT a.segment1
    FROM mtl_system_items_b a ,
    wip_discrete_jobs b
    WHERE a.inventory_item_id = b.primary_item_id
    AND a.organization_id = b.organization_id
    AND b.wip_entity_id = wta.wip_entity_id
    AND b.organization_id = wta.organization_id
    ) "Item Name" ,
    gcc.segment1 "Company" ,
    gcc.segment2 "Department" ,
    gcc.segment3 "Account" ,
    (SELECT description
    FROM fnd_flex_values_vl fnd
    WHERE flex_value_set_id = 1009707
    AND TO_CHAR (fnd.flex_value) = gcc.segment3
    ) "Account Description" ,
    gcc.segment4 "Intercompany" ,
    NULL "JE Category Description" ,
    NULL "AP Invoice Line Description"
    TRUNC (wta.transaction_date) "JE Creation/Inv Trans Date" ,
    NULL "JE Created By" ,
    NULL "Reversal Flag" ,
    NULL "Reason Code" ,
    NULL "Subinventory Name" ,
    wta.primary_quantity "Quantity" ,
    NVL (wta.base_transaction_value, 0) "Value" ,
    gje.context ,
    gje.attribute1 ,
    gje.attribute2 ,
    gje.attribute3 ,
    gje.attribute4 ,
    gje.attribute5 ,
    DECODE(gjh.status,'P','Posted','U','Unposted','Error') "Post Status"
    FROM gl_je_lines gje ,
    gl_je_headers gjh ,
    gl_je_batches gjb ,
    wip_transaction_accounts wta ,
    gl_code_combinations gcc ,
    gl_periods glp
    WHERE 1 =1
    AND gje.je_header_id = gjh.je_header_id
    AND gjh.je_batch_id = gjb.je_batch_id
    AND UPPER (gjh.je_source) = 'INVENTORY'
    AND gcc.code_combination_id = gje.code_combination_id
    AND wta.reference_account = gje.code_combination_id
    AND TRUNC(wta.transaction_date) BETWEEN glp.START_DATE AND glp.END_DATE
    AND gje.reference_1 = wta.gl_batch_id
    AND gje.gl_sl_link_table = 'WTA'
    AND glp.period_name = gjh.period_name
    AND glp.period_set_name = 'OVT_US_CAL'
    ==
    2.
    SELECT mtt.transaction_type_name "Transaction Type" ,to_char(mmt.transaction_date,'mm/dd/yyyy hh:mi:ss') as teas, mta.transaction_id,mta.organization_id, mmt.organization_id,
    gjb.NAME "Journal Batch Name" ,
    gjh.NAME "Journal Name" ,
    gjh.je_source "JE Source" ,
    gjh.je_category "JE Category" ,
    glp.period_num "GL Month" ,
    glp.period_year "GL Year" ,
    gjh.default_effective_date "GL Date" ,
    NVL(gje.accounted_dr,0) - NVL(gje.accounted_cr,0) "GL Line Amount" ,
    gje.description "GL Line Description" ,
    (SELECT organization_code
    FROM mtl_parameters
    WHERE organization_id = mta.organization_id
    ) "ORG Name" ,
    NULL "Vendor/Customer Name" ,
    NULL "SO/PO Number" ,
    mmt.TRANSACTION_REFERENCE "Reference Number" ,
    NULL "AP/AR Invoice Number" ,
    NULL "Doc Sequence Value" ,
    NULL "Invoice Type Lookup Code" ,
    NULL "Check Number" ,
    NULL "Line Type" ,
    NULL "Category" ,
    NVL (mta.base_transaction_value, 0) "Transaction Amt" ,
    NULL "AR/PO Receipt Number" ,
    NULL "Applied Invoice Number" ,
    (SELECT segment1
    FROM mtl_system_items_b
    WHERE inventory_item_id = mmt.inventory_item_id
    AND organization_id = mmt.organization_id
    ) "Item Name" ,
    gcc.segment1 "Company" ,
    gcc.segment2 "Department" ,
    gcc.segment3 "Account" ,
    (SELECT description
    FROM fnd_flex_values_vl fnd
    WHERE flex_value_set_id = 1009707
    AND TO_CHAR (fnd.flex_value) = gcc.segment3
    ) "Account Description" ,
    gcc.segment4 "Intercompany" ,
    NULL "JE Category Description" ,
    NULL "AP Invoice Line Description"
    TRUNC (mta.transaction_date) "JE Creation/Inv Trans Date" ,
    NULL "JE Created By" ,
    NULL "Reversal Flag" ,
    (SELECT MGD.SEGMENT1
    FROM MTL_GENERIC_DISPOSITIONS MGD,
    MTL_MATERIAL_TRANSACTIONS MMTT
    WHERE MGD.DISPOSITION_ID = MMTT.TRANSACTION_SOURCE_ID
    AND MMTT.TRANSACTION_SOURCE_TYPE_ID = 6
    AND MGD.ORGANIZATION_ID = MMTT.ORGANIZATION_ID
    AND MMTT.TRANSACTION_ID = MMT.TRANSACTION_ID
    ) "Reason Code" ,
    mmt.SUBINVENTORY_CODE "Subinventory Name" ,
    mta.primary_quantity "Quantity" ,
    NVL (mta.base_transaction_value, 0) "Value" ,
    gje.context ,
    gje.attribute1 ,
    gje.attribute2 ,
    gje.attribute3 ,
    gje.attribute4 ,
    gje.attribute5 ,
    DECODE(gjh.status,'P','Posted','U','Unposted','Error') "Post Status" -
    FROM gl_je_lines gje ,
    gl_je_headers gjh ,
    gl_je_batches gjb ,
    mtl_transaction_accounts mta ,
    mtl_material_transactions mmt ,
    mtl_transaction_types mtt ,
    gl_code_combinations gcc ,
    gl_periods glp
    WHERE 1 =1
    AND mta.transaction_id = mmt.transaction_id
    AND gje.je_header_id = gjh.je_header_id
    AND gjh.je_batch_id = gjb.je_batch_id
    AND UPPER (gjh.je_source) = 'INVENTORY'
    AND gcc.code_combination_id = gje.code_combination_id
    AND mta.reference_account = gje.code_combination_id
    AND TRUNC(mta.transaction_date) BETWEEN glp.START_DATE AND glp.END_DATE
    and mta.request_id=mmt.request_id
    and mta.inventory_item_id=mmt.inventory_item_id
    AND gje.reference_1 = mta.gl_batch_id
    AND gje.gl_sl_link_table = 'MTA'
    AND mtt.transaction_type_id = mmt.transaction_type_id
    AND glp.period_name = gjh.period_name
    AND glp.period_set_name = 'OVT_US_CAL'
    When gl_je_lines have multiple line this script on the same transaction_id it will get wrong information. By setup all wip_transaction_accounts.GL_SL_Lind_id and mtl_transaction_accounts.GL_SL_Lind_id is null
    Anyone can help?

    Hello.
    How are you reaching the inconsistencies ? Are you comparing report's results?
    Octavio

  • Relation between idoc and EDI,ALE

    Hi,
           I am new to IDOC.  Can u please tell me whether there is any relation between IDOC and ALE or IDOC and EDI. If so what is the relation between them.

    Hi Buvana,
    ALE is a means of migrating DATA between SAP systems. You configure logical destinations, establish some migrating criteria and the system takes care of getting your data from point A to point B.
    EDI ->
    EDI is described as the interchange of structured data according to agreed message standards between computer systems, by electronic means. Structured data equates to a simple and direct method of presenting the data content of a document. The method of ensuring the correct interpretation of the information by the computer system is defined by the EDI standard."
    EDI is a technique used to communicate business transactions between computer systems of different companies and organizations. Note that sometimes the EDI mechanism deployed at a company is often used to interface to other systems within the same organization."
    Info shuttle - "I was implementing EDI at a customer and I wanted some real data to test with in Development, so I asked the Project Manager to see if he could arrange something. Literally, 20 minutes later he came back and said that there were 10 new orders in Development, together with the customer master records related to the order, the carrier vendor master record, the material master records, ... were all there. I asked him how he had done it so quickly which led me to my first exposure with Infoshuttle, a tool that does this for you using the ALE functionality. As an ALE consultant I know how to do this stuff but I also know how long it would take to set that up.
    www.sapgenie.com/sapedi/edi_sap_training.htm
    www.sap-img.com/basis/ difference-between-edi-and-idoc.htm
    www.help.sap.com/saphelp_nw04/helpdata/ en/35/26b592afab52b9e10000009b38f974/content.htm
    www.help.sap.com/saphelp_nw04/helpdata/ en/35/26b594afab52b9e10000009b38f974/content.htm
    http://www.onestopsap.com/interview-Question/edi/
    IDocs act as data containers in an ALE scenario. They are just a format of storing application data and then transferred using ALE. Take a look at http://www.sapgenie.com/ale/why_ale.htm and http://www.sapgenie.com/sapgenie/docs/ale_whitepaper.doc for further details on ALE.
    Check this link.
    http://www.sapgenie.com/sapedi/idoc_abap.htm
    http://www.henrikfrank.dk/abapexamples/IDOC/IDOC.htm
    http://help.sap.com/printdocu/core/Print46c/en/Data/Index_en.htm
    http://help.sap.com/saphelp_47x200/helpdata/en/0b/2a6095507d11d18ee90000e8366fc2/frameset.htm
    Regards,
    Priyanka.

  • Relation between plant and shipping point

    Hi all,
    my simple question is,  what is the relation between plant and shipping point ?
    in book i read, it is                      many to many,
    but in some web link it is showing 1 to many
    what is the logic behind these two ans.
    pls tell me the ans...
    Regards
    Avinash

    Hi Hegal . K . Charles,
    it is correct,  as we discussed above it may be MANY to MANY .
    Suppose, we hv 2 plant 1 in delhi and 2nd in nagpur. and this company want to export via inland water ways, in this case we can assign mumbai port to delhi and nagpur plant as well.
    as you said it is correct, It is just  not to overload the master data maintenance 
    Ex- In case of Sales Org. Its is always advisable to keep it minimum to per company code. Because master data records are multiplied by each additional org. element.

  • Relation between Rollup and compression.

    Hi All,
    Is there a relation between rollup and compression ? i.e if i compress the cube, will the rollup job be faster ?

    Hi,
    Thanks for all your replies. Now the picture is crystal clear.
    1. A compression job moves data from F table to E table. This is also applicable for aggregates. If you schedule a compression job for a cube, the aggregates are automatically compressed( You can also compress aggregates of a cube without compressing the cube. There are programs avalibale for the same).
    2. Assume that you have not done compression . If you load the data to base cube ,it takes more time to create the indexes after loading.
    3) Assume that you have done compression . If you load the data to base cube ,it takes less time than 2 point to create the indexes after loading.
    On similar lines, if aggregates of a cube  are compressed before rollup, then rollup will be faster. This is what exactly I have experienced. I had a cube which had 1000 request. It was never compressed. The rollup job for the cube used to take around 60,000 seconds. I compressed 900 requests. Now the rollup job gets over in 2400 seconds.So the conclusion is the following:
    <b>When aggregates of a cube are compressed, the rollup job runs faster.</b>
    Message was edited by: Tej Trivedi

Maybe you are looking for