GL Account Master Data for all the inventory Accounts

Hi,
What is common & unique feature in the GL Master data of all the Inventory related accounts? Is it "POST AUTOMATICALLY" or some thing else
Thanks,
Lavanya

In addition to the 'post automatically' option, please also ensure the field status group selected is relating to 'material accounts'.

Similar Messages

  • Can I adjust all the master gains for all the keyframes at once?

    I edited a whole scene and keyframed some corrections but I wanted to change a few settings for the keyframes such as master gain. Is there anyway to fix this so I can just adjust all the master gians for all the keyframes at once? Thanks!

    If its the vignette geometry that is keyframed, and not the corrections, as written, (there is a distinction) then the color values are presumably NOT keyframed, as the keyframe only applies to that room/component of the grade, and not the whole grade. Your Primary In correction, in this case, is NOT keyframed and all the normal rules for revising corrections apply.
    Apple COLOR has a systemic definition: a change in values for a ROOM is a "correction", and the accumulation of corrections for a clips is a 'grade'.
    Keyframes only govern the behaviour of the room that they are created in, and don't apply across the board to the whole grade. The advantage is that the corrections can act independently.. the disadvantage is that the grade itself cannot be copy-and-pasted into another keyframed clip without any preparations.
    There are ways of doing that that are not in the Manual.
    I concur with your view of tracking, and you do have to do a wetware estimate to decide whether the investment is worth it... maybe a start and end keyframe is all you need, and adjust the trajectory as required.
    Wetware= that 'goo' between most people's ears, that starts right behind their eyes.
    jPo

  • How to generate test data for all the tables in oracle

    I am planning to use plsql to generate the test data in all the tables in schema, schema name is given as input parameters, min records in master table, min records in child table. data should be consistent in the columns which are used for constraints i.e. using same column value..
    planning to implement something like
    execute sp_schema_data_gen (schemaname, minrecinmstrtbl, minrecsforchildtable);
    schemaname = owner,
    minrecinmstrtbl= minimum records to insert into each parent table,
    minrecsforchildtable = minimum records to enter into each child table of a each master table;
    all_tables where owner= schemaname;
    all_tab_columns and all_constrains - where owner =schemaname;
    using dbms_random pkg.
    is anyone have better idea to do this.. is this functionality already there in oracle db?

    Ah, damorgan, data, test data, metadata and table-driven processes. Love the stuff!
    There are two approaches you can take with this. I'll mention both and then ask which
    one you think you would find most useful for your requirements.
    One approach I would call the generic bottom-up approach which is the one I think you
    are referring to.
    This system is a generic test data generator. It isn't designed to generate data for any
    particular existing table or application but is the general case solution.
    Building on damorgan's advice define the basic hierarchy: table collection, tables, data; so start at the data level.
    1. Identify/document the data types that you need to support. Start small (NUMBER, VARCHAR2, DATE) and add as you go along
    2. For each data type identify the functionality and attributes that you need. For instance for VARCHAR2
    a. min length - the minimum length to generate
    b. max length - the maximum length
    c. prefix - a prefix for the generated data; e.g. for an address field you might want a 'add1' prefix
    d. suffix - a suffix for the generated data; see prefix
    e. whether to generate NULLs
    3. For NUMBER you will probably want at least precision and scale but might want minimum and maximum values or even min/max precision,
    min/max scale.
    4. store the attribute combinations in Oracle tables
    5. build functionality for each data type that can create the range and type of data that you need. These functions should take parameters that can be used to control the attributes and the amount of data generated.
    6. At the table level you will need business rules that control how the different columns of the table relate to each other. For example, for ADDRESS information your business rule might be that ADDRESS1, CITY, STATE, ZIP are required and ADDRESS2 is optional.
    7. Add table-level processes, driven by the saved metadata, that can generate data at the record level by leveraging the data type functionality you have built previously.
    8. Then add the metadata, business rules and functionality to control the TABLE-TO-TABLE relationships; that is, the data model. You need the same DETPNO values in the SCOTT.EMP table that exist in the SCOTT.DEPT table.
    The second approach I have used more often. I would it call the top-down approach and I use
    it when test data is needed for an existing system. The main use case here is to avoid
    having to copy production data to QA, TEST or DEV environments.
    QA people want to test with data that they are familiar with: names, companies, code values.
    I've found they aren't often fond of random character strings for names of things.
    The second approach I use for mature systems where there is already plenty of data to choose from.
    It involves selecting subsets of data from each of the existing tables and saving that data in a
    set of test tables. This data can then be used for regression testing and for automated unit testing of
    existing functionality and functionality that is being developed.
    QA can use data they are already familiar with and can test the application (GUI?) interface on that
    data to see if they get the expected changes.
    For each table to be tested (e.g. DEPT) I create two test system tables. A BEFORE table and an EXPECTED table.
    1. DEPT_TEST_BEFORE
         This table has all EMP table columns and a TEST_CASE column.
         It holds EMP-image rows for each test case that show the row as it should look BEFORE the
         test for that test case is performed.
         CREATE TABLE DEPT_TEST_BEFORE
         TESTCASE NUMBER,
         DEPTNO NUMBER(2),
         DNAME VARCHAR2(14 BYTE),
         LOC VARCHAR2(13 BYTE)
    2. DEPT_TEST_EXPECTED
         This table also has all EMP table columns and a TEST_CASE column.
         It holds EMP-image rows for each test case that show the row as it should look AFTER the
         test for that test case is performed.
    Each of these tables are a mirror image of the actual application table with one new column
    added that contains a value representing the TESTCASE_NUMBER.
    To create test case #3 identify or create the DEPT records you want to use for test case #3.
    Insert these records into DEPT_TEST_BEFORE:
         INSERT INTO DEPT_TEST_BEFORE
         SELECT 3, D.* FROM DEPT D where DEPNO = 20
    Insert records for test case #3 into DEPT_TEST_EXPECTED that show the rows as they should
    look after test #3 is run. For example, if test #3 creates one new record add all the
    records fro the BEFORE data set and add a new one for the new record.
    When you want to run TESTCASE_ONE the process is basically (ignore for this illustration that
    there is a foreign key betwee DEPT and EMP):
    1. delete the records from SCOTT.DEPT that correspond to test case #3 DEPT records.
              DELETE FROM DEPT
              WHERE DEPTNO IN (SELECT DEPTNO FROM DEPT_TEST_BEFORE WHERE TESTCASE = 3);
    2. insert the test data set records for SCOTT.DEPT for test case #3.
              INSERT INTO DEPT
              SELECT DEPTNO, DNAME, LOC FROM DEPT_TEST_BEFORE WHERE TESTCASE = 3;
    3 perform the test.
    4. compare the actual results with the expected results.
         This is done by a function that compares the records in DEPT with the records
         in DEPT_TEST_EXPECTED for test #3.
         I usually store these results in yet another table or just report them out.
    5. Report out the differences.
    This second approach uses data the users (QA) are already familiar with, is scaleable and
    is easy to add new data that meets business requirements.
    It is also easy to automatically generate the necessary tables and test setup/breakdown
    using a table-driven metadata approach. Adding a new test table is as easy as calling
    a stored procedure; the procedure can generate the DDL or create the actual tables needed
    for the BEFORE and AFTER snapshots.
    The main disadvantage is that existing data will almost never cover the corner cases.
    But you can add data for these. By corner cases I mean data that defines the limits
    for a data type: a VARCHAR2(30) name field should have at least one test record that
    has a name that is 30 characters long.
    Which of these approaches makes the most sense for you?

  • From where to get "First day of the week" data for all the locales, is it present in CLDR spec 24?

    I am trying to get "First day of the week" data from CLDR spec24 but cannot find where to look for it in the spec. I need this data to calculate numeric value of "LOCAL day of the week".
    This data to implement "c" and "cc" day formats that equals numeric local day of the week.
    e.g if "First day of the week" data for a locale is 2 (Monday) , it means numeric value for local day of the week will be 1 if it is Monday that day, 2 if it is Tuesday that day and likewise.

    Hi
    If you want to week to be started with Sunday then use the following formula:
    TimestampAdd(SQL_TSI_DAY, 1-DAYOFWEEK(Date'@{var_Date}'), Date'@{var_Date}') if it's retail week(starts from Monday) then the follow below:
    TimestampAdd(SQL_TSI_DAY, 1-DAYOFWEEK(Date'@{var_Date}'), Date'@{var_Date}')
    I'm assuming var_Date is the presentation variable for prompt...
    Edited by: Kishore Guggilla on Jan 3, 2011 4:48 PM

  • GL account master data creation control based on Account group

    While creating GL account in FS00 we need to control to select P&L account or Balance sheet account. At present after selecting P&L accunt group it is allowing to check Balance sheet account. If we select account group belongs to Balance sheet but it is allowing to check P&L. How to control to select P&L or Balance based on account group.
    For example: If we select account group beongs to P&L it should allow to check P&L statement account.
    If we select account group belongs to BS it should allow to check Balance sheet account

    I don't there is any control for this at 'Account Group' level, please use userexit EXIT_SAPMF02H_001 for this.
    Regards,
    Ganesh

  • How to Export VO query data for all the columns.

    Hi All,
    I have advanced Table where i will be displaying only few of the columns from the VO query, when i export it will display only those columns which are displayed in advanced Table, so is it possible to export other columns also?
    Thanks
    Babu

    I faced this similar issue ...as per my knowledge if the widget for the underlying VO attribute is rendered then only one can export the data...a possible workaround is..u can create a similar page showing the respective columns intended to be exported and redirect to tht page and export it from there .....

  • Updation of contract data for all the follow up documents

    In crm I have created a service contract. From this I have created a sales-service order as follow up document. Then from this service order one suspension order.
    When I am opening the sales-service order in item details tab and then in contract tab we get the contract details. When i press the follow up document button and choose the suspension order the suspension order is opening and this contract details should come here also  but it is not coming. Please suggest whether this can be done by configuration or by badi. If by badi then suggest its name.

    Please not that the contract determination works fine when creating a new complaint ... but when the I am trying to create a complaint from service ticket the item level contract determination is not working fine ...
    We have configured that the contract should be determined at the item level only option `F`.
    I checked in one of the sap document that this cant be achieved by standard .
    http://help.sap.com/saphelp_crm70/helpdata/en/46/5cd7335bbd516fe10000000a114a6b/frameset.htmhttp://help.sap.com/saphelp_crm70/helpdata/en/46/5cd7335bbd516fe10000000a114a6b/frameset.htm
    Is there a way we can put this code in the BADI eg; CRM_SERVICE_CONTRACT ?? Please note that there is already an implementation of CRM_SERVICE_CONTRACT which is working fine for new service document creation .

  • In mdx how to get max date for all employees is it posible shall we use group by in mdx

    in mdx how to get max date for all employees is it posible shall we use group by in mdx
    example
    empno  ename date
    1         hari        12-01-1982
    1         hari        13-06-2000
    by using above data i want to get max data

    Hi Hari3109,
    According to your description, you want to get the max date for the employees, right?
    In your scenario, do you want to get the max date for all the employees or for each employee? In MDX, we have the Max function to achieve your requirement. You can refer to Naveen's link or the link below to see the details.
    http://www.sqldbpros.com/2013/08/get-the-max-date-from-a-cube-using-mdx/
    If this is not what you want, please provide us more information about the structure of you cube, so that we can make further analysis.
    Regards,
    Charlie Liao
    TechNet Community Support

  • What are the T-codes that contain the master data for material and vendor?

    what are the T-codes that contain the master data for material and vendor?

    hi ,
    - Display Material  tcodes...
    MM01 - Create Material
    MM02 - Change Material
    MM03 - Display Material
    MM50 - List Extendable Materials
    MMBE - Stock Overview
    MMI1 - Create Operating Supplies
    MMN1 - Create Non-Stock Material
    MMS1 - Create Service
    MMU1 - Create Non-Valuated Material
    ME51N - Create Purchase Requisition
    ME52N - Change Purchase Requisition
    ME53N - Display Purchase Requisition
    ME5A - Purchase Requisitions: List Display
    ME5J - Purchase Requisitions for Project
    ME5K - Requisitions by Account Assignment
    MELB - Purch. Transactions by Tracking No.
    ME56 - Assign Source to Purch. Requisition
    ME57 - Assign and Process Requisitions
    ME58 - Ordering: Assigned Requisitions
    ME59 - Automatic Generation of POs
    ME54 - Release Purchase Requisition
    ME55 - Collective Release of Purchase Reqs.
    ME5F - Release Reminder: Purch. Requisition
    MB21 - Create Reservation
    MB22 - Change Reservation
    MB23 - Display Reservation
    MB24 - Reservations by Material
    MB25 - Reservations by Account Assignment
    MB1C - Other Goods Receipts
    MB90 - Output Processing for Mat. Documents
    MB21 - Create Reservation
    MB22 - Change Reservation
    MB23 - Display Reservation
    MB24 - Reservations by Material
    MB25 - Reservations by Account Assignment
    MBRL - Return Delivery per Mat. Document
    MB1C - Other Goods Receipts
    MB90 - Output Processing for Mat. Documents
    MB1B - Transfer Posting
    MIBC - ABC Analysis for Cycle Counting
    MI01 - Create Physical Inventory Document
    MI02 - Change Physical Inventory Document
    MI03 - Display Physical Inventory Document
    MI31 - Batch Input: Create Phys. Inv. Doc.
    MI32 - Batch Input: Block Material
    MI33 - Batch Input: Freeze Book Inv.Balance
    MICN - Btch Inpt:Ph.Inv.Docs.for Cycle Ctng
    MIK1 - Batch Input: Ph.Inv.Doc.Vendor Cons.
    MIQ1 - Batch Input: PhInvDoc. Project Stock
    MI01 - Create Physical Inventory Document
    MI02 - Change Physical Inventory Document
    MI03 - Display Physical Inventory Document
    MI31 - Batch Input: Create Phys. Inv. Doc.
    MI32 - Batch Input: Block Material
    MI33 - Batch Input: Freeze Book Inv.Balance
    MICN - Btch Inpt:Ph.Inv.Docs.for Cycle Ctng
    MIK1 - Batch Input: Ph.Inv.Doc.Vendor Cons.
    MIQ1 - Batch Input: PhInvDoc. Project Stock
    MI01 - Create Physical Inventory Document
    MI02 - Change Physical Inventory Document
    MI03 - Display Physical Inventory Document
    MI31 - Batch Input: Create Phys. Inv. Doc.
    MI32 - Batch Input: Block Material
    MI33 - Batch Input: Freeze Book Inv.Balance
    MICN - Btch Inpt:Ph.Inv.Docs.for Cycle Ctng
    MIK1 - Batch Input: Ph.Inv.Doc.Vendor Cons.
    MIQ1 - Batch Input: PhInvDoc. Project Stock
    MI21 - Print physical inventory document
    MI04 - Enter Inventory Count with Document
    MI05 - Change Inventory Count
    MI06 - Display Inventory Count
    MI09 - Enter Inventory Count w/o Document
    MI34 - Batch Input: Enter Count
    MI35 - Batch Input: Post Zero Stock Balance
    MI38 - Batch Input: Count and Differences
    MI39 - Batch Input: Document and Count
    MI40 - Batch Input: Doc., Count and Diff.
    MI08 - Create List of Differences with Doc.
    MI10 - Create List of Differences w/o Doc.
    MI20 - Print List of Differences
    MI11 - Physical Inventory Document Recount
    MI07 - Process List of Differences
    MI37 - Batch Input: Post Differences
    for vendor..
    XKN1  Display Number Ranges (Vendor)
    XK01  Create vendor (centrally)
    XK02  Change vendor (centrally)
    XK03  Display vendor (centrally)
    XK04  Vendor Changes (Centrally)
    XK05  Block Vendor (Centrally)
    XK06 Mark vendor for deletion (centrally
    XK07  Change vendor account group
    XK11  Create Condition
    XK12  Change Condition
    XK13  Display Condition
    XK14  Create with cond. ref. (cond. list)
    XK15  Create Conditions (background job)
    reward points if useful,
    venkat.

  • Crystal reports 2011 failure to pull the master data for AP Items

    I am using businessobjects 4.0 and i am also running crystal report 2011.I am connecting to a sap erp ecc 6.0 ehp4 without any issue.The challenge comes when trying to input or pick the master data for example company code .The crystal report does not populate the company code field with master data.Any suggestions?

    Hi Venkateswaran,
    FIrst of all thanks for offering help...
    uploding screen shots, below mentioned screens are captured when I was creating a new report on CR2011, usning SAP standard table "T001", there is no parameters, it is simple report but I didnot get data, if i browse the field through Cystal report is says Data Size is 0.
    I also would like to inform you that If i use crystal report 2008 SP3 for same ERP, Same table and same fields, then i get proper data/report.

  • LSMW for GL account Master data

    Hi all,
    I have to create one lsmw tool in GL ACCOUNT master data.
    Please suggest me which method i should adopt.
    direct input, BAPI , Recording OR IDOC.
    IF bapi or recording, please tell me BAPI or transaction for GL account master data.
    Reagrds,
    Ruchika

    Hi Ruchika,
    BAPI_ACC_GL_POSTING_POST caters to your requirement.
    Regards,
    Younus
    <b>Reward Helpful Answers:-)</b>

  • Master data for the BP doesnu00B4t exist

    i have a big problem, when i create a BP in ECC, all the data is copied to the CRM but when i try to use this BP to create a quotation, the message "master data for this BP doesn't exist"
    could you help me to solve this BIG problem please
    regards

    Hi Sandeep,
              You can use  EXIT_/SAPAPO/SAPLDM_LOCATI_001 for populating the master data on the vendor location. Basically all the details while be CIFFED from R/3 to APO. If u need to add your custom fields u can make use of this exit. Check the LOCTYPE field with the value '1011' (Vendor).
    Regards,
    Siva.

  • Why files in Lightroom mobile and files in the creative cloud not the same? idea-  one place(stored date) for all application ??????

    Why files in Lightroom mobile and files in the creative cloud not the same? idea- one place(stored date) for all application ??????

    Lightroom Mobile is not a cloud storage service. You shouldn't treat it as a way of backing up your files. You are merely storing Smart Previews of your files in the cloud space, high-quality JPEGs of your files regardless of their original format on your desktop. The point is that they are there so you can continue editing them in a Lightroom-like environment even while away from your desktop/laptop computer. The files in Lightroom Mobile can only be used in Lightroom Mobile.
    This is very different from what is offered by the Creative Cloud storage, which can be used to synchronize your files between any device (that can support the individual files).
    The vast difference in how each service works and its intended use is why they are separate.

  • Looking for a specific data in all the cubes and ods

    Hi Gurus
    "i am looking for all the cubes/ods that contain a specific Controlling area(lets say 0123) and a specific 0plant (lets say plant 4567), now i can go down to every cube and ods and search for it in its contents but i have like hundereds of cubes it will take days, is there a simple way to look for some particular data in all the cubes/ods, and it tells me which cube/ods contains these plants and controlling area."
    <b>now based on this above post i got a reply that abaping can help.</b>
    "you could write an ABAP where you call for every InfoProvider function RSDRI_INFOPROV_READ_RFC like
    loop at <infoprov-table> assigning <wa>.
    call function 'RSDRI_INFOPROV_READ_RFC'
    exporting
    i_infoprov = <wa>
    tables
    i_t_sfc = i_t_rsdri_t_sfc
    i_t_range = l_t_rsdri_t_range
    e_t_rfcdata = l_t_rsdri_t_rfcdata
    exceptions
    illegal_input = 1
    illegal_input_sfc = 2
    illegal_input_sfk = 3
    illegal_input_range = 4
    illegal_input_tablesel = 5
    no_authorization = 6
    generation_error = 7
    illegal_download = 8
    illegal_tablename = 9
    illegal_resulttype = 10
    x_message = 11
    data_overflow = 12
    others = 13.
    endloop.
    i_t_sfc should contain 0PLANT and i_t_range the restriction on you plant value.
    with a describe table statement on l_t_rsdri_t_rfcdata you can get the hits.
    check test program RSDRI_INFOPROV_READ_DEMO for details
    best regards clemens "
    <b>now my question is how do  i use this code to check each and every cube in bw, it seems like it is meant to be for only one cube at a time. and what does he  mean by  "for every infoprovider function"</b>
    thanks

    THANKS

  • Smart form printing same data  in all the 10 pages for tcode va03

    Hi Experts
    with tcode va03 i am taking print preview of sales document when i go for print preview it is showing all the 10 pages
    with line items but when i take print for all 10 pages it is displaying first page data in all the 10 pages
    i had checked with basis consultant regarding print problem it is fine from their part
    what can be the reason
    Thanks & Regards
    anu

    Hi Anu,
    Welcome to SDN.
    In hope your Sales Document Print Preview is a Smartform. So, in the Smartform first page check for the next page assignement. Also, debug your Smartform for any issues with reference to the looping of data.
    Regards,
    Pranav.

Maybe you are looking for

  • G4 MDD CD/DVD not recognized

    I have a G4/867 DP MDD. The CD quit working and doesn't show up in the profile. I think it is toast. What is a seamless replacement? I also want to use WI-FI in this computer, what can I plug into the inside that will work?

  • Viewing iCloud Drive in Finder

    When I open iCloud Drive in the Finder I HAVE seen folders for the apps that can save files to iCloud Drive and files within them that I have saved there. But usually all I see in iCloud Drive is a folder "iCloud Drive Upgrade - Recovered Documents".

  • Blank category screens in the iTunes App store.

    If I select ANY category the screens are blank.  Is Apple aware of this issue and if so - when will it be fixed? I need to research travel apps before my trip next week to Hungary. I am running Windows 7 - 64 bit. iTunes 11.

  • Loadtesting/eTester recording of AMF

    hey, I need to write some loadtest scripts that uses AMF protocols (using weborb) for a RIA project that we're doing but i'm having some problems using eTester with capturing any data when i record. Does eTester support this? I've tried to use the pr

  • Idoc - port xml

    hi people, I have an internal table with 1308 entries. I divide it in internal tables with 300 entries that i send through an idoc in xml format. I run the report but seeing in we05 only i send 3 idocs when should be 5 (4 of 300 segments and 1 of 108