Volume of Line Item

Hi,
Currently, volume of line item in sales order VBAP-VOLUM is calculated based on order quantity.
However, I need the calculation based on confirmed quantity.
Any user exits or possible solutions. Thanks in advance.

Hi,
Check below User Exit.
Include - MV45AFZB
User Exit - USEREXIT_SOURCE_DETERMINATION
Else
Include - MV45AFZZ
User Exit - USEREXIT_SAVE_DOCUMENT_PREPARE
Good Luck !
Best regards,
Anupa

Similar Messages

  • Profit Center population in the Vendor and Customer Line items

    hello
    our client is asking for  getting profit center in the vendor and customer line items  where in the view FBL5n and fbl1n we are not getting the profit center populated - in the new gl i understand that there is a standard report based on the gl account.
    but our business is not satisfied with the report and expecting report at profit center level.
    Can any one suggest any way of doing this.
    regards,
    Vijay

    Dear Vijay,
    Let me provide you my view of solutioning for this. This is an enahcement that needs to be done
    1. You can get the profit center from the given vendor and customer line item at the time of posting, using an enahcement you will be able to capture it.
    2. Existing the profit center field is not populated in the BSIK,BSAK,BSID and BSAD tables
    3. Hence, in the same enhancement once you capture the profit center , you can write the code that profit center is updated in these tables also.
    4. This will help you to do the vendor line item wise selection in the FBL1N,  FBL5N profit center wise.
    Constraints of this solution:
    The only constraint remains where in the for a given document if there are multiple profit center, then the system will do the splitting profit center wise for a vendor line item, which will not populate the profit center in those tables as there is only one field available in the bsid etc.. tables.
    This basically would be the one the soltuion where in as seeen from the end user ther eis no change in the front end interface , the way they are doing always they can do.
    You need to also take care the % of document splitting means cross profit center postings /cross document splitting charactericstics postings and the volume involved in this. so that you can suggest this to your client.
    Regards,
    Bharathi.

  • General Ledger Accounting (New): Line Items 0FIGL_O14  Performace issue

    Dear Forum,
    We are facing a performance issue while loading the data to 0FIGL_O14 General Ledger Accounting (New): Line Items from  CUBE ZMMPRC01 -> ODSO 0FIGL_O14 DSO.
    Please see my requirement below for updating the data to 0FIGL_O14 DSO.
    This report is generated to display Dry Dock and Running Repair expenses for the particular Purchase orders with respective G/L's.
    1) The G/L DSO will provide us the 0DEBIT_LC and    0DEB_CRE_DC Foreign currency amount with signs (+/-) amounts and.
    2) ZMMPRC01 Cube   will provide us the 0ORDER_VALUE  (Purchse order value)and    0INVCD_AMNT Invoice  amount.
    While we are loading the data from  CUBE ZMMPRC01 -> ODSO 0FIGL_O14 DSO ,we have created nearly 19 InfoObject  level routine to derive the below mentioned fields data for MM Purchase Order related records.
    0CHRT_ACCTS    Chart of accounts
    0ITEM_NUM      Number of line item within accounting documen
    0AC_DOC_NO     Accounting document number
    0GL_ACCOUNT    G/L Account
    0COMP_CODE     Company code
    0COSTCENTER    Cost Center
    0CO_AREA       Controlling area
    0COSTELMNT     Cost Element
    0SEGMENT       Segment for Segmental Reporting
    0BUS_AREA      Business area
    0FUNC_AREA     Functional area
    0AC_DOC_NR     Document Number (General Ledger View)
    0AC_DOC_TYP    Document type
    0POST_KEY      Posting key
    0PSTNG_DATE    Posting date in the document
    0DOC_CURRCY    Document currency
    0LOC_CURTP2    Currency Type of Second Local Currency
    0CALQUART1     Quarter
    0CALYEAR       Calendar year
    For reference Please see the below logic to derive the data for PO related record.
    DATA:
          MONITOR_REC    TYPE rsmonitor.
    $$ begin of routine - insert your code only below this line        -
        ... "insert your code here
        types : begin of ty_FIGL,
                    CHRT_ACCTS type /BI0/OICHRT_ACCTS,
                    ITEM_NUM type /BI0/OIITEM_NUM,
                    AC_DOC_NO type /BI0/OIAC_DOC_NO,
                    GL_ACCOUNT type /BI0/OIGL_ACCOUNT,
                end of ty_FIGL.
        data :it_figl type STANDARD TABLE OF ty_figl,
              wa_figl type ty_figl.
        SELECT single CHRT_ACCTS
                        ITEM_NUM
                        AC_DOC_NO
                        GL_ACCOUNT from /BI0/AFIGL_O1400
                          into wa_figl
                          where DOC_NUM = SOURCE_FIELDS-DOC_NUM and
                                DOC_ITEM = SOURCE_FIELDS-DOC_ITEM and
                                /BIC/Z_PCODE = SOURCE_FIELDS-/BIC/Z_PCODE
                                and
                                /BIC/Z_VOY_NO = SOURCE_FIELDS-/BIC/Z_VOY_NO
                                and
                                FISCYEAR = SOURCE_FIELDS-FISCYEAR.
        if sy-subrc = 0.
          RESULT = wa_figl-AC_DOC_NO.
        ENDIF.
        clear wa_figl.
    Please note the same kind of logic is applied for all the above mentioned fields.
    Here is my concerns and issue.
    For the all above all routines i am referring BI0/AFIGL_O1400
    DSO and finally loading to the Same DSO(BI0/AFIGL_O1400
    The worried part is my DSO  0FIGL_O1400 is currecnly having nearly 60 Lacks records and MM cube is having nearly 55 requests which are required to update to the Above DSO for PO related PO value and Invoice amount.
    The big issue here is while uploading data from MM cube to DSO say for example if the request is having  25,000 records from this  nearly 500-600 records will be updated to DSO.
    But here it is taking huge time ( nearly 3 days for request ) for updating  these records , like this i have to pull 50 more requests from Cube to DSO as per the requirement.
    Please note as of now i haven't created any indexes on DSO to improve this loads.
    Please note am facing this issue in Production environment and need your help ASAP.
    Thanks & Regards,
    Srinivas Padugula

    Hi,
    If selecting data from 0FIGL_O14 is taking long time then you can create secondary indexes on DSO.
    0FIGL_O14 would be huge as data volume directly corresponds to data volume in BSEG.
    But for you requirement, I think what you can do is,
    1. create multiprovider on top of DSO and Cube and create Bex report to give you the fields requried from both the infoproviders, you can then use open hub or APD approach to keep the data in the staging table or direct update DSO and then load the data to the DSO
    2. Create secondary indexes on DSO so that fetching would be faster.
    3. Do the enhancment at R/3 level to fetch fields from MM during load of G/L
    Regards,
    Pravin Karkhanis.

  • How to create new line item in credit memo request in rebate settlement

    HI friends,
                 I need aimmediate help on following.
    Requirenment is for given agreement all invoices for that needs to be retrieved. In those invoices we need to check the tax code or tax city for all invoices. If all are same then do nothing.
               But if any of these 2 fields are different in all retrieved invoices, then we need to sum up the qty for all invoices having same tax code/city & like wise ned to create a credit memo request which will have one line item for each tax code/city.
    Example:
    Agreement 9999 with the relevant sales volume:
    Invoice 1: Material A, tax key X, quantity 1 kg
    Invoice 2: Material B, tax key X, quantity 2 kg
    Invioce 3: Material C, tax key Z, quantity 4 kg
    The credit memo request for agreement 9999 should look as such:
    Thanks & regards,
    Abhijeet.
    Item line 1: SETTLEMENT MATERIAL, tax key X, 3 kg
    Item line 2: SETTLEMENT MATERIAL, tax key Z, 4 kg
    Please help me . I am not getting the user exit / badi for this development.
    Please let me know if any one of you know from where to do this?

    Hi Vinod,
    The relationship with main item is stored with CRMD_ORDERADM_I- Parent.
    You need to pass the guid of main item to orderadm_i-parent. This will keep the relationship with main item.
    Thanks
    Ajay

  • Line Item Dimention Vs Hig Cardinality

    HI ALL
               I have a problem with a cube and the below are the challenges...
                       1) can we use High cardinality / line item dimension if we already used one of these(High cardinality / line item dimension)?
                       2) while loading 4lakh records in to cube, i tried optimizing the cube by segregating characteristics into different dimension and i am successful in bringing down the percentage levels but all dimension levels are almost near to 30%. I Have only 1 key figure , will it cause any problem....
             please let me know the concept....
    Thanks
    vijay

    Hi Vijay,
    High Cardinality Purpose: If data volume stored in a table is very huge, it's difficult to retrieve particular record or sets of records. For fast retrieval purpose system creates Indexes on various columns, by default SAP BW creates 'Bitmap' index on each dimension.
    If dimension doesn't contains large number of distinct values 'Bit Map Indexes' are fine, but in case distinct values are high and the dimension table is at least 20% of the size of the fact table, SAP recommends checking the 'High Cardinality' check box to create 'B-Tree' indexes. When compared to Bitmap indexes B-Tree indexes data retrial is fast, this is recommended on Oracle Data base.
    Line Item Dimension: By checking this the relational path of Fact Table -> Dim_ID -> SID will be reduced to Fact -> SID, so you must create a dimension which doesn't create more relations between included characteristics in dimension.
    So you can check Line Item dimension to make the system aware of not creating the Dim_ID path and you can check the High Cardinality to make the system to generate B-Tree indexes.
    Hope this helps.
    Regards
    -Sunil         

  • Reset Cleared Items - for huge number of line items

    Hi Everyone,
    Is there any way where we could reset cleared items for one clearing document (cleared close to 19000 line items) aside from FBRA? Perhaps another program which can be executed in the background? We are timing out due to the sheer volume of data.
    Are there any other options that we can explore?
    Thanks and regards,
    Joey

    Hi Tony,
    Unfortunately, it is not an executable program in background when we do SE38. Since it has no selection screens, we cannot setup a variant as well. In FBRA, you are only allowed to hit save as soon as you confirm that you want the clearing reset. It is during this time when we hit the time out.
    We have thought of this as well but asking the basis team to extend the time out session would probably be hard to get due to stringent approvals needed.
    Thanks for your advise.
    Regards,
    Joey

  • Line items  AND GENERIC EXTRACTION

    what does line items exactly mean in DATASOURCES
    AND WHAT ARE DELTA TYPE EXTRACTIONS IN GENERIC
    TIME STAMPING
    CALENDAR DAY
    NUMERICAL POINTER
    COULD ANY PLEASE EXPLAIN THE DIFFERNCE  AND IN WHAT SCENARIOS WE USE IT ...
    LOOKING FOR YOUR REPLY

    Hi Guru,
    Check below doc & thread for Line Item Dimension:
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/a7f2f294-0501-0010-11bb-80e0d67c3e4a
    Line Item Dimenstion
    If a field (date, progressive document number, timestamp) exists in the extract structure of a DataSource that contains values which increase monotonously over time, you can define delta capability for this DataSource. If such a delta-relevant field exists in the extract structure, such as a timestamp, the system determines the data volume transferred in the delta method by comparing the maximum value transferred with the last load with the amount of data that has since entered the system. Only the data that has newly arrived is transferred.
    To get the delta, generic delta management translates the update mode into a selection criterion. The selection of the request is enhanced with an interval for the delta-relevant field. The lower limit of the interval is known from the previous extraction. The upper limit is taken from the current value, such as the timestamp or the time of extraction. You can use security intervals to ensure that all data is taken into consideration in the extractions (The purpose of a security interval is to make the system take into consideration records that appear during the extraction process but which remain unextracted -since they have yet to be saved- during the next extraction; you have the option of adding a security interval to the upper limit/lower limit of the interval).
    After the data request was transferred to the extractor, and the data was extracted, the extractor then informs generic delta management that the pointer can be set to the upper limit of the previously returned interval.
    To have a clear idea:
    1. If delta field is Date (Record Create Date or change date), then use Upper Limit of 1 day.
    This will load Delta in BW as of yesterday. Leave Lower limit blank.
    2. If delta field is Time Stamp, then use Upper Limit of equal to 1800 Seconds (30 minutes).
    This will load Delta in BW as of 30 minutes old. Leave Lower limit blank.
    3. If delta field is a Numeric Pointer i.e. generated record # like in GLPCA table, then use
    Lower Limit. Use count 10-100. Leave upper limit blank. If value 10 is used then last 10
    records will be loaded again. If a record is created when load was running, those records
    may get lost. To prevent this situation, lower limit can be used to backup the starting
    sequence number. This may result in some records being processed more than once.
    Refer this link from help.sap.com
    http://help.sap.com/saphelp_erp2005/helpdata/en/3f/548c9ec754ee4d90188a4f108e0121/frameset.htm
    Difference between timestamp used in copa and generic data extraction?
    COPA timestamps vs Generic timestamps
    Check this doc for more info:
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/84bf4d68-0601-0010-13b5-b062adbb3e33
    Hope these helps u...
    Regards,
    KK.

  • 5 YEARS LINE ITEM DATA TO BE LOADED..(init)

    Hi
    I have set up GL Line item ODS (0FI_GL_4).
    R/3 has data for last 5 years...
    What should be the data load strategy....???
    Unless and until i load all the 5 years data, the g/l account balances will not match....
    Will it be ok if i load only open line items (Item Status = 'O')  and load trial balance through flat file to ODS..... this would reduce data volume in ODS and account balances would also tally.
    Please advise.
    Thank you.
    Best regards,
    Deepak

    Deepak,
    I am not sure what your requirement going to be in future. If you are going to load it is a good idea to load them all. Like suggested you can load period by period.
    1. Init without data transfer.
    2. Load period by period by full load with Repair option.
    3. It is a good idea to have a logical/physical partition for each years data and connect them in one multi provider if needed.
    4. If you are going to load something with some status (open status) sometime in the future you may get request for the closed status or even tracking of entries.
    My avaergare daily FI-GL load is around 5 millions, there are not too many clients have that much volume. So from my experience go through the pain once.
    Hope this helps. Award points if helpful.
    Goodluck
    Alex.

  • Converting Line item dimension to non line item dimension

    I have some BI 7.0 issues.
    For some of the existing cubes I had certain dimension as line item dimension  - I wanted to add some characteristic to these dimensions. I am able to do this change and activate the cube.
    But we noticed that the data which is a part of the line item dimension is lost. If the data set is small then we could have done a reload but in my case the volumes are huge, so reloading data is not a possible solution. We actually did the same exercise in BW 3.5 and noticed that the data is intact. We have looked into OSS notes and not found anything so far.
    Kindly help me to find a way to convert Line item dimension to non line item dimension in BI 7.0 without loosing Data.

    Hi,
    the only way to do is to generate a new cube and transfer the data from the first cube via transformation rule. A conversion within the existing cube is not possible.
    Regards,
    C. Aslan.

  • Affect of using Line Item Dimensions

    Hi Guys!!
    My Cube model involves 4 line item dimensions.
    I wanted to know if this will have any affect on the performance during loading/reporting on the cube.
    Marking as a line item would  cut down on one join during reporting. Please do let me know ur views on this!
    Thanks,
    Dennis.

    Hi,
      Line item dimension is one way good for reporting. The impact on loading needs to be considererd keeping in view the volume of records for each load.
    Also SAP recommends to use ODS objects, where possible instead of cubes for line items."High Cardinality" is used to adjust indexing for a dimension. Whichmeans there may be some performance impact I believe.
    Hope this helps you.....
    Message was edited by: AVR - IGA

  • Report from two line item

    Hi Experts,
    Opportunities (0crm_c04) InfoCube getting updated from Header and Item ODS. One of the Key figure Value (Expected Sales Volume) coming from Header ODS (It is not available in Item ODS). And two Characteristics (Product and Status) coming from Item ODS (It is not available in Header ODS).
    Now the records are maintained in two line item.
    In my query I want to see the key figure value based on Product and Status.
    I tried through multicube, but it is not working.
    Thanks advance,
    Amrit

    Is there a common key characteristic between header and item?
    You can use an infoset on the two ODS to get what you want.

  • [solved] Where did XFCE's volume control panel item go?

    I did an update recently and noticed that the volume control panel item disappeared.  So I went to add it again ... and it's not in the list!  I looked in the repos and there isn't a package for it either.  Where did it go?
    Last edited by synthead (2008-10-30 10:50:20)

    peart wrote:
    Hi synthead,
    Is this the problem you are having?
    peart wrote:
    Got it.  Change the last line of /usr/share/xfce4/panel-plugins/xfce4-mixer.desktop to :
    X-XFCE-Exec=/usr/lib/xfce4/xfce4/panel-plugins/xfce4-mixer-plugin
    I'll file a bug report.
    Yes!  It was.  Was, as in, it's fixed now.  Thanks!

  • Unable to capture field values in Master agreement  Line Items

    I am trying to get the field values from master agreement line items page. I have written the following code to get the line item details which is working fine to capture part number and quantity->
    agreementBean = doc.getRootParentIBean();
    lineItemCollection = agreementBean.getLineItems();
    collectionsize = lineItemCollection.size();
    for(i=0;i<collectionsize;i++){
         collectionbean = lineItemCollection.get(i);
         partNum = ""+collectionbean.getPartNumber() ;
         Quantity = ""+collectionbean.getQuantity();     
         throw doc.createApplicationException(null,partNum + " , " + Quantity );
    Now, my problem is that I am unable to capture the following field item values in the master agreement > line items page.
    1. Price Unit
    2. Unit Price
    3. Product Category
    4. Plant
    Please help me to find the values.
    Thankx in Advance.
    Su
    Edited by: subrataindra on Aug 10, 2010 11:11 AM

    This will return the name of the plant.
    .getExtensionField("PLANT").get().getDisplayName(session);
    Check if this returns the product category
    collectionbean.getExtensionField("ITEM_CAT").get().getDisplayName();
    This will Return the Price
    collectionbean.getExtensionField("PRICE").get().getPrice();
    (return type :BigDecimal)
    This will return the currency
    collectionbean.getExtensionField("PRICE").get().getCurrency;
    (Return type:String)
    Similarly to retrieve other fields for which there are no standard functions, use .getExtensionField("Field ID")
    Hope this helps
    Regards,
    Immanuel

  • Vendor Line Item not cleared after F110

    Hi Gurus,
    We encountered an issue. The payment run was processed last April 18 and the vendor invoices were captured during the payment run and in the aging report, these invoices didn't appear as Open Item.
    In between April 19 until August 19, these invoices are not part of the proposal list until the when they run payment on August 20. The value is zero and in the proposal list, there is a value in the Reason to Print Advice "A".
    What is the reason for this?
    Thank you.

    Hi Dianne Obviar,
    For that invoice payment delete uncompleted payment
    proposal.
    U go through this path select F4 in run date field
    There check it any un completed proposal is there and delete
    that invoice line item will come for payable
    Regards
    Surya

  • Open Item and Line Items are not matching

    Hi Gurus,
    I ran the EFT for vendor #1600109 for payment on 10/18/07.  If I run the vendor open items report (ZFI019) for today, it shows the balance net of the EFT that is $203,466 ( $263869 - EFT Value is $60,403). 
    When I ran the vendor line items display (FBL1N), it shows all the invoices including the invoices paid on the Oct 18 EFT $263,869 which is (2034,66+60,403). 
    User wants the open item report should tally the line item display.
    I will assign good points.
    Thanks for your help
    Prashanth

    Hi Gurus,
    EFT is working fine. But the Vendor Open item Report (Custom Report) and Vendor line item report (FBL1N) is different.
    We are looking a balance for a vendor as on 15th where the balances are not matching between the reports but if i run FBL1N on 18th the balances are matching.
    Clearning date is 18th
    Posting Date is 15th.
    Thanks
    Prashanth

Maybe you are looking for