AP check set-up table info in R12

Hello all;
Does anybody knows where the information for AP check formats:
From the Apps Menu
Payables Manager>Payment Administrator (Under Setup : Payment)> Goto Task: Formats>
That particular screen information in what TABLE is stored (IBY_?)
or what API handle the information for that screen.
I got a duplicate records which I need to delete them.
Thank you.
Shaun S.

check it out at
http://www.oracleappshub.com/release12/know-the-changes-because-of-r12-oracle-payment-module-fund-disbursement-in-ebs/

Similar Messages

  • How to check whether set up tables are filled or not

    Hi Gurus,
    How to check whehter set up tables are filled or not?
    Cheers,
    Reddy.

    Hi Reddy.......
    And another point I want to make u clear............LBWQ is not the Delta Queue.............Delta queue is RSA7...............it comes into play in delta loading.......LBWQ is Extraction Queue.........While delta loading....if we use the Update mode Queued Delta........then any Changes will not directly get updated in the Delta queue.....For the Delta set up.........
    1) First u hav to run init with data transfer(After filling the set up table)
    2) After that any Changed ........will be recorded in the Extraction Queue(LBWQ)........
    3) After a certain number of records get accumulated in LBWQ........We hav to run V3 jobs to bring the data in the Delta queue(RSA7)...
    Check this.....
    Re: question on LBWQ and RSA7
    Regards,
    Debjani.....

  • RE: Authority checks included in the info set of the query

    Hi all,
    I am checking the program code for one of our custom tcodes and i asked ABAP team to add authority check to the program code because there is no auth check in the code and abapers told me that the authority check is included inside the info set of the query and not in the program . the program is used to execute the query in the Tcode.
    how to find the Authority checks included in the info set of the query.
    Thanks in advance,
    Sun.

    If you have the BI support roles assigned to you  and the security admin  roles please login to the BI system
    execute transaction RSECADMIN, click on the analysis tab and execute as the user who is assigned the role with restrictions.
    For variables in authorizations like ( type customer exit )
    use RSECADMIN - maintain authorization tab - Click on value authorization tab.
    Keytransaction is RSECADMIN  & infoobject maintenance details you can get from RSD1.
    Regards

  • Not getting set up record in Set up tables

    Hi ,
    I am able to see revaluation records in MB5B (Stock on posting date) for one perticular posting date. but this record i am not able to see in Bw side ,even its not avail in RSA3 level.
    So wat cud be the reason its avail in data base table but not filled in Set up tables ?
    regards
    raja

    Hi vishwanath,
    This is bcoz of 
    U have not done the delte delta queue( In LBWQ "MCEX03" entries).
    first U have to delete the MCEX03 Entries ---> then only u have to do fill the setup tables.
    follow these steps:
    1.First goto LBWQ --> delete the MCEX03 Entries.
    2.Refresh .
    3.Check the datasources in RSA3 ( Datasources Should show Zero Records )
    4.Do the Fill setup table.
    5.Then check once again LBWQ ---> It should not have MCEX03 Entries.
    6. Check Datasources in RSA3 ( Datasources should have data).
    7.Do replication & Extarct data.
    for ur problem this is the solution ( I have faced this).
    Thanks,
    kiran.

  • Error while filling the set up table for 2LIS_11_VAITM

    Dear Experts,
    i am facing an error while filling the set up table for 2LIS_11_VAITM in the source system.
    The error was "Company code for sales org 9000 does not exist (document 1326)".
    when i checked the document in the Header table VBAK,company code was blank for the document .
    Company code was there in the selections for the setup table.
    The error means that company code should be maintained for the sales document as per my knolwedge.i think document was not created properly in the source system.
    Could you please help me out to over come the problem.
    Regards,
    Sunil...

    in the selection screen for filling the setup table, there is a parameter No. tolerated faulty documents
    you'll have to enter at least '1' here in your case
    it might be that there are other erronuous records, in that case it will crash again, and you'll have to increase the value for this parameter
    (under "normal" circumstances, all documents in the source should be correct... but apparently it's possible to circumvent this)

  • Error while filling set up table SD-Billing Documents - Perform Setup

    Hello friends,
    while filling set up table of SD-Billing Documents - Perform Setup
    i am getting error 'TSV_TNEW_PAGE_ALLOC_FAILED' . i am executing this job in background process . Please suggest me a solution .
    Regards
    Nilesh Vakil

    Hi,
    Try some other LO extractor in RSA3 to check if you get the same error.
    If you dont then it may not be system memory issue, check if thr is some user exit written for the extractor.The problem could be in the user exit whr the internal table may get overflown and lead to the page overfloen dump.
    If this is also not the case try reducing the packet size.
    Let me know if the issue is resolved.
    Regards,
    Ashwin

  • Error while filling up set up tables in SAP R/3

    Experts,
    I trying to fill up set up tables for application component 11 (SD) for my LO extraction. It gives me the following error
    <b>Error determining rate: foreign curr. RMB local curr. AUD date 06/19/2007 (doc. 100025787)</b>
    when I check in RSA3, I have data before 06/19/2007 but nothing is filled up after that.
    Does anyone know how to fix this?
    Thanks
    Ashwin

    Hi Ashwin,
    Did you get the solution for issue(filling the setup tables for applicaion component 11)
    "Error determining rate: foreign curr. RMB local curr. AUD date 06/19/2007 (doc. 100025787)"
    thanks in advance.
    Bhaskar.

  • JOIN ON 2 different sets of table depending on the result of first set

    <br>
    I have a query where it returns results. I want to join this query to
    2 different sets of table depending upon the first set has a result or not.
    if first set didnt had a results or records then check for the second set.
    SELECT 
    peo.email_address,
    r.segment1 requistion_num,
    to_char(l.line_num) line_num,
    v.vendor_name supplier, 
    p.CONCATENATED_SEGMENTS category,
    to_char(round((nvl(l.quantity, 0) * nvl(l.unit_price, 0))),'99,999,999,999.99'),
    TO_CHAR(l.need_by_date,'MM/DD/YYYY') need_by_date,
    pe.full_name requestor,
    l.item_description,
    pr.segment1 project_num,
    t.task_number,
    c.segment1,
    c.segment2
    FROM po_requisition_headers_all r,
         po_requisition_lines_all l,  
    (SELECT project_id,task_id,code_combination_id, distribution_id,requisition_line_id,creation_date  FROM
    (SELECT project_id,task_id,code_combination_id,distribution_id,creation_date,requisition_line_id,ROW_NUMBER ()
    OVER (PARTITION BY requisition_line_id ORDER BY requisition_line_id,distribution_id ) rn
    FROM po_req_distributions_all pod) WHERE rn = 1) d,
    gl_code_combinations c,
    POR_CATEGORY_LOV_V p,
    per_people_v7 pe,
    PA_PROJECTS_ALL pr,
    PA_TASKS_ALL_V t,
    ap_vendors_v v,
    WHERE  d.creation_date >= nvl(to_date(:DATE_LAST_CHECKED,
    'DD-MON-YYYY HH24:MI:SS'),SYSDATE-1)
    AND
    l.requisition_header_id = r.requisition_header_id
    AND l.requisition_line_id = d.requisition_line_id
    AND d.code_combination_id = c.code_combination_id
    AND r.APPS_SOURCE_CODE = 'POR'
    AND l.category_id = p.category_id
    AND r.authorization_status IN ('IN PROCESS','PRE-APPROVED','APPROVED')
    AND l.to_person_id = pe.person_id
    AND pr.project_id(+) = d.project_id
    AND t.project_id(+) = d.project_id
    AND t.task_id(+) = d.task_id
    AND v.vendor_id(+) = l.vendor_id
    and r.requisition_header_id in(
    SELECT requisition_header_id FROM po_requisition_lines_all pl                    
    GROUP BY requisition_header_id HAVING SUM(nvl(pl.quantity,0) * nvl(pl.unit_price, 0)) >=100000)
    group by
    peo.email_address,
    r.REQUISITION_HEADER_ID,
    r.segment1 ,
    to_char(l.line_num) ,
    v.vendor_name, 
    p.CONCATENATED_SEGMENTS ,
    to_char(round((nvl(l.quantity, 0) * nvl(l.unit_price, 0))),'99,999,999,999.99'),
    TO_CHAR(l.need_by_date,'MM/DD/YYYY') ,
    pe.full_name ,
    l.item_description,
    c.segment1,
    c.segment2,
    pr.segment1 ,
    t.task_number
    <b>I want to join this query with this first set </b>
    SELECT b.NAME, c.segment1 CO, c.segment2 CC,
              a.org_information2 Commodity_mgr,
              b.organization_id, p.email_address
         FROM hr_organization_information a, hr_all_organization_units b, pay_cost_allocation_keyflex c, per_people_v7 p
        WHERE a.org_information_context = 'Financial Approver Information'
          AND a.organization_id = b.organization_id
           AND b.COST_ALLOCATION_KEYFLEX_ID = c.COST_ALLOCATION_KEYFLEX_ID
           and a.ORG_INFORMATION2 = p.person_id
          AND NVL (b.date_to, SYSDATE + 1) >= SYSDATE
          AND b.date_from <= SYSDATE;
    <b>if this doesnt return any result then i need to join the query with the 2nd set</b>
    select lookup_code, meaning, v.attribute1 company, v.attribute2 cc,
                decode(v.attribute3,null,null,p1.employee_number || '-' || p1.full_name) sbu_controller,
                decode(v.attribute4,null,null,p2.employee_number || '-' || p2.full_name) commodity_mgr
                from fnd_lookup_values_vl v,
                per_people_v7 p1, per_people_v7 p2
                where lookup_type = 'BIO_FIN_APPROVER_INFO'
                  and v.attribute3 = p1.person_id(+)
                and v.attribute4 = p2.person_id(+)
                order by lookup_code
    How do i do it?
    [pre]                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                               

    <br>
    I have hard coded the 2 jon sets into one using UNION ALL but if one record exists in both sets how would i diferentiate between the 2 sets.
    COUNT(*) will only give the total records.
    if there r total 14
    suppose first set gives 12 records
    second set gives 4 records.
    But i want only 14 records which could 12 from set 1 and 2 from set 2 since set1  and set2 can have common records.
    SELECT 
    peo.email_address,
    r.segment1 requistion_num,
    to_char(l.line_num) line_num,
    v.vendor_name supplier, 
    p.CONCATENATED_SEGMENTS category,
    to_char(round((nvl(l.quantity, 0) * nvl(l.unit_price, 0))),'99,999,999,999.99'),
    TO_CHAR(l.need_by_date,'MM/DD/YYYY') need_by_date,
    pe.full_name requestor,
    l.item_description,
    pr.segment1 project_num,
    t.task_number,
    c.segment1,
    c.segment2
    FROM po_requisition_headers_all r,
         po_requisition_lines_all l,  
    (SELECT project_id,task_id,code_combination_id, distribution_id,requisition_line_id,creation_date  FROM
    (SELECT project_id,task_id,code_combination_id,distribution_id,creation_date,requisition_line_id,ROW_NUMBER ()
    OVER (PARTITION BY requisition_line_id ORDER BY requisition_line_id,distribution_id ) rn
    FROM po_req_distributions_all pod) WHERE rn = 1) d,
    gl_code_combinations c,
    POR_CATEGORY_LOV_V p,
    per_people_v7 pe,
    PA_PROJECTS_ALL pr,
    PA_TASKS_ALL_V t,
    ap_vendors_v v,
    WHERE  d.creation_date >= nvl(to_date(:DATE_LAST_CHECKED,
    'DD-MON-YYYY HH24:MI:SS'),SYSDATE-1)
    AND
    l.requisition_header_id = r.requisition_header_id
    AND l.requisition_line_id = d.requisition_line_id
    AND d.code_combination_id = c.code_combination_id
    AND r.APPS_SOURCE_CODE = 'POR'
    AND l.category_id = p.category_id
    AND r.authorization_status IN ('IN PROCESS','PRE-APPROVED','APPROVED')
    AND l.to_person_id = pe.person_id
    AND pr.project_id(+) = d.project_id
    AND t.project_id(+) = d.project_id
    AND t.task_id(+) = d.task_id
    AND v.vendor_id(+) = l.vendor_id
    and r.requisition_header_id in(
    SELECT requisition_header_id FROM po_requisition_lines_all pl                    
    GROUP BY requisition_header_id HAVING SUM(nvl(pl.quantity,0) * nvl(pl.unit_price, 0)) >=100000)
    group by
    peo.email_address,
    r.REQUISITION_HEADER_ID,
    r.segment1 ,
    to_char(l.line_num) ,
    v.vendor_name, 
    p.CONCATENATED_SEGMENTS ,
    to_char(round((nvl(l.quantity, 0) * nvl(l.unit_price, 0))),'99,999,999,999.99'),
    TO_CHAR(l.need_by_date,'MM/DD/YYYY') ,
    pe.full_name ,
    l.item_description,
    c.segment1,
    c.segment2,
    pr.segment1 ,
    t.task_number
    UNION ALL
    SELECT 
    r.segment1 requistion_num,
    to_char(l.line_num) line_num,
    v.vendor_name supplier, 
    p.CONCATENATED_SEGMENTS category,
    to_char(round((nvl(l.quantity, 0) * nvl(l.unit_price, 0))),'99,999,999,999.99'),
    TO_CHAR(l.need_by_date,'MM/DD/YYYY') need_by_date,
    pe.full_name requestor,
    l.item_description,
    pr.segment1 project_num,
    t.task_number,
    c.segment1,
    c.segment2
    FROM po_requisition_headers_all r,
         po_requisition_lines_all l,  
    (SELECT project_id,task_id,code_combination_id, distribution_id,requisition_line_id,creation_date  FROM
    (SELECT project_id,task_id,code_combination_id,distribution_id,creation_date,requisition_line_id,ROW_NUMBER ()
    OVER (PARTITION BY requisition_line_id ORDER BY requisition_line_id,distribution_id ) rn
    FROM po_req_distributions_all pod) WHERE rn = 1) d,
    gl_code_combinations c,
    POR_CATEGORY_LOV_V p,
    per_people_v7 pe,
    PA_PROJECTS_ALL pr,
    PA_TASKS_ALL_V t,
    ap_vendors_v v,
    fnd_lookup_values_vl flv,
    per_people_v7 p1,
    per_people_v7 p2
    WHERE  d.creation_date >= nvl(to_date('11-APR-2008',
    'DD-MON-YYYY HH24:MI:SS'),SYSDATE-1)
    AND
    l.requisition_header_id = r.requisition_header_id
    AND l.requisition_line_id = d.requisition_line_id
    AND d.code_combination_id = c.code_combination_id
    AND r.APPS_SOURCE_CODE = 'POR'
    AND l.org_id = 141
    AND l.category_id = p.category_id
    AND r.authorization_status IN ('IN PROCESS','PRE-APPROVED','APPROVED')
    AND l.to_person_id = pe.person_id
    AND pr.project_id(+) = d.project_id
    AND t.project_id(+) = d.project_id
    AND t.task_id(+) = d.task_id
    AND v.vendor_id(+) = l.vendor_id
    AND flv.attribute1=c.segment1
    AND flv.attribute2=c.segment2
    AND flv.lookup_type = 'BIO_FIN_APPROVER_INFO'
    and flv.attribute3 = p1.person_id(+)
    and flv.attribute4 = p2.person_id(+)
    and r.requisition_header_id in(
    SELECT requisition_header_id FROM po_requisition_lines_all pl                    
    GROUP BY requisition_header_id HAVING SUM(nvl(pl.quantity,0) * nvl(pl.unit_price, 0)) >=100000)
    group by
    r.REQUISITION_HEADER_ID,
    r.segment1 ,
    to_char(l.line_num) ,
    v.vendor_name, 
    p.CONCATENATED_SEGMENTS ,
    to_char(round((nvl(l.quantity, 0) * nvl(l.unit_price, 0))),'99,999,999,999.99'),
    TO_CHAR(l.need_by_date,'MM/DD/YYYY') ,
    pe.full_name ,
    l.item_description,
    c.segment1,
    c.segment2,
    pr.segment1 ,
    t.task_number

  • Dataset from two set of tables based on condition

    I have two queries that will return same columns from two different set of tables ( column mapping has been taken care of). The return type is out ref cursor. (P_SUPPLY_REORDER )
    Query 1-SO
    select
    so.SMO_NO,
    so.SPLY_ORD_DT,
    so.fk_CUST_ID as CUST_ID,
    so.CUST_PO_NO,
    so.ATTENTION_NAME,
    sum(sol.SPLY_ORD_QTY) as SPLY_ORD_QTY,
    --sum(sol.sply_shp_qty),
    so.ST_NAME,
    so.ADDR_LN_2,
    so.ADDR_LN_3,--sta.SHIP_TO_ADDRESS_LINE_3,
    so.CITY_NAME,
    so.ST_TERR_CD,
    so.ZIP_CD,
    so.SPCL_SHP_INSTR_TXT,
    so.SHP_CNFRM_DT
    ,XCOM_ORDER_NO
    from
    supply_order so,
    supply_order_line sol,
    XCOM_ORDER_HEADER
    where
    so.FK_CUST_ID =in_cust_id
    and so.pc_ord_no = sol.fk_pc_ord_no and
    XCOM_ORDER_HEADER.FK_PC_ORD_NO = so.PC_ORD_NO
    group by so.SMO_NO, so.SPLY_ORD_DT, so.fk_CUST_ID,
    so.CUST_PO_NO, so.ATTENTION_NAME,
    so.ST_NAME, so.ADDR_LN_2, so.ADDR_LN_3, so.CITY_NAME, so.ST_TERR_CD,
    so.ZIP_CD, so.SPCL_SHP_INSTR_TXT, so.SHP_CNFRM_DT, XCOM_ORDER_NO;
    Query-2 Xcom
    select
    null as sMO_NO,
    xso.created_date as SPLY_ORD_DT,
    xso.fk_cust_id as cust_id,
    cust.cust_po_no as cust_PO_NO
    ,(sta.SHIP_TO_ATTN_FIRST_NAME||''||sta.SHIP_TO_ATTN_LAST_NAME) as attention_name,
    xsol.CARTONS_ORDERED as SPLY_ORD_QTY,
    --sum(sol.sply_shp_qty),
    sta.SHIP_TO_ADDRESS_LINE_1 as ST_NAME,
    sta.SHIP_TO_ADDRESS_LINE_2 as ADDR_LN_2,
    --sta.SHIP_TO_ADDRESS_LINE_3,
    NULL as ADDR_LN_3,
    sta.ship_to_city as CITY_NAME,
    sta.SHIP_TO_STATE as ST_TERR_CD,
    sta.SHIP_TO_POSTAL_CODE as ZIP_CD,
    sta.SHIPPING_INSTRUCTIONS as SPCL_SHP_INSTR_TXT,
    null  as SHP_CNFRM_DT,
    xso.XCOM_ORDER_NO as XCOM_ORDER_NO
    from
    XCOM_ORDER_HEADER xso,
    XCOM_ORDER_LINES xsol,
    customer cust,
    ship_to_address sta
    where
    cust.cust_id = xso.fk_cust_id and
    sta.fk_cust_id = xso.fk_cust_id
    and xso.FK_CUST_ID =in_cust_id
    and xso.FK_PC_ORD_NO is null
    and xso.xcom_order_no = xsol.fk_xcom_order_no;Now the requirement is
    One of four conditions are possible for each Supply Reorder Number:
    •     Both table queries return no records
    -     Populate all the P_SUPPLY_REORDER output fields with nulls
    •     SUPPLY_ORDER returns a record, but XCOM_ORDER_HEADER returns no records
    -     Populate output fields with values from the join of SUPPLY_ORDER and SUPPLY_ORDER_LINE.
    •     SUPPLY_ORDER returns no records, but XCOM_ORDER_HEADER returns one record
    -     Populate output fields with values from the join of XCOM_ORDER_HEADER and XCOM_ORDER_LINES.
    •     SUPPLY_ORDER returns a record, and XCOM_ORDER_HEADER returns a record; find out the latest order by comapring max(SPLY_ORD_DT)
    from SUPPLY_ORDER with max(CREATED_DATE) from XCOM_ORDER_HEADER.
    -     If the latest order is in SUPPLY_ORDER, then populate output fields with values from the join of SUPPLY_ORDER and SUPPLY_ORDER_LINE.
    -     If order dates are equal from both join results, then populate output fields with values from the join of SUPPLY_ORDER and SUPPLY_ORDER_LINE.
    -     If the latest order is in XCOM_ORDER_HEADER, then populate output fields with values from the join of XCOM_ORDER_HEADER and XCOM_ORDER_LINES.
    Question is how can we switch over the queries to pull respective dataset based on these conditions ( checking that which table join is going to return a row and then based upon latest order if both tables return a row) and all this logic as part of single SQL statement that is returned as OUT Ref Cursor.
    Your help will be really appreciated. Thanks much in advance.

    It would be easier if you would supply a dataset we can work with. And I assume the queries are called within a other cursor which returns in_cust_id right?
    The best/fasters way would be to say goodbye to nested cursors.
    You need:
    the in_cust_id_query
    extend SUPPLY_ORDER with a new field
    max(CREATED_DATE) over (partiontion by cust_id) max_created_dateextend XCOM_ORDER_HEADER with a new field
    max(SPLY_ORD_DT)over (partiontion by cust_id) max_sply_ord_dateThen you make a 3 level SELECT
    the innerst () join all 3 sources
    the second level names the rule
    the outermost level applies the rule to each field:
    SELECT
    -- case statement to each field:
    CASE rule
    WHEN 1 THEN null
    WHEN 2 THEN s_smo_no
    WHEN 3 .... END  SMO_NO,
    SELECT
    -- check all rules:
    case
    --• Both table queries return no records
    when s_in_cust_id IS NULL and h_in_cust_id IS NULL then 1
    --• SUPPLY_ORDER returns a record, but XCOM_ORDER_HEADER returns no records
    when s_in_cust_id IS NOT NULL and h_in_cust_id IS NULL then 2
    --• SUPPLY_ORDER returns no records, but XCOM_ORDER_HEADER returns one record
    when s_in_cust_id IS NULL and h_in_cust_id IS NOT NULL then 3
    ELSE 4 end rule_id,
    i.*
    (SELECT c.in_cust_id,
                 s.*, (with alias of course eg. s_...)
                h.*, (with alias of course  e.g. h_...)
    from  in_cust_id_query c
      LEFT JOIN SUPPLY_ORDER s
       on (c.in_cust_id = s.in_cust_id)
    LEFT XCOM_ORDER_HEADER h
      on (c.in_cust_id = h.in_cust_id) i;I hope you get it.
    Of course you can do what you probable currently do. Open both cursors one by one and compare the results. But when you have 10000 in_cust_ids you start 20000 queries and you have a lot of code.
    This method I'm showing need these roles coded as well, but you'll have ONE single query and that will be as fast as ORACLE can be. And I believe it's easier to read too. (But not simple!)
    If you need more help, try to supply an example we can work on.
    -- andy

  • Problem about fill set up table

    Hi all,
    I got a problem when I run the fill set up table for sales billing data source. what I did is:
    first delete the setup table then oli9bw->type in a sales document no.(as I only want this order data), then give a run name, last execute, but I got the message as follow:
    Data source 2LIS_13_VDITM contains data still to be transferred
    Could you pls explain what this means, and what should I do to solve the problem
    Thanks

    Set up table definition
    Setup table is store the historical data, where as the delta records are updated in delta queue not to set up table.
    So once historical data is loaded you can delete the contents of the set up table.
    name is extract sturcture of your data source + setup
    also data is taken from setup table when an init / full/ full repair load is done.
    so in order to take ur records from the table delete the setup table(LBWG) and initialize it with ur records.
    You can find the contents SE16
    Filling set up tables
    transac LBWG
    + SBIW -> Logistics -> Managing Transfer Information Structures-> Setup of Statistical Data-> Application-Specific Setup of Statistical Data
    set up tables
    Set up tables
    Set up tables
    view of set up tables data in se11??????????
    Set Up tables..
    lo: delete set up tables: DOUBT
    LBWQ is the extraction queue and RSA7 is delta queue. Data is sent to delta queue from extraction queue through collective job scheduled in transaction LBWE.
    when we want to extract the data using LO Cockpit, the data will be come to the extraction queue first and from there it will processed to the delta queue. SO lbwq works as a outbound queue.
    If the update mode is Unserialised V3 then as soon as the document is posted it comes update table which you can see in Tr. Code SM13. After the jonb is scheduled the records come to RSA7 i.e delta queue from which BW pulls the data
    If you use u201CQueued Deltau201Dupdate method then the data moves to Extraction queue(LBWQ). Then run Collective update to move the data from LBWQ into Delta Queue (RSA7). Then schedule the data using the infopackage by selecting Delta Load in the update tab.
    So in If you are going to do Setup table filling, delete the data from LBWQ.
    The usage of Extraction queue(LBWQ) comes in to picture in delta loading. But actually system starts collecting the data (whenever there is document creation in R/3) after activating the EXtraction structure.
    Steps to Be Performed
    We know that there are 4 types of delta extraction that are available in LO. If you use u201CQueued Deltau201Dupdate method then the data moves to Extraction queue(LBWQ). Then run Collective update to move the data from LBWQ into Delta Queue (RSA7). Then schedule the data using the infopackage by selecting Delta Load in the update tab.
    Here is Roberto's weblog:
    /people/sap.user72/blog/2005/02/14/logistic-cockpit--when-you-need-more--first-option-enhance-it
    /people/sap.user72/blog/2004/12/16/logistic-cockpit-delta-mechanism--episode-one-v3-update-the-145serializer146
    /people/sap.user72/blog/2005/01/19/logistic-cockpit-delta-mechanism--episode-three-the-new-update-methods
    check this out...
    LBWQ - T code is for what?
    Ttransaction LBWQ
    What is LBWQ?
    Assign points if it helps
    Hope it helps
    regards
    Bala

  • Mass check scale price of info record

    Hi there:
    Is there any transaction or table to mass check scale price of info record?

    Any updates?

  • Extractor and set-up table

    I have VAITM enhanced with few new fields.
    now when do those fields get populated... after set-up tables are filled or during set-up tables filling.
    I mean to know  that when in delta que, extraction que and application table reading, those CMOD user exits populate extract structures.
    Rita

    Hi,
    As per my understanding, the extractor program there is a code which looks up to CMOD to see if there is any enhanced code involved.If it is there it will execute the code and populate the additional fields too.
    So whenever the extractor program is executed, it checks for the CMOD code which if found is executed for every record populated.
    In this case it should work in all the cases you have mentioned.
    Hope this helps,
    Regards,
    Anoop P

  • How do you display the tables of EBS R12 (VIS)?

    Hi,
    My sqlplus does not work but I have an Oracle sqldeveloper. How do you display the tables info of order management module in EBS R12 for windows?
    For example:
    Base tables
    oe_order_headers_all
    oe_lines_all
    oe_lot_serial_numbers
    oe_sales_credits
    oe_price_adjustments
    Interface tables:
    OE_HEADERS_IFACE_ALL
    OE_LINES_IFACE_ALL
    OE_RESERVTNS_IFACE_ALL
    OE_CREDITS_IFACE_ALL
    OE_PRICE_ADJS_IFACE_ALL
    OE_LOTSERIALS_IFACE_ALL
    OE_ACTIONS_IFACE_ALL
    Thanks,
    Fernando

    Hi Fernando,
    The tables for Order Management will mostly be in ONT schema - APPS schema only holds Views, Synonyms, PL/SQL Packages and similar abstractions.
    In SQL Developer make a connection to the APPS schema, then go to Synonyms and you'll see the tables (note you should filter as there are lots of objects in Apps schema)!
    In the synonym SQL tab you can see the base schema for the table, e.g.
    CREATE OR REPLACE SYNONYM "APPS"."OE_ORDER_HEADERS_ALL" FOR *"ONT"*."OE_ORDER_HEADERS_ALL";
    Make a connection to the base schema, e.g. ONT.
    Then you'll see the ONT tables under the Tables folder.
    Alternatively, connect as apps and execute SQL "desc OE_ORDER_HEADERS_ALL"
    Or look at the APPS schema views instead, e.g. OE_ORDER_HEADERS
    Regards,
    Gareth
    Edited by: gareth.roberts on Aug 18, 2009 11:52 AM

  • Set up table data extracted from R/3 not visible in data target of BW

    Hai friends,
             I am currently working on extracting data from R/3 to BW. I read the several docs given in the forum and did the following:
      1) In the LBWE transaction, my extract structure is already active.
      2) In SBIW, i went to the filling of set up tables for QM
      3) I executed the set up table extraction
      4) Then, i checked in RSA3. The extraction was successful.
      5) In BW, i replicated the datasource, and in  the infopackage, i selected in the  
          PROCESSING TAB, "PSA and then into Data targets (Package by Package)
      6) In UPDATE tab, i selected FULL UPDATE
      7) And then i did immediate load.
      8) In RSMO, it showed successful. (It showed the same number of records as in the
          RSA3 of R/3)
              But when i went into the data target (ODS) and checked for its contents, nothing is visible. Why is it so? HAve i skipped any step? Please help.
    Regards,
    Neha Solanki

    Hai,
           U r rite. It is an NW2004 system.This is what is displayed in the status tab in RSMO.
    Data successfully updated
    Diagnosis
    The request has been updated successfully.
    InfoSource : 2LIS_05_QE2
    Data type : Transaction Data
    Source system: QAS678
       And i can find no such button as u said.
    Regards,
    neha Solanki

  • 2LIS_02_ITM - Set up Table

    All,
        Can we run a set up table job (LO datasources) (in case of purchasing program RMCENEUA) parallely in (R/3)production environment by checking the block documnet option with different document date in selection ?
        The reason for this is that if we go with this option we can schedule maximun set up jobs in parallel and complete the setup table fill up in few hours instead of running series of jobs with different document date...
        please provide your inputs
    Thanks
    Kamal

    Hi kamal,
    We can do parallel execution ,but we have to keep in mind
    that, in case of jobs that were scheduled in parallel, we
    must assign different run names by all means. If a job is
    canceled (on purpose or through an exceptional situation)
    it can only be restarted by means of the correct run name.
    With rgds,
    Anil Kumar Sharma .P

Maybe you are looking for