How to determine how many Sales Quote turned to Sales Order

Hello Friends,
Is there any possibility to track how many Sales Quotation has been turned into an Sales Order in the system. If not then can you please help me in making the query for it as I am not able to make it.
Thanks,
Rahul

Dear M. Jenkins,
The query which you have sent is working but I need a diffrent report.
I want that from Sales Quotation the followings fields should be displayed :
1) Document Series Name
2) Sales Quotation Document Number
3) Sales Quotation Date
4) Business Partner Code
5) Business Partner Name 
6) Sales Employee Name
7) Document Total
8) Sales Quotation Status
9) Customer Refrence Number
10) Remarks
11) Territory Name
and from Sales Order
1) Sales Order Document Series Name
2) Sales Order Document Number
3) Sales Order Date
4) Business Partner Code
5) Business Partner Name 
6) Sales Employee Name
7) Document Total
8) Sales Order Status
9) Customer Refrence Number
10) Remarks
11) Territory Name
and parameter should have Start Date and End Date, Territory Name, Sales Employee Name.
Means I want a report by which I get to know how many sales Quotation are turned into Order with respect to Territory Name, Sales Employee with Start and End date.
Would be thankful if you can kindly help me out in it.
I have a below query but would need an help to make it as per the above requirement.
SELECT distinct 'Sales Quotation' "Type",
D0.DocStatus,
D0.DocNum SalesQuotation#,
D0.DocDate,
D0.DocDueDate,
D0.DocTotal,
'Sales Order' "Type",
I0.DocStatus,
I0.DocNum  SalesOrder#,
I0.DocDate,
I0.DocDueDate,
I0.DocTotal
FROM
((dbo.OQUT D0 inner Join dbo.QUT1 D1 on D0.DocEntry = D1.DocEntry)
left join
(dbo.ORDR I0 inner join dbo.RDR1 I1 on I0.DocEntry = I1.DocEntry)
on (I1.BaseType=23 AND D1.DocEntry = I1.BaseEntry AND D1.LineNum=I1.BaseLine))
WHERE
(D1.BaseType=-1 and D0.DocStatus = 'O' AND D1.DocDate>='[%0]' AND D1.DocDate<='[%1]')
OR (I1.BaseType=23 AND I1.BaseEntry IN
(SELECT Distinct DocEntry
FROM dbo.QUT1 WHERE D0.DocStatus = 'O' and BaseType=-1 AND DocDate>='[%0]' AND DocDate<='[%1]'))
ORDER BY D0.DocNum
Please guide and help.
Thanks,
Rahul

Similar Messages

  • How to determine how long an Oracle patch will take to complete?

    Hello,
    Does anybody know of a way as to how to determine how long a certain Oracle patch will take when it is applied?
    Regards.

    Another thing to consider is if the patch does any data conversion/inserts/updates/validation/etc. Running a patch against a small test instance with 100 users and 10,000 invoices may take considerably less time than it does against a production size instance with 22,000+ users and millions of invoices.
    We try to test every patch against a production size test instance prior to implementing it into production.

  • How to determine how old an iPod is

    My iPod is not working correctly with my Cadillac CUE system. I'm being told the iPod is too old.  How can I determine how old my iPod is?

    Hello there Leah,
    There is an article that talks about Identifying iPod models found here http://support.apple.com/kb/ht1353.
    Usually I quote the relevant section of the article but in this case, what I would quote would be a mile long heh.
    Regards,
    Sterling

  • How to determine how long dbms_job jobs take to execute?

    I have a number of dbms_Jobs (snapshot refresh jobs created automatically by creating the refresh groups) and I'm required to determine how long the jobs are running to verify some timelines. I was hoping it would be as easy as recording the last_date and next_date frequently and when there are changes, use the delta to determine (since the jobs are supposed to be scheduled from the completion of the previous job) but this doesn't appear to be working. I created a simple job that just sleeps for 15 seconds with an interval of 5 minutes, and (next_date - last_date) is always 5 minutes. What I would have expected was
    Job starts at 00:00:00 (last_date)
    Runs for 15 seconds, ends at 00:00:15
    Scheduled out 5 minutes: 00:05:15 (next_date)
    In that example next_date - last_date would be 5:15, and knowing the interval was 5 minutes I could determine it ran for 15 seconds.
    In practice though the current next_date becomes the next last_date (as if the interval is being calculated when the job starts, not completes), so it always shows 5 minutes which is just the interval.
    Short of not using the default jobs, and just submitting my own to do the refresh where I add some logic to record the duration, is there a good way to get this information? I'm using 10.2.0.4.
    Thanks

    Did you check out the TOTAL_TIME column of the DBA_JOBS view?
    :p

  • How to determine the delivery plant in a CRM Sales Order

    Hi experts:
    In CRM how to work the delivery plant determination? and I can change the Plan in a Sales Order before its download to ECC?
    I download the DNL_PLANT object but i don't understand what is the functionality of these are created like business partner in CRM.
    Thanks in advanced by your orientation about these questions that have me a little bit confuse.
    Sussie

    Hi Reddy
    Thanks for your reply, but the note says that only applies for India or Brazil legal purposes.
    In my case I still have confuse, in a normal Sales scenario CRM to R/3 how to works the determination of the delivery plant.
    I know that in R/3 is determined by:
    1. Customer-Material info record (Sales Area data)
    2. Customer Master data (Sales area data)
    3. Material Master data (Sales-area data)
    I can think that in CRM is determinate bye Material Master Data, but I still don't know if I can change
    it??  and where in the crm sales order before it is download to R/3??
    Thanks
    Sussie

  • How to determine how many times result set columns have same value

    Hi -
    I'm doing a report which will be used for payment trend analyses.
    My initial result set looks like this:
    HOUSEHOLD_ID     JAN_PMT     FEB_PMT     MAR_PMT     APR_PMT     MAY_PMT     JUN_PMT     JUL_PMT     AUG_PMT     SEP_PMT     OCT_PMT     NOV_PMT     DEC_PMT
    90026845409     1     1     1     1     2     1     1     1     1     0     1     0(many rows, of course; result set pivoted)
    I need to determine the households that have a > 0 value in three or more consecutive months.
    I'm hoping someone will have some suggestions because the only solutions I'm coming up with right now would be a coding nightmare (lots of "OR's"), and I'm assuming (hoping) there's a better solution out there.
    Thanks!
    Christine

    Hi Frank,
    I'm not sure I'm understanding how I would use those analytic functions. Here is my select statement:
    SELECT HOUSEHOLD_ID,
           SUM(CASE WHEN ( TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'MM'))     =  1 AND TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'DD')) > 19 ) OR
                         ( TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'MM')) - 1 =  1 AND TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'DD')) < 6  )
                    THEN 1 ELSE 0 END) AS JAN_PMT,
           SUM(CASE WHEN ( TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'MM'))     =  2 AND TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'DD')) > 19 ) OR
                         ( TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'MM')) - 1 =  2 AND TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'DD')) < 6  )
                    THEN 1 ELSE 0 END) AS FEB_PMT,
           SUM(CASE WHEN ( TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'MM'))     =  3 AND TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'DD')) > 19 ) OR
                         ( TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'MM')) - 1 =  3 AND TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'DD')) < 6  )
                    THEN 1 ELSE 0 END) AS MAR_PMT,
           SUM(CASE WHEN ( TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'MM'))     =  4 AND TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'DD')) > 19 ) OR
                         ( TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'MM')) - 1 =  4 AND TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'DD')) < 6  )
                    THEN 1 ELSE 0 END) AS APR_PMT,
           SUM(CASE WHEN ( TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'MM'))     =  5 AND TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'DD')) > 19 ) OR
                         ( TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'MM')) - 1 =  5 AND TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'DD')) < 6  )
                    THEN 1 ELSE 0 END) AS MAY_PMT,
           SUM(CASE WHEN ( TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'MM'))     =  6 AND TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'DD')) > 19 ) OR
                         ( TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'MM')) - 1 =  6 AND TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'DD')) < 6  )
                    THEN 1 ELSE 0 END) AS JUN_PMT,
           SUM(CASE WHEN ( TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'MM'))     =  7 AND TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'DD')) > 19 ) OR
                         ( TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'MM')) - 1 =  7 AND TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'DD')) < 6  )
                    THEN 1 ELSE 0 END) AS JUL_PMT,
           SUM(CASE WHEN ( TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'MM'))     =  8 AND TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'DD')) > 19 ) OR
                         ( TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'MM')) - 1 =  8 AND TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'DD')) < 6  )
                    THEN 1 ELSE 0 END) AS AUG_PMT,
           SUM(CASE WHEN ( TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'MM'))     =  9 AND TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'DD')) > 19 ) OR
                         ( TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'MM')) - 1 =  9 AND TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'DD')) < 6  )
                    THEN 1 ELSE 0 END) AS SEP_PMT,
           SUM(CASE WHEN ( TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'MM'))     = 10 AND TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'DD')) > 19 ) OR
                         ( TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'MM')) - 1 = 10 AND TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'DD')) < 6  )
                    THEN 1 ELSE 0 END) AS OCT_PMT,
           SUM(CASE WHEN ( TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'MM'))     = 11 AND TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'DD')) > 19 ) OR
                         ( TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'MM')) - 1 = 11 AND TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'DD')) < 6  )
                    THEN 1 ELSE 0 END) AS NOV_PMT,
           SUM(CASE WHEN ( TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'MM'))     = 12 AND TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'DD')) > 19 ) OR
                         ( TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'MM')) - 1 = 12 AND TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'DD')) < 6  )
                    THEN 1 ELSE 0 END) AS DEC_PMT
      FROM MONETARY_TRANS
    WHERE MONETARY_TRANS_TYPE_ID    = 1                             --payment
    --TESTING
    AND HOUSEHOLD_ID = 90026845409
       AND RESPONSIBLE_PARTY_TYPE_ID = 1                             --household
       AND RECEIVED_DATE > '01-JAN-2008'
       AND ( TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'DD')) < 6
        OR   TO_NUMBER(TO_CHAR(RECEIVED_DATE, 'DD')) > 19 )
       AND PREMIUM_AMOUNT            > 0
       AND BILLING_TRANS_TYPE_ID     = 6
       AND MONETARY_TRANS_TYPE_ID    = 1
       AND RESPONSIBLE_PARTY_TYPE_ID = 1
    GROUP BY HOUSEHOLD_IDAnd from this I get the results originally posted. From there I need to figure out the households that have values greater than 0 for three or more consecutive months.
    Thanks for your help........
    -Christine

  • How to determine:how many connections will be needed in RFC adapter setting

    Hi XI guru,
    We are working on following secnario:
    1. JDBC sender selects data from Oracle database, and send this data to SAP XI, XI use async RFC receiver adapter to call a self made RFC, in this RFC, we will call some BAPI and process some ABAP code, then if the reture code is ok, nothing will happen( means data successfully import into SAP ). if the reture code is not ok, another RFC will be called from ABAP code, send an asyn message to XI, XI use RFC receiver to update the message back to Oracle with specific flag.
    2. Why we use this secnario is, we want to avoid using BPM, this scenario are mass happening in our project, BPM will slow down the performance.
    3. We also don't want to use sync RFC, because BAPI and ABAP code will take sometime to run, we don;t want XI want for the RFC reture message.
    The question is:
    We successfully run this secnario in unit testing, but when we work on stress testing,found some times we will miss the RFC call. For example:JDBC sender sends message to XI for 10 times in a very short time( we use fault data to make  every message has a send back RFC message in stress testing), let's say poll interval 5 sec one time. but send back RFC only be called 7 or 8 times, 2-3 message were missing.
    we get back to SXMB_MONI, see all JDBC sender 10 message are successful, and 7- 8 send back RFC messages also successful. No log in SM58 RFC monitor. Then we go to modify maximum connection, and set it to 5, means 5 RFC coconnection possible(as far as I know, this setting means like this, but of course I maybe wrong)
    Question is: how can I know how many coconnections I need to setup, and if some RFC calls are missing, is there any place whatever in XI or R/3 can help me to monitor the log, then I can know they are not successful for some reasons,but not just simply miss as they never happened?
    That will be helpful, if some one can tell the story of what happen RFC works at background.
    Thank you very much for your time and effort on this

    HI,
    Maximum possible number of initial connections required between sender system and adapter. This value is limited internally to a maximum of 50
    See the below links
    http://help.sap.com/saphelp_nw2004s/helpdata/en/67/6d0540ba5ee569e10000000a155106/content.htm
    http://help.sap.com/saphelp_nw2004s/helpdata/en/33/c6e63b60c25767e10000000a11402f/content.htm
    http://help.sap.com/saphelp_nw2004s/helpdata/en/a4/db833d4c47ea4ea1a9b7af3c535ff2/content.htm
    http://help.sap.com/saphelp_nw2004s/helpdata/en/6e/3f560a3e0f6f4a90fb1e7db1388d4d/frameset.htm
    http://help.sap.com/saphelp_nw2004s/helpdata/en/25/90d038f58f863de10000000a11402f/frameset.htm
    /people/swaroopa.vishwanath/blog/2005/03/11/rfc-migration-to-bapi-in-r3--step-by-step
    Regards
    Chilla

  • How to determine output based on the content of purchase order

    Hi Gurus,
    I've a typical requirement at my present project. With vendor X we have punchout catalogs. while creation of shopping cart, we are not allowing the user to mix catalog items and non catalog items into the same shopping cart. Once the shopping cart all shopping carts are automatically coverted into PO.
    Now the challenge we are facing is, if the PO is created based on items from catalog, we want to send such PO via XML interface to the vendor. If the PO is created based of items which are free-form, we want to send such PO via Smartform email attachment.
    As this condition can only be determined at runtime, I'm finding a challenge to configure this. In both PO cases, vendor is same but the content of the PO is different. One contains all items that are catalog based and other contains items which are non catalog.
    Can someone help me find solution for this issue.
    Regards,
    Sam K

    Change Forms for Document Output
    You can use the Business Add-In BBP_OUTPUT_CHANGE_SF to override the Smart Forms that are defined in the activity Define Actions for Document Output.
    This BAdI serves as an alternative to the IMG activity Condition-Dependent Document Output.
    Note that no Smart Forms are used for output via XML.
    Standard settings
    The following parameters are available in the method CHANGE_FORMS. Depending on the input parameters, you can define other Smart Forms to output the document type.
    Importing
    IV_GUID
    Globally Unique Identifier of the document.
    IV_MEDIUM
    Output medium found for this document according to Define Actions for Document Output (print: PRN, fax: FAX, e-mail: MAI, XML: MET).
    IS_PARTNER
    Business partner to whom the document is sent.
    IV_OBJECT_TYPE
    Business transaction type.
    Changing
    CV_SMARTFORM
    Smart Form used to output the document.
    E-mail output
    Smart Form with which the document is added to the e-mail in the form of a PDF attachment.
    Note:
    Note that the BAdI with this parameter overrides the parameter form name (technical name of the Smart Form BBP_PO for example) irrespective of the defined determination technology (SRM: Purchase Order Output Determination, for example).
    CV_SMARTFORM_MAIL
    Smart Form used to define content and layout of the e-mail text in the case of e-mail output.
    CV_SUBJECT_MAIL
    Smart Form that determines the text in the (short description of the e-mail content).
    As an alternative to this BAdI, you can control document output via determination technology.
    For more information, see Condition-Dependent Document Output.
    EXPLORE THIS badi

  • SAP how to determine the account in Goods receipt for purchase order?

    We know when we make a Goods receipt for purchase order,we will use the account in GBB and BSX define in OBYC. But where can we find the rule? I mean why system can find the account in GBB and BSX , not the other TEcode?

    HI
    This is again based on the movement type which SAP has created a link internally.
    So if you see when a document is posted it basically derives the GL accounts based on the Movement type,Valuation class,valaution area and so on.
    Good luck
    Hari

  • Quick question on how CBO determines hows to access table partitions

    Hey all,
    I have a large table that is range partitioned on a "code" within the table (about 23 unique values). I'm able to query it very quickly and easily when I specify the code in a query... ie:
    select *
    from large_table
    where code = 'X'
    However, I am looking to calculate the code based on a reference table that I have. I can easily re-create the performance drop doing:
    select *
    from large_table
    where code = (Select 'X' from dual)
    When running the 2nd SQL, my query takes a very long time. I think I understand why (CBO can't "guess" the results of the subquery) and I understand that the easiest to fix this would be to just run the subquery first and pass its results to a second query.
    My question is if anyone knows of a way to provide the CBO a hint or any other trick that would allow Oracle to scan a single partition instead of the whole table using a method like this?

    Thanks Tubby, that link at least helps confirm my original reasoning for the degraded performance and appears to be for similar reasons
    As for the explain plans, here they are:
    (Slow Query -- Object names have been changed)
    | Id | Operation | Name | Rows | Bytes |TempSpc| Cost (%CPU)| Time | Pstart| Pstop | TQ |IN-OUT| PQ Distrib |
    | 0 | SELECT STATEMENT | | 413 | 219K| | 18794 (1)| 00:03:46 | | | | | |
    | 1 | PX COORDINATOR | | | | | | | | | | | |
    | 2 | PX SEND QC (RANDOM) | :TQ10013 | 413 | 219K| | 18794 (1)| 00:03:46 | | | Q1,13 | P->S | QC (RAND) |
    |* 3 | HASH JOIN RIGHT SEMI BUFFERED | | 413 | 219K| | 18794 (1)| 00:03:46 | | | Q1,13 | PCWP | |
    | 4 | BUFFER SORT | | | | | | | | | Q1,13 | PCWC | |
    | 5 | PX RECEIVE | | 2 | 8 | | 15 (7)| 00:00:01 | | | Q1,13 | PCWP | |
    | 6 | PX SEND HASH | :TQ10003 | 2 | 8 | | 15 (7)| 00:00:01 | | | | S->P | HASH |
    | 7 | VIEW | VW_NSO_1 | 2 | 8 | | 15 (7)| 00:00:01 | | | | | |
    | 8 | VIEW | view_ref_table | 2 | 26 | | 15 (7)| 00:00:01 | | | | | |
    |* 9 | TABLE ACCESS FULL | ref_table | 2 | 22 | | 15 (7)| 00:00:01 | | | | | |
    | 10 | PX RECEIVE | | 4543 | 2391K| | 18779 (1)| 00:03:46 | | | Q1,13 | PCWP | |
    | 11 | PX SEND HASH | :TQ10012 | 4543 | 2391K| | 18779 (1)| 00:03:46 | | | Q1,12 | P->P | HASH |
    |* 12 | HASH JOIN | | 4543 | 2391K| | 18779 (1)| 00:03:46 | | | Q1,12 | PCWP | |
    | 13 | BUFFER SORT | | | | | | | | | Q1,12 | PCWC | |
    | 14 | PX RECEIVE | | 11 | 187 | | 5 (0)| 00:00:01 | | | Q1,12 | PCWP | |
    | 15 | PX SEND BROADCAST | :TQ10002 | 11 | 187 | | 5 (0)| 00:00:01 | | | | S->P | BROADCAST |
    |* 16 | TABLE ACCESS FULL | threshold_table | 11 | 187 | | 5 (0)| 00:00:01 | | | | | |
    |* 17 | VIEW | | 430K| 214M| | 18774 (1)| 00:03:46 | | | Q1,12 | PCWP | |
    | 18 | WINDOW SORT | | 430K| 130M| 280M| 18774 (1)| 00:03:46 | | | Q1,12 | PCWP | |
    | 19 | PX RECEIVE | | 430K| 130M| | 18772 (1)| 00:03:46 | | | Q1,12 | PCWP | |
    | 20 | PX SEND HASH | :TQ10011 | 430K| 130M| | 18772 (1)| 00:03:46 | | | Q1,11 | P->P | HASH |
    |* 21 | HASH JOIN BUFFERED | | 430K| 130M| | 18772 (1)| 00:03:46 | | | Q1,11 | PCWP | |
    | 22 | BUFFER SORT | | | | | | | | | Q1,11 | PCWC | |
    | 23 | PX RECEIVE | | 7332 | 157K| | 16 (13)| 00:00:01 | | | Q1,11 | PCWP | |
    | 24 | PX SEND HASH | :TQ10001 | 7332 | 157K| | 16 (13)| 00:00:01 | | | | S->P | HASH |
    | 25 | VIEW | view_ref_table | 7332 | 157K| | 16 (13)| 00:00:01 | | | | | |
    | 26 | WINDOW SORT | | 7332 | 80652 | | 16 (13)| 00:00:01 | | | | | |
    | 27 | TABLE ACCESS FULL | ref_table | 7332 | 80652 | | 14 (0)| 00:00:01 | | | | | |
    | 28 | PX RECEIVE | | 129K| 36M| | 18756 (1)| 00:03:46 | | | Q1,11 | PCWP | |
    | 29 | PX SEND HASH | :TQ10010 | 129K| 36M| | 18756 (1)| 00:03:46 | | | Q1,10 | P->P | HASH |
    | 30 | BUFFER SORT | | 413 | 219K| | | | | | Q1,10 | PCWP | |
    | 31 | NESTED LOOPS | | 129K| 36M| | 18756 (1)| 00:03:46 | | | Q1,10 | PCWP | |
    | 32 | NESTED LOOPS | | 124K| 30M| | 14444 (1)| 00:02:54 | | | Q1,10 | PCWP | |
    | 33 | NESTED LOOPS | | 122K| 25M| | 10173 (1)| 00:02:03 | | | Q1,10 | PCWP | |
    |* 34 | HASH JOIN | | 123K| 23M| | 5893 (1)| 00:01:11 | | | Q1,10 | PCWP | |
    | 35 | PX RECEIVE | | 124K| 22M| | 5629 (1)| 00:01:08 | | | Q1,10 | PCWP | |
    | 36 | PX SEND HASH | :TQ10008 | 124K| 22M| | 5629 (1)| 00:01:08 | | | Q1,08 | P->P | HASH |
    | 37 | BUFFER SORT | | 413 | 219K| | | | | | Q1,08 | PCWP | |
    | 38 | NESTED LOOPS | | 124K| 22M| | 5629 (1)| 00:01:08 | | | Q1,08 | PCWP | |
    |* 39 | HASH JOIN | | 122K| 20M| | 1797 (2)| 00:00:22 | | | Q1,08 | PCWP | |
    | 40 | PX RECEIVE | | 119K| 17M| | 1101 (3)| 00:00:14 | | | Q1,08 | PCWP | |
    | 41 | PX SEND HASH | :TQ10006 | 119K| 17M| | 1101 (3)| 00:00:14 | | | Q1,06 | P->P | HASH |
    |* 42 | HASH JOIN | | 119K| 17M| | 1101 (3)| 00:00:14 | | | Q1,06 | PCWP | |
    | 43 | BUFFER SORT | | | | | | | | | Q1,06 | PCWC | |
    | 44 | PX RECEIVE | | 86 | 602 | | 3 (0)| 00:00:01 | | | Q1,06 | PCWP | |
    | 45 | PX SEND BROADCAST | :TQ10000 | 86 | 602 | | 3 (0)| 00:00:01 | | | | S->P | BROADCAST |
    | 46 | TABLE ACCESS FULL | OFFICE | 86 | 602 | | 3 (0)| 00:00:01 | | | | | |
    |* 47 | HASH JOIN | | 119K| 16M| | 1098 (3)| 00:00:14 | | | Q1,06 | PCWP | |
    | 48 | PX RECEIVE | | 47 | 940 | | 2 (0)| 00:00:01 | | | Q1,06 | PCWP | |
    | 49 | PX SEND BROADCAST | :TQ10004 | 47 | 940 | | 2 (0)| 00:00:01 | | | Q1,04 | P->P | BROADCAST |
    | 50 | PX BLOCK ITERATOR | | 47 | 940 | | 2 (0)| 00:00:01 | | | Q1,04 | PCWC | |
    |* 51 | TABLE ACCESS FULL | REF | 47 | 940 | | 2 (0)| 00:00:01 | | | Q1,04 | PCWP | |
    |* 52 | HASH JOIN | | 701K| 84M| | 1095 (3)| 00:00:14 | | | Q1,06 | PCWP | |
    | 53 | PX RECEIVE | | 12 | 240 | | 2 (0)| 00:00:01 | | | Q1,06 | PCWP | |
    | 54 | PX SEND BROADCAST | :TQ10005 | 12 | 240 | | 2 (0)| 00:00:01 | | | Q1,05 | P->P | BROADCAST |
    | 55 | PX BLOCK ITERATOR | | 12 | 240 | | 2 (0)| 00:00:01 | | | Q1,05 | PCWC | |
    |* 56 | TABLE ACCESS FULL | REF | 12 | 240 | | 2 (0)| 00:00:01 | | | Q1,05 | PCWP | |
    | 57 | PX BLOCK ITERATOR | | 16M| 1617M| | 1090 (2)| 00:00:14 | 1 | 24 | Q1,06 | PCWC | |
    | 58 | TABLE ACCESS FULL | Fact_Table | 16M| 1617M| | 1090 (2)| 00:00:14 | 1 | 24 | Q1,06 | PCWP | |
    | 59 | PX RECEIVE | | 13M| 323M| | 693 (1)| 00:00:09 | | | Q1,08 | PCWP | |
    | 60 | PX SEND HASH | :TQ10007 | 13M| 323M| | 693 (1)| 00:00:09 | | | Q1,07 | P->P | HASH |
    | 61 | PX BLOCK ITERATOR | | 13M| 323M| | 693 (1)| 00:00:09 | | | Q1,07 | PCWC | |
    | 62 | TABLE ACCESS FULL | Dimension 1 | 13M| 323M| | 693 (1)| 00:00:09 | | | Q1,07 | PCWP | |
    | 63 | TABLE ACCESS BY INDEX ROWID | Dimesnion 2 | 1 | 12 | | 0 (0)| 00:00:01 | | | Q1,08 | PCWP | |
    |* 64 | INDEX UNIQUE SCAN | Dimension 2 Index | 1 | | | 0 (0)| 00:00:01 | | | Q1,08 | PCWP | |
    | 65 | PX RECEIVE | | 15M| 151M| | 261 (2)| 00:00:04 | | | Q1,10 | PCWP | |
    | 66 | PX SEND HASH | :TQ10009 | 15M| 151M| | 261 (2)| 00:00:04 | | | Q1,09 | P->P | HASH |
    | 67 | PX BLOCK ITERATOR | | 15M| 151M| | 261 (2)| 00:00:04 | | | Q1,09 | PCWC | |
    | 68 | INDEX FAST FULL SCAN | Dimension 1 Index | 15M| 151M| | 261 (2)| 00:00:04 | | | Q1,09 | PCWP | |
    | 69 | TABLE ACCESS BY INDEX ROWID | Dimension 3 | 1 | 18 | | 0 (0)| 00:00:01 | | | Q1,10 | PCWP | |
    |* 70 | INDEX UNIQUE SCAN | Dimension 3 Index | 1 | | | 0 (0)| 00:00:01 | | | Q1,10 | PCWP | |
    | 71 | PARTITION HASH ITERATOR | | 1 | 39 | | 0 (0)| 00:00:01 | KEY | KEY | Q1,10 | PCWP | |
    | 72 | TABLE ACCESS BY LOCAL INDEX ROWID| Dimension 4 | 1 | 39 | | 0 (0)| 00:00:01 | KEY | KEY | Q1,10 | PCWP | |
    |* 73 | INDEX UNIQUE SCAN | dimension 6 Index | 1 | | | 0 (0)| 00:00:01 | KEY | KEY | Q1,10 | PCWP | |
    | 74 | TABLE ACCESS BY INDEX ROWID | Dimension 5 | 1 | 38 | | 0 (0)| 00:00:01 | | | Q1,10 | PCWP | |
    |* 75 | INDEX UNIQUE SCAN | Dimension 5 index | 1 | | | 0 (0)| 00:00:01 | | | Q1,10 | PCWP | |
    (Fast Query -- Object Names Changed)
    | Id | Operation | Name | Rows | Bytes |TempSpc| Cost (%CPU)| Time | Pstart| Pstop | TQ |IN-OUT| PQ Distrib |
    | 0 | SELECT STATEMENT | | 107 | 57673 | | 1302 (1)| 00:00:16 | | | | | |
    | 1 | PX COORDINATOR | | | | | | | | | | | |
    | 2 | PX SEND QC (RANDOM) | :TQ10009 | 107 | 57673 | | 1302 (1)| 00:00:16 | | | Q1,09 | P->S | QC (RAND) |
    |* 3 | HASH JOIN BUFFERED | | 107 | 57673 | | 1302 (1)| 00:00:16 | | | Q1,09 | PCWP | |
    | 4 | BUFFER SORT | | | | | | | | | Q1,09 | PCWC | |
    | 5 | PX RECEIVE | | 11 | 187 | | 5 (0)| 00:00:01 | | | Q1,09 | PCWP | |
    | 6 | PX SEND HASH | :TQ10002 | 11 | 187 | | 5 (0)| 00:00:01 | | | | S->P | HASH |
    |* 7 | TABLE ACCESS FULL | Threshold Table | 11 | 187 | | 5 (0)| 00:00:01 | | | | | |
    | 8 | PX RECEIVE | | 10163 | 5180K| | 1296 (1)| 00:00:16 | | | Q1,09 | PCWP | |
    | 9 | PX SEND HASH | :TQ10008 | 10163 | 5180K| | 1296 (1)| 00:00:16 | | | Q1,08 | P->P | HASH |
    |* 10 | VIEW | | 10163 | 5180K| | 1296 (1)| 00:00:16 | | | Q1,08 | PCWP | |
    | 11 | WINDOW SORT | | 10163 | 3185K| 6792K| 1296 (1)| 00:00:16 | | | Q1,08 | PCWP | |
    | 12 | PX RECEIVE | | 10163 | 3185K| | 1295 (1)| 00:00:16 | | | Q1,08 | PCWP | |
    | 13 | PX SEND HASH | :TQ10007 | 10163 | 3185K| | 1295 (1)| 00:00:16 | | | Q1,07 | P->P | HASH |
    |* 14 | HASH JOIN BUFFERED | | 10163 | 3185K| | 1295 (1)| 00:00:16 | | | Q1,07 | PCWP | |
    | 15 | BUFFER SORT | | | | | | | | | Q1,07 | PCWC | |
    | 16 | PX RECEIVE | | 148 | 3256 | | 15 (7)| 00:00:01 | | | Q1,07 | PCWP | |
    | 17 | PX SEND HASH | :TQ10001 | 148 | 3256 | | 15 (7)| 00:00:01 | | | | S->P | HASH |
    | 18 | VIEW | view_ref table | 148 | 3256 | | 15 (7)| 00:00:01 | | | | | |
    | 19 | WINDOW SORT | | 148 | 2072 | | 15 (7)| 00:00:01 | | | | | |
    |* 20 | TABLE ACCESS FULL | ref table | 148 | 2072 | | 14 (0)| 00:00:01 | | | | | |
    | 21 | PX RECEIVE | | 6867 | 2005K| | 1279 (1)| 00:00:16 | | | Q1,07 | PCWP | |
    | 22 | PX SEND HASH | :TQ10006 | 6867 | 2005K| | 1279 (1)| 00:00:16 | | | Q1,06 | P->P | HASH |
    | 23 | BUFFER SORT | | 107 | 57673 | | | | | | Q1,06 | PCWP | |
    | 24 | NESTED LOOPS | | 6867 | 2005K| | 1279 (1)| 00:00:16 | | | Q1,06 | PCWP | |
    | 25 | NESTED LOOPS | | 6591 | 1679K| | 1050 (1)| 00:00:13 | | | Q1,06 | PCWP | |
    |* 26 | HASH JOIN | | 6604 | 1567K| | 821 (1)| 00:00:10 | | | Q1,06 | PCWP | |
    | 27 | BUFFER SORT | | | | | | | | | Q1,06 | PCWC | |
    | 28 | PX RECEIVE | | 86 | 602 | | 3 (0)| 00:00:01 | | | Q1,06 | PCWP | |
    | 29 | PX SEND HASH | :TQ10000 | 86 | 602 | | 3 (0)| 00:00:01 | | | | S->P | HASH |
    | 30 | TABLE ACCESS FULL | OFFICE | 86 | 602 | | 3 (0)| 00:00:01 | | | | | |
    | 31 | PX RECEIVE | | 6604 | 1522K| | 817 (1)| 00:00:10 | | | Q1,06 | PCWP | |
    | 32 | PX SEND HASH | :TQ10005 | 6604 | 1522K| | 817 (1)| 00:00:10 | | | Q1,05 | P->P | HASH |
    | 33 | NESTED LOOPS | | 6604 | 1522K| | 817 (1)| 00:00:10 | | | Q1,05 | PCWP | |
    | 34 | NESTED LOOPS | | 6541 | 1258K| | 590 (1)| 00:00:08 | | | Q1,05 | PCWP | |
    | 35 | NESTED LOOPS | | 6612 | 1207K| | 475 (1)| 00:00:06 | | | Q1,05 | PCWP | |
    | 36 | NESTED LOOPS | | 6507 | 1112K| | 272 (2)| 00:00:04 | | | Q1,05 | PCWP | |
    |* 37 | HASH JOIN | | 6352 | 924K| | 51 (4)| 00:00:01 | | | Q1,05 | PCWP | |
    | 38 | PX RECEIVE | | 47 | 940 | | 2 (0)| 00:00:01 | | | Q1,05 | PCWP | |
    | 39 | PX SEND BROADCAST | :TQ10003 | 47 | 940 | | 2 (0)| 00:00:01 | | | Q1,03 | P->P | BROADCAST |
    | 40 | PX BLOCK ITERATOR | | 47 | 940 | | 2 (0)| 00:00:01 | | | Q1,03 | PCWC | |
    |* 41 | TABLE ACCESS FULL | REF | 47 | 940 | | 2 (0)| 00:00:01 | | | Q1,03 | PCWP | |
    |* 42 | HASH JOIN | | 37236 | 4690K| | 48 (3)| 00:00:01 | | | Q1,05 | PCWP | |
    | 43 | PX RECEIVE | | 12 | 240 | | 2 (0)| 00:00:01 | | | Q1,05 | PCWP | |
    | 44 | PX SEND BROADCAST | :TQ10004 | 12 | 240 | | 2 (0)| 00:00:01 | | | Q1,04 | P->P | BROADCAST |
    | 45 | PX BLOCK ITERATOR | | 12 | 240 | | 2 (0)| 00:00:01 | | | Q1,04 | PCWC | |
    |* 46 | TABLE ACCESS FULL | REF | 12 | 240 | | 2 (0)| 00:00:01 | | | Q1,04 | PCWP | |
    | 47 | PX BLOCK ITERATOR | | 849K| 88M| | 46 (3)| 00:00:01 | KEY | KEY | Q1,05 | PCWC | |
    | 48 | TABLE ACCESS FULL | Fact Table | 849K| 88M| | 46 (3)| 00:00:01 | 5 | 5 | Q1,05 | PCWP | |
    | 49 | TABLE ACCESS BY INDEX ROWID | Dim 1     | 1 | 26 | | 0 (0)| 00:00:01 | | | Q1,05 | PCWP | |
    |* 50 | INDEX UNIQUE SCAN | Dim 1 index | 1 | | | 0 (0)| 00:00:01 | | | Q1,05 | PCWP | |
    | 51 | TABLE ACCESS BY INDEX ROWID | Dim 2 | 1 | 12 | | 0 (0)| 00:00:01 | | | Q1,05 | PCWP | |
    |* 52 | INDEX UNIQUE SCAN | Dim 2 index | 1 | | | 0 (0)| 00:00:01 | | | Q1,05 | PCWP | |
    |* 53 | INDEX UNIQUE SCAN | Dim 3 Index | 1 | 10 | | 0 (0)| 00:00:01 | | | Q1,05 | PCWP | |
    | 54 | PARTITION HASH ITERATOR | | 1 | 39 | | 0 (0)| 00:00:01 | KEY | KEY | Q1,05 | PCWP | |
    | 55 | TABLE ACCESS BY LOCAL INDEX ROWID| Dim 4 | 1 | 39 | | 0 (0)| 00:00:01 | KEY | KEY | Q1,05 | PCWP | |
    |* 56 | INDEX UNIQUE SCAN | Dim 4 Index | 1 | | | 0 (0)| 00:00:01 | KEY | KEY | Q1,05 | PCWP | |
    | 57 | TABLE ACCESS BY INDEX ROWID | Dim 5 | 1 | 18 | | 0 (0)| 00:00:01 | | | Q1,06 | PCWP | |
    |* 58 | INDEX UNIQUE SCAN | Dim 5 Index | 1 | | | 0 (0)| 00:00:01 | | | Q1,06 | PCWP | |
    | 59 | TABLE ACCESS BY INDEX ROWID | Dim 6 | 1 | 38 | | 0 (0)| 00:00:01 | | | Q1,06 | PCWP | |
    |* 60 | INDEX UNIQUE SCAN | Dim 6 Index | 1 | | | 0 (0)| 00:00:01 | | | Q1,06 | PCWP | |
    DB Version is 10g R2:
    Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bi
    PL/SQL Release 10.2.0.4.0 - Production
    "CORE     10.2.0.4.0     Production"
    TNS for IBM/AIX RISC System/6000: Version 10.2.0.4.0 - Productio
    NLSRTL Version 10.2.0.4.0 - Production

  • How to determine how long a User Profile incremental and a Audience incremental take to complete?

    I'm looking at the ManageUserProfileServiceApplication.aspx page for my User Profile service. Just wanted to confirm a couple things based on what I'm seeing here.
    Audience Compilation Schedule looks like it takes under a minute to complete:
    Profiles
    Number of User Profiles
    3,467
    Number of User Properties
    72
    Number of Organization Profiles
    1
    Number of Organization Properties
    15
    Audiences
    Number of Audiences
    66
    Uncompiled Audiences
    0
    Audience Compilation Status
    Idle
    Audience Compilation Schedule
    Every day at 05:00 AM
    Last Compilation Time
    Ended at 1/15/2015 5:00
    AM
    Profile Synchronization Settings
    Synchronization Schedule (Incremental)
    Every day at 03:00 AM
    Profile Synchronization Status
    Idle
    User Profiles seemed to take just a few minutes, even though the times are off (8:5x am instead of 3 am).
    Stage History Cleanup
    Start Time 1/15/2015 8:51:57 AM
    End Time 1/15/2015 8:51:58 AM
    MX
    Stage Active directory import
    Additions 0
    Updates 0
    Unchanged 0
    Successes 0
    Failures 0
    Start Time 1/15/2015 8:51:58 AM
    End Time 1/15/2015 8:52:14 AM
    MXCB
    Stage Active directory import
    Additions 0
    Updates 0
    Unchanged 0
    Successes 0
    Failures 0
    Start Time 1/15/2015 8:51:58 AM
    End Time 1/15/2015 8:52:14 AM
    US2
    Stage Active directory import
    Additions 0
    Updates 0
    Unchanged 0
    Successes 0
    Failures 0
    Start Time 1/15/2015 8:51:58 AM
    End Time 1/15/2015 8:51:58 AM
    MOSS
    Stage SharePoint Server import
    Additions 0
    Updates 0
    Unchanged 0
    Successes 0
    Failures 0
    Start Time 1/15/2015 8:51:58 AM
    End Time 1/15/2015 8:52:32 AM
    US
    Stage Active directory import
    Additions 0
    Updates 0
    Unchanged 0
    Successes 0
    Failures 0
    Start Time 1/15/2015 8:51:59 AM
    End Time 1/15/2015 8:52:00 AM
    MOSS
    Stage SharePoint Server synchronization
    Additions 0
    Updates 0
    Unchanged 0
    Successes 0
    Failures 0
    Start Time 1/15/2015 8:52:32 AM
    End Time 1/15/2015 8:52:33 AM
    US
    Stage Active directory synchronization
    Additions 0
    Updates 0
    Unchanged 0
    Successes 0
    Failures 0
    Start Time 1/15/2015 8:52:33 AM
    End Time 1/15/2015 8:52:43 AM
    MX
    Stage Active directory synchronization
    Additions 0
    Updates 0
    Unchanged 0
    Successes 0
    Failures 0
    Start Time 1/15/2015 8:52:58 AM
    End Time 1/15/2015 8:53:01 AM
    US2
    Stage Active directory synchronization
    Additions 0
    Updates 0
    Unchanged 0
    Successes 0
    Failures 0
    Start Time 1/15/2015 8:53:01 AM
    End Time 1/15/2015 8:53:02 AM
    MXCB
    Stage Active directory synchronization
    Additions 0
    Updates 0
    Unchanged 0
    Successes 0
    Failures 0
    Start Time 1/15/2015 8:53:02 AM
    End Time 1/15/2015 8:53:04 AM
    MOSS
    Stage SharePoint Server export
    Additions 0
    Updates 0
    Unchanged 0
    Successes 0
    Failures 579
    Start Time 1/15/2015 8:53:04 AM
    End Time 1/15/2015 8:53:51 AM
    Stage Post-processing
    Start Time 1/15/2015 8:53:51 AM
    End Time 1/15/2015 8:53:51 AM

    Audience Compilation is fairly intense job. That is why it has its own timer
    job(User Profile service application - audience compilation)to
    which runs weekly basis and not suggested to run everyday. You can use below blog to find out exact time duration of that job
    Extract Timer Job History Using PowerShell
    Please 'propose as answer' if it helped you, also 'vote helpful' if you like this reply.

  • How to determine max netstreams before failure

    Hi
    Everything on Adobe's website is listed as "unlimited" and scalable but I would like to know approximate actual numbers to estimate server hardware required
    I can't find out how to determine how many live audio/video netstreams per instance, specifically
    1) How many broadcasters per instance?
    2) How many subscribers/viewers per  broadcast?
    3) Does seperate application/instance provide better performance?
    I understand the exact number is based on the publish settings so everyone's setup is different, but is it just bandwidth consideration?
    Is there any server side functions and/or metrics besides general OS cpu/ram stats?

    • How do I determine how many songs will fit on a CD?
    Already answered by others.
    • I heard that the LAME encoder is the best out there, is there an easy way to use that encoder to convert AAC files?
    I don't think LAME is part of itunes. You need to find a different encoder. I think Max CDripper uses LAME. It's pretty common in other encoding tools so you need to search for one that you like and at a price you like.
    • Can I use the LAME encoder to convert the files and then use those files in iTunes instead of using iTunes to convert them?
    Yes. Understand though that nothing will convert iTunes AAC format DRM protected files except iTunes and then only to audio CD.
    • I have a play list I want to burn to an MP3 CD for the car. The playlist is composed of songs with both AAC and MP3 formats. Is there an easy way to convert an then use the AAC files for this playlist?
    iTunes will do this, or other audio encoders. The DRM limitation still applies, even to iTunes.
    Is there a high quality way of recording streamed songs along with the title/artist info?
    It depends upon the source. Different sources stream different ways, so what works with one will not always work with another. The one exception is something that captures any audio output from your computer. That doesn't capture with artist and song information, just any sound that happens to be coming from your computer. If it isn't easy to do it is probably because the people streaming the music really don't want to you do it, probably because of licensing, in which case the Terms of Use of this website don't let us discuss how to do it.

  • Determine how much to allocate for /etc/project

    Question on how to determine how much resources a particular is actually being used on a solaris 10 server because I don't want to over allocate, esp for max-shm-memory to a particular project.
    Current setting in /etc/project
    user.test:101:Test project:::project.max-sem-ids=(privileged,1024,deny);project.max-sem-nsems=(privileged,512,deny);project.max-shm-ids=(privileged,512,deny);project.max-shm-memory=(privileged,2048000000,deny)
    Thank you
    Mike

    Thanks wsanders.
    I did ipcs -b and added the values in column 7 of the shared memory section to determine the amount of shared memory used.

  • Query to determine sales quote and sales order

    Hi
    I am working on User exits.Whenever a sales quote or a sales order is saved the user exit USEREXIT_SAVE_DOC_PREPARE  gets triggered.I need to check whether a sales quote is being saved or a sales order is being saved.
    Can anyone tell me where i can find whether the save operation of sales quote or the save operation of  the sales order has triggered the user exit.

    Hi,
    Check for the field VBTYP of XVBAK or YVBAK structure, if it is 'C" then it is Sales Order,if it is 'B', then quotation.
    Regards,
    Anji

  • HT4059 Being new to iBook publishing I need to know if stats are available to determine how many times my iBook has been downloaded. Can anyone help please? Mr Ignorant

    Being new to iBook publishing I need to know if stats are available to determine how many times my iBook has been downloaded. Can anyone help please? I need to persuade someone that this really is worthwhile. Mr Ignorant.

    The only people who can possibly assist you with this is Apple Customer Relations, call your local Apple contact number and ask for Customer Relations then explain your situation clearly and politely (be firm but don't rant).
    You might want to investiage what the local laws are regarding defective goods and 'fit for use' definitions on warranties etc. Consumer Protection can be a useful tool to use or bargain with if needed ...

Maybe you are looking for

  • How to change the layout of a category view

    I don't know how to change the layout of a category. I want to put 3 image's behind each other followed by some text, like a table with 4 columns. Instead of puting then in 4 rows

  • Tagged items web part issue SP 2013

    Hi, we have configured Content tag cloud and we have two different issues. 1. we have configured to show the tags that are tagged under the current url by all users, but when we click on the tag, it shows all the documents that are tagged with that s

  • How can i validate an xml against an dtd which are both in a clob ?

    Hi, I think my question says it all. I would like to validate my xml which has been stored into a clob against a dtd also stored in a clob. It has to be done with the xdk for plsql. Regards, Geert.

  • Total pages in a form from the program

    Hi all      I have got a special situation, That I need  to know the total number pages that a sapscript will print with in a program. If i want it it it with in a form then I use &SAPSCRIPT-FORMPAGES&, what would be the opetion if i want to get the

  • Q10 Problems with icloud!

    This is very disappointed to me that Blackberry said that, icloud is supported but... i found it is so-called: support... What problems i have are: 1) The same with other guys here! no calendar sync with my icloud and my Q10 (may be all bb10 devices?