Rule Based distribution of leads

Hi
For maintaining the rules for rule- based distribution of leads, which business roles are relevant? Can we do it by using roles marketpro or salespro ?? Please guide.
Regards
Aakshi Suri

hi
can you plz explain your question elaborately
if you to use business role----
den used MARKETINGPRO bt before using the b.role convert into "z-transaction" den use it
thanking you
debi prasad

Similar Messages

  • Rule-based distribution for opportunities

    Hello,
           I've configured a rule for opportunity distribution in 'Rule Policies' in SAP CRM Web UI. The rule is shown as 'Active' and 'Released'. However, when I create an opportunity and select the distribution method as 'Rule-Based distribution', then the dropdown the field 'Rule' stays blank and I don't get any options to select a 'Rule'. Please provide your inputs on how the rule I created can be made available for an opportunity.
    Regards,
    Aditya Mishra

    Hi Aditya,
    Have you configured a service profile and assigned your rule policy to it? You can set this up in:
    SPRO --> E-mail Response Management System --> Service Manager --> Define Service Manager Profiles
    1. Create a new service profile
    2. Add in service ID DD_RE
    3. Under property ID add in CONTEXT using value DD_OPPORTUNITY
    4. Under property ID add in POLICY using the rule policy ID you created.
    I hope this helps.
    Regards,
    David.

  • Rule based Lead distribution is not working.

    Hi friends,
    I have created Lead distribution rule for assigment of Leads to Sales representative using below condition :
    If
    Lead Main Partner Postal Code Contains "12001"
    Then
    Partner by ID ( Assign Business Partner ID = Will Smith(BP id)
    Set Partner Function To = Sales Representative
    Set Lead Status To = In Process
    Execution behavior = Stop after execution if partner(s) found )
    I am using standard service Mgr Profile for lead distribution, when i am executing standard report CRM_DD_BATCH  by giving values Object Type, Transaction Type and Service Mgr Profile it does not work i.e. assignment not happens but in addition to this if i mention specific Lead no. into report field ID it works perfectly.
    I dont understand what i m missing, please help me if anyone has come across the same scenario.
    Thanks in Advance,
    Dipesh.

    Hi Dipesh,
    Have you ever fixed this issue? I have exactly similar problem...Appreciate if you can share the solution/root cause of this problem, if you found any. Thanks.

  • Lead Assignment Rule Based on Industry and Zip Code

    Can we set up a lead assingment rule based on multiple conditions?
    For example, if Industry = Retail and
    State = VA assign to Bob
    State = MD assign to Jim
    State = CA assign to Karen
    Industry = Commercial and
    State = NY assign to Lam
    State = VA assign to John
    State = CA assign to Karen.....

    Hi Bob,
    In continuation to your reply, can we add more complexity to these rules like with industry and state, if i want to assign a lead based on to the tyoe of product and the revenue of the company.
    e.g.
    Industry - Retail
    State - MA
    Product - ABC & Revenue of company - $100M - assign it to Sales person 1
    OR
    Industry - Retail
    State - MA
    Product - ABC & Revenue of company - $1M - assign it to Sales person 2
    Regards
    Nisman

  • Documentation about Rule-Based Account Assignment Distribution

    Hello dear PSM Experts,
    In our project (using ECC6), we want to split in a PO the FM account assignment between several lines, according to the fund type of the initial account assignment (based from an referenced earmarked document).
    After looking in the customizing and in ECC6 PSM Public Sector Management
    Release Notes, we are investigating Documentation about Rule-Based Account Assignment Distribution.
    Can anybody provides me a link to a documentation, if any?
    Otherwise, I would greatly appreciate any help to use this functionality.
    Regards,
    François

    Hello,
    I am configuring this functionality currently in an ECC 6.0 environment for a public sector organization and have had an issue with the FMSPLITMAINT transaction.  When attempting to create a rule the 'create' 'copy' 'change' options are not available.
    Configuration has already been completed for Field Profiles, Action Profiles, & Rule Groups.
    The olny documentation I have been able to find was an appendix to the SAP Grants Management - Grantee course.
    Please respond if you there is any additional documentation which may be of assistance.  I have already tried notes: 991661 and 991962.
    Thank you

  • Questions on Rules-Based ATP and Purchase Requisitions for STOs

    Hello experts,
    We are working on rules-based ATP configuration and have several questions about the functionality.  Iu2019m hoping that some of you are using this functionality and can help give us direction.
    In our environment we have multiple distribution centers and multiple manufacturing plants.  We want to confirm sales orders against stock and production orders in any of those plants, depending on the locations that have stock or planned production.  For example, we will place a sales order against plant A.  If there is not enough stock in plant A then rules-based ATP will use location determination to check in plant B, then C.  The scope of check on the ATP check will include stock and released production orders.  We will configure plant A as the u201Cconsolidation locationu201D so if stock is found in plants B or C then stock transport orders will automatically be created to move the stock to plant A before shipping to the customer.
    We have configured rules-based ATP and this functionality is working well in our Development system.  The ATP check is executed and uses the rules-based ATP to find eligible stock in other plants.  The system is also creating purchase requisitions to move the stock to the consolidation plant. 
    Our first concern is that there doesnu2019t appear to be any firm linkage between the sales order and the resulting purchase requisition.  For example, if we create sales order 123 for plant A and the rules-based ATP finds stock in plant B it automatically creates a purchase requisition 987 to move the stock from plant B to plant A.  However, there doesnu2019t appear to be a linkage between sales order 123 and purchase requisition 987.  For instance, if we delete sales order 123 the purchase requisition doesnu2019t get deleted. 
    Our second concern is that the quantity on the purchase requisition can still be confirmed against later sales orders.  For example, say the above scenario resulted in a purchase requisition 987 that consumed all the stock available in plant B.  We then create a second sales order 456 for the same product.  Plant A is out of stock so the rules-based ATP looks in plant B.  We would expect that plant B would also not have any stock because itu2019s all been consumed by the purchase requisition.  Instead, the system creates a second purchase requisition to move quantity from plant B to plant A.  Itu2019s as if the system doesnu2019t realize that the purchase requisition 987 is already planning to move stock out of plant B.
    Does anyone have any thoughts or suggestions on these two scenarios?  Is there a way to configure the system so there is a hard linkage between the sales order and the purchase requisition so that if the sales order is deleted then the purchase requisition is also deleted?  Should ATP realize that purchase orders are consuming inventory and not allow later sales orders to confirm against that same inventory?  Any advice or experience would be greatly appreciated.
    Thanks,
    David Eady
    Application Delivery Team Lead
    Propex, Inc.

    Hi,
    The scheduling is done in SCM, and from there, whenever the RBA is triggered, the calculation is done always with the old route in SCM. Until you get back to R/3 this is when your route is determined. But the ATP check is always with the original route. So the idea would be that you change the values of the route while still in APO, this is possible via the user exit. Should be done in scheduling in APO.  
    Hope this information is helpful.
    Regards,
    Tibor

  • Rule based location substitution:

    Hi,
    We have the following scenario:
    Material  : X
    Location : A ->B -> C  (B is a substitute of A and C is a substitute of B)
    ATP position      A = 2
                               B = 2
                               C = 2
    Sales Order of 10 raised at A : Location substitution happens and partially confirms sales order with 6 quantities.
    Therefore new ATP position is as below
              A = 0
                               B = 0
                               C = 0
    Now we want to give confirmation(of 4 quantities) based on external lead time to vendor. This will be based on source-list in R/3.  The PO  to vendor can be based on any of the location based on  source-list. i.e. it can be from A,B, or C. The source list is dynamic and can change based on business decision.  i.e today  the PO plant can be A. It can change tomorrow and be B.
    What is the best possible way to model the scenario?
    The constraint is that we are not allowed to change the checking rule.
    Regards
    Sameer

    Hi Sameer,
    Your scenario is rules based location substitution and then MRP based upon source list.
    No matter whether plant A or plant B or C  comes on line item , demand should be placed on plant suggested by source list.
    I think you can control it using source list field MRP "1     Record relevant to MRP"  and planning Plant.
    Example : sales order line item plant is A and in source list planning plant is B then , PO will be created at plant B.
    Then it will be a two step MRP  process:
    1.P.O from Plant B to vendor
    2. STO from plant B to A.
    Thanks
    Pavan Verma

  • Rule Based GATP

    Dear All,
    I'm using Rule Based ATP check for Sales Order.
    When I conduct ATP simulation in APO (SCM 4.0), the results are perfect. the check triggers the rule and displays schedule lines as per rule. (Location substitution).
    But when I conduct availability check in Sales order (R/3 4.7), the results are completely different. For Sales Order at Warehouse, it confirms all the requirement irespective of transportation lane lead time from plant to warehouse. Moreover, when I save the sales order, the PReqs created at Warehouse are in future instead of PReqrel in past. (compared to requirement date). Ideally, the Preqs should have same date as that of requirement date and Preqrel should be created after deducting Transporation lead time and GR/GI time.
    I would like to know, what parameters in Sales Order document are used to trigger the rule and how is it different from ATP simulation.
    Thank you in advance.
    Regards,
    Bipin K Umarale

    Hi
    Take a look at the following SAP Note, it gives all the FAQ's on ATP configuration in R/3, hopefully it will help you with your problem:
    [Note 547512 - FAQ: Customizing of the ATP in R/3|https://websmp230.sap-ag.de/sap(bD1lbiZjPTAwMQ==)/bc/bsp/spn/sapnotes/index2.htm?numm=547512&nlang=EN&smpsrv=https%3a%2f%2fwebsmp208%2esap-ag%2ede]
    Regards
    Ian

  • Rules based ATP to Consider Checking Horizon

    Hello SAP Experts,
    I am a novice to GATP functionality. I have one scenario tried to work on it and Standard SAP system doesn't seem to work this way.
    Rules based ATP for location Determination is used for order confirmation.
    Loc A   Mat 1  500 Pcs
    Loc B   Mat 1  200 Pcs
    Rule 1 LocA --> Loc B
    Order comes in for Mat 1 at location A for 1000 Pcs and confirmation was given for 700 Pcs out of which 200 Pcs will be sourced from Loc B. I'm trying to confirm remaining 300 at the end of Checking horizon as per Customer requirement. Please suggest a way to acheive this.
    Regards,
    Priyanka

    Hi,
    You will start from ECC, Material master MRP3 view   MARC-WZEIT   Total replenishment lead time (in workdays).
    In APO , SPRO--GATP-Maintain check control - make sure that ATP group and business event example 01 and A should have "consider Checking horizon" .
    RLT from R3 will become Checking horizon in product master in APO. In APO product master ATP tab CHKHOR is the field name.
    Make sure that in the rules -- calculation profile -allowed delay is not maintained or should not be less than RLT /Checking horizon.
    Thanks,
    Pavan Verma

  • Rule based ATP is not working for Components

    Hi All,
    Our requirement is to do availability check through APO for Sales order created in ECC,so we are using gATP.
    Requirement: We are creating salesorder for BOM header (Sales BOM) and avaialbility check should happen for components i.e. Product avalaibility & Rule based substitution.
    Issue: Product availiabilty is working for components but rules based substituion is working,  mean Rules are not getting determind for components.
    Settings:
    - Header doesnot exist in APO and compnents do exist in APO
    - Availability check is not enabled for header item category and enabled for Item category for components
    - Rules have been created for Components in APO
    - Rule base ATP is activated in Check instructions
    We have also tried MATP for this i.e. PPM created in APO but still didn't get the desired result.
    If we create salesorder for the component material directly then Rule based ATP is happening, so for components Rule based ATP is not working.
    How do we enable enable Rulesbased ATP for components, i mean is there any different way to do the same.
    Thanks for help.
    Regards,
    Jagadeesh

    Hi Jagdeesh,
    If you are creating BOM in ECC and CIFing PPM of FG/Header material to APO, I think you need to CIF Header material, too, with material integration model.
    Please include header material in you integration models for material, SO and ATP check as well.
    For component availability check, you can use MATP; but for MATP, FG should be in APO. You need not to CIF any receipts of FG (stock, planned orders, POs etc), so that MATP will be triggered directly. Then maintaining Rules for RMs will enable to select available RMs according to the rule created.
    Regards,
    Bipin

  • Dynamic Rule based implementation in PL/SQL

    Hi,
    We are trying to implement a dynamic rule based application in Oracle 9i. Its simple logic where we store expressions as case statments and actions seperated by commas as follows.
    Rule: 'Age > 18 and Age <65'
    True Action: 'Status = ''Valid'' , description = ''age in range'''
    False Action: 'Status =''Invalid'', Description=''Age not in range'''
    Where Age,Status, description are all part of one table.
    One way of implementing this is fire rule for each record in the table and then based on true or false call action as update.
    i.e
    select (case when 'Age > 18 and Age <65' then 1 else 0 end) age_rule from tableX
    (above query will in in a cursor xcur)
    Then we search for
    if age_rule = 1 then
    update tablex set Status = ''Valid'' , description = ''age in range'' where id=xcur.id;
    else
    update tablex set Status =''Invalid'', Description=''Age not in range'' where id=xcur.id;
    end if;
    This method will result in very slow performance due to high i/o. We want to implement this in collection based method.
    Any ideas on how to dynamically check rules and apply actions to collection without impact on performance. (we have nearly 3million rows and 80 rules to be applied)
    Thanks in advance

    Returning to your original question, first of all, there is a small flaw in the requirements, because if you apply all the rules to the same table/cols, than the table will have results of only last rule that was processed.
    Suppose rule#1:
    Rule: 'Age > 18 and Age <65'
    True Action: 'Status = ''Valid'' , description = ''age in range'''
    False Action: 'Status =''Invalid'', Description=''Age not in range'''
    and Rule#2:
    Rule: 'Name like ''A%'''
    True Action: 'Status = 'Invalid'' , description = ''name begins with A'''
    False Action: 'Status =''Invalid'', Description=''name not begins with A'''
    Then after applying of rule#1 and rule#2, results of the rule#1 will be lost, because second rule will modify the results of the first rule.
    Regarding to using collections instead of row by row processing, I think that a better approach would be to move that evaluating cursor inside an update statement, in my tests this considerably reduced processed block count and response time.
    Regarding to the expression filter, even so, that you are not going to move to 10g, you still can test this feature and see how it is implemented, to get some ideas of how to better implement your solution. There is a nice paper http://www-db.cs.wisc.edu/cidr2003/program/p27.pdf that describes expression filter implementation.
    Here is my example of two different methods for expression evaluation that I've benchmarked, first is similar to your original example and second is with expression evaluation moved inside an update clause.
    -- fist create two tables rules and data.
    drop table rules;
    drop table data;
    create table rules( id number not null primary key, rule varchar(255), true_action varchar(255), false_action varchar(255) );
    create table data( id integer not null primary key, name varchar(255), age number, status varchar(255), description varchar(255) );
    -- populate this tables with information.
    insert into rules
    select rownum id
    , 'Age > '||least(a,b)||' and Age < '||greatest(a,b) rule
    , 'Status = ''Valid'', description = ''Age in Range''' true_action
    , 'Status = ''Invalid'', description = ''Age not in Range''' false_action
    from (
    select mod(abs(dbms_random.random),60)+10 a, mod(abs(dbms_random.random),60)+10 b
    from all_objects
    where rownum <= 2
    insert into data
    select rownum, object_name, mod(abs(dbms_random.random),60)+10 age, null, null
    from all_objects
    commit;
    -- this is method #1, evaluate rule against every record in the data and do the action
    declare
    eval number;
    id number;
    data_cursor sys_refcursor;
    begin
    execute immediate 'alter session set cursor_sharing=force';
    for rules in ( select * from rules ) loop
    open data_cursor for 'select case when '||rules.rule||' then 1 else 0 end eval, id from data';
    loop
    fetch data_cursor into eval, id;
    exit when data_cursor%notfound;
    if eval = 1 then
    execute immediate 'update data set '||rules.true_action|| ' where id = :id' using id;
    else
    execute immediate 'update data set '||rules.false_action|| ' where id = :id' using id;
    end if;
    end loop;
    end loop;
    end;
    -- this is method #2, evaluate rule against every record in the data and do the action in update, not in select
    begin
    execute immediate 'alter session set cursor_sharing=force';
    for rules in ( select * from rules ) loop
    execute immediate 'update data set '||rules.true_action|| ' where id in (
    select id
    from (
    select case when '||rules.rule||' then 1 else 0 end eval, id
    from data
    where eval = 1 )';
    execute immediate 'update data set '||rules.false_action|| ' where id in (
    select id
    from (
    select case when '||rules.rule||' then 1 else 0 end eval, id
    from data
    where eval = 0 )';
    end loop;
    end;
    Here are SQL_TRACE results for method#1:
    call count cpu elapsed disk query current rows
    Parse 37 0.01 0.04 0 0 0 0
    Execute 78862 16.60 17.50 0 187512 230896 78810
    Fetch 78884 3.84 3.94 2 82887 1 78913
    total 157783 20.46 21.49 2 270399 230897 157723
    and this is results for method#2:
    call count cpu elapsed disk query current rows
    Parse 6 0.00 0.00 0 0 0 0
    Execute 6 1.93 12.77 0 3488 170204 78806
    Fetch 1 0.00 0.00 0 7 0 2
    total 13 1.93 12.77 0 3495 170204 78808
    You can compare this two methods using SQL_TRACE.

  • Document rule based classification

    from the example in oracle text developers guide i tried to build a rule based document classification, using the code given below:
    create or replace package classifier as
    procedure this;
    end;
    show errors
    create or replace package body classifier as
    procedure this
    is
    v_document     blob;
    v_item          number;
    v_doc           number;
    begin
    for doc in (select document_id, content from documents)
         loop
              v_document :=doc.content;
              v_item:=0;
              v_doc:=doc.document_id;
              for c in (select category_id, category_name from docs_cats_rule_based_class
                   where matches(query,v_document)>0)
              loop
                   v_item:=v_item +1;
                   insert into doc_cat_rule_based_class values (doc.document_id, category_id);
              end loop;
         end loop;
    end this;
    end;
    show errors
    exec classifier.this
    this gives the following errors:
    package classifier Compiled.
    line 5: SQLPLUS Command Skipped: show errors
    package body Compiled.
    line 32: SQLPLUS Command Skipped: show errors
    Error starting at line 33 in command:
    exec classifier.this
    Error report:
    ORA-04063: package body "STARDOC.CLASSIFIER" has errors
    ORA-06508: PL/SQL: could not find program unit being called: "STARDOC.CLASSIFIER"
    ORA-06512: at line 1
    i think i am missing some grant to package. please help!

    What version of Oracle are you using? Did you create the required tables and index in the earlier steps? What did you run it from? It appears that you did not run it from SQL*Plus. Please see the following demonstration that shows that it works fine on Oracle 10g when run from SQL*Plus with minimal privileges. I did not use any data.
    SCOTT@10gXE> CREATE USER stardoc IDENTIFIED BY stardoc
      2  /
    User created.
    SCOTT@10gXE> GRANT CONNECT, RESOURCE TO stardoc
      2  /
    Grant succeeded.
    SCOTT@10gXE> CONNECT stardoc/stardoc
    Connected.
    STARDOC@10gXE>
    STARDOC@10gXE> create table news_table
      2    (tk    number primary key not null,
      3       title varchar2(1000),
      4       text  clob)
      5  /
    Table created.
    STARDOC@10gXE> create table news_categories
      2    (queryid  number primary key not null,
      3       category varchar2(100),
      4       query      varchar2(2000))
      5  /
    Table created.
    STARDOC@10gXE> create table news_id_cat
      2    (tk         number,
      3       category_id number)
      4  /
    Table created.
    STARDOC@10gXE> create index news_cat_idx on news_categories (query)
      2  indextype is ctxsys.ctxrule
      3  /
    Index created.
    STARDOC@10gXE> create or replace package classifier
      2  as
      3    procedure this;
      4  end classifier;
      5  /
    Package created.
    STARDOC@10gXE> show errors
    No errors.
    STARDOC@10gXE> create or replace package body classifier
      2  as
      3    procedure this
      4    is
      5        v_document    clob;
      6        v_item        number;
      7        v_doc            number;
      8    begin
      9        for doc in (select tk, text from news_table)
    10        loop
    11          v_document := doc.text;
    12          v_item := 0;
    13          v_doc  := doc.tk;
    14          for c in
    15            (select queryid, category from news_categories
    16             where matches (query, v_document) > 0)
    17          loop
    18             v_item := v_item + 1;
    19             insert into news_id_cat values (doc.tk,c.queryid);
    20          end loop;
    21        end loop;
    22    end this;
    23  end classifier;
    24  /
    Package body created.
    STARDOC@10gXE> show errors
    No errors.
    STARDOC@10gXE> exec classifier.this
    PL/SQL procedure successfully completed.
    STARDOC@10gXE>

  • Re: Oracle 8i (8.1.7.4) Rule based v/s Cost based

    Hi,
    I would like to know the advantages/disadvantages of using RULE based optimizer v/s COST based optimizer in Oracle 8i. We have a production RULE based database and are experiencing performance issues on some queries sporadically.
    TKPROF revealed:
    call       count       cpu    elapsed             disk                  query                current              rows
    Parse        0      0.00       0.00                0                      0                      0                      0
    Execute      3     94.67    2699.16            1020421            5692711            51404               0
    Fetch       13    140.93    4204.41             688482             4073366            0                      26896
    total       16       235.60    6903.57           1708903            9766077            51404               26896
    Please post your expert suggestions as soon as possible.
    Thanks and Regards,
    A

    I think the answer you are looking for is that Rule Based optimizer is predictive, but Cost Based optimizer results may vary depending on statistics of rows, indexes, etc. But at the same time, you can typically get better speed for OLTP relational databases with CBO, assuming you have correct statistics, and correct optimizer settings set.

  • Improving performace for a Rule Based Optimizer DB

    Hi,
    I am looking for information on improving the current performance of an ancient 35GB Oracle 7.3.4 using RULE based optimizer mode. It is using 160 MB SGA and the physical memory on the system is 512MB RAM.
    As of now, all the major tasks which take time, are run after peak hours so that the 130 user sessions are not affected significantly.
    But recently am told some procedures take too long to execute ( procedure has to do with truncating tables and re-populating data into it ) and I do see 54% of the pie chart for WAITS are for "sequential reads" followed by "scattered reads" of 36%. There are a couple of large tables of around 4GB in this DB.
    Autotrace doesn't help me much in terms of getting an explain plan of slow queries since COST option doesnt show up and am trying to find ways of improving the performance of DB in general.
    Apart from the "redo log space requests" which I run into frequently (which btw is something I am trying to resolve ..thanks to some of you) I dont see much info on exactly how to proceed.
    Is there any info that I can look towards in terms of improving performance on this rule based optimizer DB ? Or is identifying the top sql's in terms of buffer gets the only way to tune ?
    Thank you for any suggestions provided.

    Thanks Hemant.
    This is for a 15 minute internal under moderate load early this morning.
    Statistic                   Total        Per Transact Per Logon    Per Second 
    CR blocks created                    275          .95         5.19          .29
    Current blocks converted fo           10          .03          .19          .01
    DBWR buffers scanned               74600       258.13      1407.55        78.44
    DBWR free buffers found            74251       256.92      1400.96        78.08
    DBWR lru scans                       607          2.1        11.45          .64
    DBWR make free requests              607          2.1        11.45          .64
    DBWR summed scan depth             74600       258.13      1407.55        78.44
    DBWR timeouts                        273          .94         5.15          .29
    OS Integral shared text siz   1362952204   4716097.59  25716079.32   1433177.92
    OS Integral unshared data s    308759380   1068371.56   5825648.68    324668.12
    OS Involuntary context swit       310493      1074.37      5858.36       326.49
    OS Maximum resident set siz       339968      1176.36      6414.49       357.48
    OS Page faults                      3434        11.88        64.79         3.61
    OS Page reclaims                    6272         21.7       118.34          6.6
    OS System time used                19157        66.29       361.45        20.14
    OS User time used                 195036       674.87      3679.92       205.09
    OS Voluntary context switch        21586        74.69       407.28         22.7
    SQL*Net roundtrips to/from         16250        56.23        306.6        17.09
    SQL*Net roundtrips to/from           424         1.47            8          .45
    background timeouts                  646         2.24        12.19          .68
    bytes received via SQL*Net        814224      2817.38     15362.72       856.18
    bytes received via SQL*Net         24470        84.67        461.7        25.73
    bytes sent via SQL*Net to c       832836      2881.79     15713.89       875.75
    bytes sent via SQL*Net to d        42713        147.8       805.91        44.91
    calls to get snapshot scn:         17103        59.18        322.7        17.98
    calls to kcmgas                      381         1.32         7.19           .4
    calls to kcmgcs                      228          .79          4.3          .24
    calls to kcmgrs                    20845        72.13        393.3        21.92
    cleanouts and rollbacks - c           86           .3         1.62          .09
    cleanouts only - consistent           40          .14          .75          .04
    cluster key scan block gets         1051         3.64        19.83         1.11
    cluster key scans                    376          1.3         7.09           .4
    commit cleanout failures: c           18          .06          .34          .02
    commit cleanout number succ         2406         8.33         45.4         2.53
    consistent changes                   588         2.03        11.09          .62
    consistent gets                   929408      3215.94        17536        977.3
    cursor authentications              1746         6.04        32.94         1.84
    data blocks consistent read          588         2.03        11.09          .62
    db block changes                   20613        71.33       388.92        21.68
    db block gets                      40646       140.64       766.91        42.74
    deferred (CURRENT) block cl          668         2.31         12.6           .7
    dirty buffers inspected                3          .01          .06            0
    enqueue conversions                  424         1.47            8          .45
    enqueue releases                    1981         6.85        37.38         2.08
    enqueue requests                    1977         6.84         37.3         2.08
    execute count                      20691         71.6        390.4        21.76
    free buffer inspected               2264         7.83        42.72         2.38
    free buffer requested             490899      1698.61      9262.25       516.19
    immediate (CR) block cleano          126          .44         2.38          .13
    immediate (CURRENT) block c          658         2.28        12.42          .69
    logons cumulative                     53          .18            1          .06
    logons current                         1            0          .02            0
    messages received                    963         3.33        18.17         1.01
    messages sent                        963         3.33        18.17         1.01
    no work - consistent read g       905734      3134.03     17089.32        952.4
    opened cursors cumulative           2701         9.35        50.96         2.84
    opened cursors current               147          .51         2.77          .15
    parse count                         2733         9.46        51.57         2.87
    physical reads                    490258      1696.39      9250.15       515.52
    physical writes                     2265         7.84        42.74         2.38
    recursive calls                    37296       129.05        703.7        39.22
    redo blocks written                 5222        18.07        98.53         5.49
    redo entries                       10575        36.59       199.53        11.12
    redo size                        2498156      8644.14     47135.02      2626.87
    redo small copies                  10575        36.59       199.53        11.12
    redo synch writes                    238          .82         4.49          .25
    redo wastage                      104974       363.23      1980.64       110.38
    redo writes                          422         1.46         7.96          .44
    rollback changes - undo rec            1            0          .02            0
    rollbacks only - consistent          200          .69         3.77          .21
    session logical reads             969453      3354.51     18291.57       1019.4
    session pga memory              35597936    123176.25    671659.17     37432.11
    session pga memory max          35579576    123112.72    671312.75      37412.8
    session uga memory               2729196      9443.58     51494.26      2869.82
    session uga memory max          20580712     71213.54    388315.32     21641.13
    sorts (memory)                      1091         3.78        20.58         1.15
    sorts (rows)                       12249        42.38       231.11        12.88
    table fetch by rowid               57246       198.08      1080.11         60.2
    table fetch continued row            111          .38         2.09          .12
    table scan blocks gotten          763421       2641.6     14404.17       802.76
    table scan rows gotten          13740187      47543.9    259248.81     14448.15
    table scans (long tables)            902         3.12        17.02          .95
    table scans (short tables)          4614        15.97        87.06         4.85
    total number commit cleanou         2489         8.61        46.96         2.62
    transaction rollbacks                  1            0          .02            0
    user calls                         15266        52.82       288.04        16.05
    user commits                         289            1         5.45           .3
    user rollbacks                        23          .08          .43          .02
    write requests                       331         1.15         6.25          .35Wait Events :
    Event Name                       Count         Total Time    Avg Time    
    SQL*Net break/reset to client                7             0             0
    SQL*Net message from client              16383             0             0
    SQL*Net message from dblink                424             0             0
    SQL*Net message to client                16380             0             0
    SQL*Net message to dblink                  424             0             0
    SQL*Net more data from client                1             0             0
    SQL*Net more data to client                 24             0             0
    buffer busy waits                          169             0             0
    control file sequential read                55             0             0
    db file scattered read                   74788             0             0
    db file sequential read                 176241             0             0
    latch free                                6134             0             0
    log file sync                              225             0             0
    rdbms ipc message                           10             0             0
    write complete waits                         4             0             0I did enable the timed_stats for the session but dont know why the times are 0's. Since I cant bounce the instance until weekend, cant enable the parameter in init.ora as well.

  • Partitioning on Oracle 8i (Rule Based vs. Cost Based)

    At my current engagement, we are using Oracle Financials 11.0.3 on Oracle 8.0.6. The application uses rule-based optimizer. The client wants to implement Oracle partitioning. With this in mind, we are concerned about possible performance issues that the implementation of partitioning may cause since RBO does not recognize it.
    We agree that the RBO will see a non-partitioned table the same as a partitioned. In this scenario where you gain the most is with backup/recoverability and general maintenance of the partitioned table.
    Nevertheless, we have a few questions:
    When implementing partitions, will the optimizer choose to go with Cost base vs. Rule base for these partitioned tables?
    Is it possible that the optimizer might get confused with this?
    Could it degrade performance at the SQL level?
    If this change from RBO to CBO does occur, the application could potential perform poorly because of the way it has been written.
    Please provide any feedback.
    Thanks in advance.

    If the CBO is invoked when accessing these tables, you may run into problems.
    - You'll have to analyze your tables & ensure that the statistics are kept up to date.
    - It's possible that any SQL statements which invoke the CBO rather than the RBO will have different performance characteristics. The SYSTEM data dictionary tables, for example, must use the RBO or their performance suffers dramatically. Most of the time, the CBO beats the RBO, but applications which have been heavily tuned with the RBO may have problems with the CBO.
    - Check your init.ora to see what optimizer mode you're in. If you're set to CHOOSE, the CBO will be invoked whenever statistics are available on the table(s) involved. If you choose RULE, you'll only invoke the CBO when the RBO encounters situations it doesn't have rules for.
    Justin

Maybe you are looking for