Rec/client parameter & Performance Impact

Hi all,
We have been asked by our audit team to set the parameter Rec/client=300 (our production client) in our production system.
But when we read few forums & notes, we feel that setting this parameter Rec/client=300 will definetly impact the performance.
So before setting this parameters we need to evaluate the  performance impact in our ECC 6.0 system by setting this parameter, can anyone help us to evaluate the performace impact of this parameter settings in our below system
- ECC 6.0 / oracle 10g / HPUX - 64 bit / 3500 users / 2.5 terabyte data
Thanks
Senthil

Hello
Normally only customizing tables should have the logging flag. You can verify it in SE13 -> <table> -> Log data changes.
To list all tables with logging on, you can use this select statement:
SQL> select tabname, protokoll from sapr3.dd09l where protokoll = 'X';
As long as only so called customizing tables are logged, you should be fine. If you have some heavy traffic Z* tables, then two things might happen:
- performance might suffer
- the logging table DBTABLOG will explode
So please make sure only the necessary tables are logged and if possible test on the QAS system, if the logging leads to performance problems.
Best regards
Michael

Similar Messages

  • Rec/client paramater for logging of table changes

    Hello all,
    I know that it was asked many times on many forums how to activate the log of a specific tables and most of the answers were about disabling every table activated by default through DD09L table and then use SE13 to enable the ones we want.
    But on the documentation on SPRO, there is this mention about "rec/client" parameter :
    Set the rec/client profile parameter to one of the following values :
    -- mmm,nnn,ppp,...: logs client-specific tables in the named clients
    If I understand well, using this value I can focus the logs on a specific table, am I right ?
    If so, how do I use it ?
    I guess "nnn" is the client number. So what does "mmm", "nnn", "ppp" means ?
    Thanks & regards,

    Thank you for your answer.
    Never mind my question, I misread the description, didn't notice the "s" on "in the named clients". I should have changed the language to my native language
    I misunderstood and thought that we could actually put the name of a table directly in this parameter.
    So then, my question is, how would do to actually activate the logs only on a specific table because looking at my production server, I can see that there is about 30 000 tables activated by default to be logged.
    Thanks

  • Rec/Client

    Hi Gurus,
    We want to activate the parameter REC/CLIENT in our production Environment.We know that activating this parametre will create a huge logs
    Please any body clarify the below doubts:
    1) where will be the logs stored at OS level
    2) What would be the amount of Logs generated for one day, we have more than 700 Users and 74000 thousand table in our Production system.
    3) We want to activate the logs for only 3 tables like USR02...
    but when I saw in table DD09L there were more than 28000 tables, already activated log data changes, So we want to deactivate this option from 28000 tables and activate for only 3 tables.Please tell me the process step by step.
    4) If we want to deacivate the loging for the above said tales, what is the process (I want to know is there any other process with out deactivating the REC/CLIENT parameter because this need system reboot)
    Thanks & Reagrds,
    Veda Gandham

    > Hi
    >
    > You can configure the client that should be logged by
    > the
    > R/3 profile parameter rec/client (see documentation
    > in your R/3 system)
    >
    > Generally, there should not be too many entries
    > logged, because
    > logging is only set for cusomizing tables (by
    > default) and
    > customizing should not happen too often.
    >
    > Also see note 41300, 1916
    >
    > Reward points if useful
    Hi, thanks for your reply. Looking at S Note 1916 this states that all servers must have the same setting for logging to work correctly, which is what i wanted to know. However there is no explanation of what occurs if settings are different or otherwise, anyone know the background on how this param (rec/client) works?
    Thanks
    Ash

  • Active Table Logging T000 performance impact

    Hi fellow SAP experts,
    I need some advice on system performance impact when switching on Table Logging for T000 - configuration in production please?
    We have decided to turn on Table Logging for auditing purposes, only allowing developer config in production following a volume of evidence being supplied.
    I need to know how much this activation is going to impact the performance of the companies production environments, how much storage, memory, performance, etc. this function is going to consume and how much of the above consumables I need to cater for now and in the future?
    We have a Dual Track environment, BAU want to switch on Table Logging for fix on fail, I want to swich it on for Project deliveries.
    Please advise, with referencing if possible?
    Thank you kindly
    Paul

    Hi Paul,
    There has been a constant debate whether table logging affects the System performance, especially in Produciton environment. Please see my comments below :
    1) To turn on logging for table T000, you will have to activate the parameter rec/client, with values for one client or all the clients in the System depending on your requirement.
    2) This parameter setting will not only log changes to T000, but also for over 28000 tables.
    3) But these are customizing tables which usually contain a relatively small amount of data which is changed occasionally.
    4) After activating, if you suddenly find performance issues, you can check which tables are causing issues via transaction SCU3.
    5) You can go to transaction SE13 and deactivate logging for a table, if you find too many entries for any particular table in SCU3.
    So, logging tables doesn't necessarily impact performance. Hope this helps. Please refer to SAP Note-608835 related to this.
    Best Regards,
    Savitha

  • Performance impact using nested tables and object

    Hi,
    Iam using oracle 11g.
    While creating a package, iam using lot of nested tables created based on objects which will be passed between multiple functions in the package..
    Will it have any performance impact since all the data is stored in the memory.
    How can i measure the performance impact when the data grows ?
    Regards,
    Oracle User
    Edited by: user9080289 on Jun 30, 2011 6:07 AM
    Edited by: user9080289 on Jun 30, 2011 6:42 AM

    user9080289 wrote:
    While creating a package, iam using lot of nested tables created based on objects which will be passed between multiple functions in the package.. Not the best of ideas in general, in PL/SQL. This is not client code that can lay sole claim to most of the memory. It is server code and one of many server processes that need to share the available resources. So capitalism is fine on a client, but you need socialism on the server? {noformat} ;-) {noformat}
    Will it have any performance impact since all the data is stored in the memory.Interestingly yes. Usually crunching data in memory is better. In this case it may not be so. The memory used is the most expensive memory Oracle can use - the PGA. Private process memory. This means each process copy running that code, will need lots of memory.
    If you're not passing the data structures by reference, it means even bigger demands on memory as the data structure needs to be copied into the call stack and duplicated.
    The worse case scenario is that such code consumes so much free server memory, and make such huge demands on having that in pysical memory, it trashes memory management as the swap daemons are unable to keep up with the demand of swapping virtual memory pages into and out of memory. Most CPU time is spend by the swap daemons.
    I have seen servers crash due to this. I have seen a single PL/SQL process causing this.
    How can i measure the performance impact when the data grows ?Well, you need to look at the impact of your code on PGA memory. It is not SQL performance or I/O performance that is a factor - just how much private process memory your code needs in order to execute.

  • Regarding performance impact if I do DB accessing coding in comp Controller

    Hi ,
    This is my project requirement, I have to use some com compoment which in turn fetches data from the database. I am using a java com bridge tool to do this. This tool is generating the java proxy classes for the VB com component.
    I am using java proxy classes( This class files are using JNI to connect to VB COM compnent and fetch the data from DB) in my webdynpro component controller.
    The architecture is aas below
    WEBDYNPRO    >>   JAVA Classes object( generated by the JAVA- COM bridge   tool )                         >>   JAVA-COM bridge  tool >> VB COM+ Component   >> SQL server.
    The issue
       Performance :-   first time it is OK but for Consecutive calls the application is going down very visibly and after 4 iteration it hangs . When I look at the log I am getting this
    Message : Exception occured during processing of Web Dynpro application com/oreqsrch/com.oreqsrchapp.OReqSrchApp.
    The causing exception is nested.
    [EXCEPTION]
    com.sap.tc.webdynpro.services.session.LockException: Thread SAPEngine_Application_Thread[impl:3]_36 failed to acquire exclusive lock on client session ClientSession(id=(J2EE9536400)ID1120562150DB11245826542790956137End_1159630423). Existing locks: LockingManager(ThreadName:SAPEngine_Application_Thread[impl:3]_36, exclusive client session lock:
    ClientSessionLock(SAPEngine_Application_Thread[impl:3]_9), shared client session locks: ClientSessionSharedLockManager([]), app session locks: ApplicationSessionLockManager([]), current request: com/oreqsrch/com.oreqsrchapp.OReqSrchApp).
    Hint: Take a thread dump of the server node to find the blocking thread that causes the problem.
    Is this issue because I have return the code data access code in the component controller rather wrting in some beans ?
    My questions regarding
    What would the performance impact if write the DB access code in the webdynpro component controller rather than writing in a bean or an EJB?( I know ideally DB access code has to write in Bean or EJB ).
    Please address  this with respedct to performance  point of view .
    thanks
    pkiran

    Hi Both,
    Thanks for the reply.
    Yes they are closed and set it to null;
    Connection max and mini properties are controlled at COM+ components in VB.
    Since I am using COM - JAVA bridge,  I am just invoking the methods defined ijn the VB code  thru the bridge tool. all the objects which are retrieving the data are closed and nullify it.
    My question is
    if I write DB access code in component control instead in EJB or Java bean, will there be any performance issue ?
    regards
    pkiran

  • Change Tracker Performance Impact

    MDM: 7.1
    CE: 7.2
    ERP: 6.0 EHP4
    Hi,
    We are currently using CE/BPM based central master data management. A custom application is being developed for collaborative master data authoring.
    As part of the MDM configuration, the client wishes to track changes on MDM records for audit purposes. We are looking at the MDM Change Tracking facility provided by SAP but not sure about the performance impact it will have on the MDM server.
    We have over 300 attributes for the object we wish to track for changes. Not all attributes will change all the time but it is expected that the overall number of changes every month will be over 1000, each change including approx 20-30 fields. The number of users is expected to be approx 15 initially but will increase over time.
    I have seen on SCN forums people talking about potential performance degrade by enabling change tracking. Has anyone actually experienced performance degrade due to enabling change tracking for MDM records? If so, have you tried any means to keep the impact low e.g. by allocating more resources to MDM server?
    Thanks and regards,
    Shehryar
    Edited by: Shehryar Khan on Dec 2, 2010 1:39 PM

    Hi,
    Change histoey does have impact on system performance. this can be controlled via regular archiving of Change History database. Lets say, Change history table data older than 3 months can be stored in another repository.
    Second option could be to export all Change History data(Older than 1-2 days) to a BI system(using regular scheduled job), and have change hisotry report there. It will bring drastci change in MDM System performance. 1-2 days data can be viewed from MDM Change history database itself.
    Regards,
    Prashant
    Edited by: Prashant Malik on Dec 3, 2010 1:58 AM

  • SSL Performance Impact

    Hi,
    We are planning to move to SSL based communication for our web based application as well as internal communication between various components using Sockets. Are there any studies done, which tell us about the performance impact of using SSL. How much degradation in performance/throughput should we expect? Are there any guidelines available which tell us how to minimize the performance impact?
    Thanks

    I did some testing on that last year over a 1.5Mb Internet connection. I found that the throughput degradation for long transfers was about 1::3. The study itself was for a client, and I can't release it to the public until at least 12 months have expired.
    The more important issue is the extra CPU load on the server. Beyond a certain number of connections, which depends on the hardware of course, you may want to investigate a hardware crypto accelerator.

  • Performance impact of using Web Services?

    As BEA and other vendors continue to add Web Services support
    to their enterprise software, what is your plan for
    quantifying the performance impact and the functional
    correctness of using web services before going live with the
    final application?
    Empirix is hosting a free one hour web event discussion on
    web services testing and automated web services testing
    solutions on Thursday, January 17, 2-3pm Eastern time.
    To sign-up for this web event or learn about other web
    events being offering by Empirix this month, go to:
    http://webevents.empirix.com
    For your convenience, here is the complete abstract:
    The advent of web services has brought the promises of
    integrating multiple software applications from
    heterogeneous networks and for exchanging information
    from vendor-to-vendor or vendor-to-consumer in a
    standardized way.
    As web service technologies are deployed within and across
    organizations over the next several years, it will be
    critical that web services undergo performance testing.
    As with any enterprise software project, the adoption of
    proper test methodologies and use of testing tools will
    play a key part in the overall success or failure of
    projects utilizing web services. In a compressed
    software project schedule, an organization must
    quickly determine if its web services will operate
    successfully under a variety of load conditions. Like other
    web-based technologies, successful web services will need
    to respond quickly and correctly when implemented.
    During our presentation, we will discuss the testing
    challenges created by this emerging technology, along with
    the variety of testing solutions available. Automated
    web service testing will be discussed and demonstrated
    using FirstACT, the first web services performance testing solution available
    on the market. Using a sample web
    service, automatic test case creation, scalability testing,
    and results analysis will be explored.
    If you wish to download FirstACT prior to the web event, you can do so at:
    http://www.empirix.com/downloads/FirstACT

    As BEA and other vendors continue to add Web Services support
    to their enterprise software, what is your plan for
    quantifying the performance impact and the functional
    correctness of using web services before going live with the
    final application?
    Empirix is hosting a free one hour web event discussion on
    web services testing and automated web services testing
    solutions on Thursday, January 17, 2-3pm Eastern time.
    To sign-up for this web event or learn about other web
    events being offering by Empirix this month, go to:
    http://webevents.empirix.com
    For your convenience, here is the complete abstract:
    The advent of web services has brought the promises of
    integrating multiple software applications from
    heterogeneous networks and for exchanging information
    from vendor-to-vendor or vendor-to-consumer in a
    standardized way.
    As web service technologies are deployed within and across
    organizations over the next several years, it will be
    critical that web services undergo performance testing.
    As with any enterprise software project, the adoption of
    proper test methodologies and use of testing tools will
    play a key part in the overall success or failure of
    projects utilizing web services. In a compressed
    software project schedule, an organization must
    quickly determine if its web services will operate
    successfully under a variety of load conditions. Like other
    web-based technologies, successful web services will need
    to respond quickly and correctly when implemented.
    During our presentation, we will discuss the testing
    challenges created by this emerging technology, along with
    the variety of testing solutions available. Automated
    web service testing will be discussed and demonstrated
    using FirstACT, the first web services performance testing solution available
    on the market. Using a sample web
    service, automatic test case creation, scalability testing,
    and results analysis will be explored.
    If you wish to download FirstACT prior to the web event, you can do so at:
    http://www.empirix.com/downloads/FirstACT

  • Performance impact of Web Services

    As WebLogic adds support for Web Services to its platform, what is
    your plan for quantifying the performance impact and the functional
    correctness of using web services before going live with the final
    application.
    Empirix is hosting a free one hour web event discussion on web
    services testing and automated web services testing solutions on
    Thursday, January 17, 2-3pm Eastern time.
    To register for this web event or learn about other web events being
    offering by Empirix this month, go to:
    http://webevents.empirix.com
    The complete abstract is below:
    The advent of web services has brought the promises of integrating
    multiple software applications from heterogeneous networks and for
    exchanging information from vendor-to-vendor or vendor-to-consumer in
    a standardized way.
    As web service technologies are deployed within and across
    organizations over the next several years, it will be critical that
    web services undergo performance testing. As with any enterprise
    software project, the adoption of proper test methodologies and use of
    testing tools will play a key part in the overall success or failure
    of projects utilizing web services. In a compressed software project
    schedule, an organization must quickly determine if its web services
    will operate successfully under a variety of load conditions. Like
    other web-based technologies, successful web services will need to
    respond quickly and correctly when implemented.
    During our presentation, we will discuss the testing challenges
    created by this emerging technology, along with the variety of testing
    solutions available. Automated web service testing will be discussed
    and demonstrated using FirstACT, the first web services performance
    testing solution available on the market. Using a sample web service,
    automatic test case creation, scalability testing, and results
    analysis will be explored.

    Hi,
    We test several frameworks and find out that usually JAXB 2.0 performs better than XMLBeans, but that is not a strict rule.
    Regards,
    LG

  • Index creation online - performance impact on database

    hi,
    I have oracle 11.1.0.7 database running on Linux as 3 node RAC.
    I have a huge table which has more than 255 columns and is about 400GB in size which is also highly fragmented because of constant DML activities.
    Questions:
    1. For now i am trying to create an index Online while the business applications are running.
    Will there be any performance impact on the database to create index Online on a single column of a table 'TBL' while applications are active against the same table? So basically my question will index creation on a object during DML operations on the same object have performance impact on the database? is there a major performance impact difference in the database in creating index online and not online?
    2. I tried to build an index on a column which has NULL value on this same table 'TBL' which has more than 255 columns and is about 400GB in size highly fragmented and has about 140 million rows.
    I requested the applications to be shutdown, but the index creation with parallel of 4 a least took more than 6 hours to complete.
    We have a Pre-Prod database which has the exported and imported copy of the Prod data. So the pre-Prod is a highly de-fragmented copy of the Prod.
    When i created the same index on the same column with NULL, it only took 15 minutes to complete.
    Not sure why on a highly fragmented copy of Prod it took more than 6 hours compared to highly defragmented copy of Pre-Prod where the index creation took only 15 minutes.
    Any thoughts would be helpful.
    Thanks.
    Phil.

    How are you measuring the "fragmentation" of the table ?
    Is the pre-prod database running single instance or RAC ?
    Did you collect any workload stats (AWR / Statspack) on the pre-prod and production systems while creating (or failing to create) the index ?
    Did you check whether the index creation ended up in-memory, single pass or multi pass in in the two environments ?
    The commonest explanation for this type of difference is two-fold:
    a) the older data needs a lot of delayed block cleanout, which results in a lot of random I/O to the undo tablespace - slowing down I/O generally
    b) the newer end of the table is subject to lots of change, so needs a lot of work relating to read-consistency - which also means I/O on the undo system
      --  UPDATED:  but you did say that you had stopped the application so this bit wouldn't have been relevant.
    On top of this, an online (re)build has to lock the table briefly at the start and end of the build, and in a busy system you can wait a long time for the locks to be acquired - and if the system has been busy while the build has been going on it can take quite a long time to apply the journal file to finish the index build.
    Regards
    Jonathan Lewis

  • Performance impact related to workbook

    Hi All,
    Please let me know the performance impact on workbooks if i insert 13 work sheets,
    How much performance impact will be there?
    If i insert all reports reports in 2 worksheets by splitting them,
    How much performance impact ll be there?
    Please let me know which one is better way
    Urgent!...
    Thanks,
    Sarau

    Hello Sarau,
    The performance is purely based on the Query selection, since that decide the number of records to be fetched.
    The more you filter the query the performance will be better.
    Since you are having many queries in a workbook, make sure that all have common variable screen.
    Thanks
    Chandran

  • Performance impact on the size of the CHM file

    Is there any impact on performance depending on the size of a
    CHM file?

    The main issues people have with help file performance
    (regardless of whether it is a CHM file) are related to the number
    of images, DHTML hotspots, bookmarks and links they have in a
    topic. The number of topics in a CHM should not be an issue. What
    exactly are you trying to access the performance impact of?

  • EBS performance impact using it as a Data Source

    I have a quick question on EBS performance. If I set up the EBS Database as a data source for SSRS (SQL Server Reporting Services), would there be a performance impact on EBS, due to SSRS accessing EBS Data for reports generation? Now, I know there'll always be a hit depending on the volume of data being accessed. But, my question is, will it be significantly higher using an external reporting tool using an ODBC connection rather than native XML Publisher.

    I have a quick question on EBS performance. If I set up the EBS Database as a data source for SSRS (SQL Server Reporting Services), would there be a performance impact on EBS, due to SSRS accessing EBS Data for reports generation? Now, I know there'll always be a hit depending on the volume of data being accessed. But, my question is, will it be significantly higher using an external reporting tool using an ODBC connection rather than native XML Publisher.Hi,
    Tough to answer without looking at data; my suggestion would be to have a test EBS environment setup, get permission from the vendors to run performance test without buying license - compare AWRs from both scenarios and then decide.
    Generally speaking, native XML publisher (BI Publisher) has less of database performance hit than external reporting tools using ODBC.
    Hope this helps.
    Regards,

  • Performance Impact with OR concatenation / Inlist Iterator

    Hello guys,
    is there any performance impact with using OR concatenations or some IN-Lists?
    The function of both is the "same":
    1) Concatenation (OR-processing)
    SELECT * FROM emp WHERE mgr# = 1 OR job = ‘YOURS’;- Similar to query rewrite into 2 seperate queries
    - Which are then ‘concatenated’
    2) Inlist Iterator
    SELECT * FROM dept WHERE d# in (10,20,30);- Iteration over enumerated value-list
    - Every value executed seperately
    - Same as concatenation of 3 “OR-red” values
    So i want to know if there is any performance impact if using IN-Lists instead of OR concatenations.
    Thanks and Regards
    Stefan

    The note is very misleading and far from complete; but there is one critical point of difference that you need to observe. It's talking about using a tablescan to deal with an IN-list (and that's NOT "in-list iteration"), my comments start by saying "if there is a suitable indexed access path."
    The note, by the way, describes a transformation to a UNION ALL - clearly that would be inefficient if there were no indexed access path. (Given the choice between one tablescan and several consecutive tablescans, which option would you choose ?).
    The note, in effect, is just about a slightly more subtle version of "why isn't oracle using my index". For "shorter" lists you might get an indexed iteration, for "longer" lists you might get a tablescan.
    Remember, Metalink is not perfect; most of it is just written by ordinary people who learned about Oracle in the normal fashion.
    Quick example to demonstrate the difference between concatenation and iteration:
    drop table t1;
    create table t1 as
    select
         rownum     id,
         rownum     n1,
         rpad('x',100)     padding
    from
         all_objects
    where
         rownum <= 10000
    create index t1_i1 on t1(id);
    execute dbms_stats.gather_table_stats(user,'t1')
    set autotrace traceonly explain
    select
         /*+ use_concat(t1) */
         n1
    from
         t1
    where
         id in (10,20,30,40,50,60,70,80,90,100)
    set autotrace offThe execution plan I got from 8.1.7.4 was as follows - showing the transformation to a UNION ALL - this is concatenation and required 10 query block optimisations (which were all done three times):
    Execution Plan
       0      SELECT STATEMENT Optimizer=ALL_ROWS (Cost=20 Card=10 Bytes=80)
       1    0   CONCATENATION
       2    1     TABLE ACCESS (BY INDEX ROWID) OF 'T1' (Cost=2 Card=1 Bytes=8)
       3    2       INDEX (RANGE SCAN) OF 'T1_I1' (NON-UNIQUE) (Cost=1 Card=1)
       4    1     TABLE ACCESS (BY INDEX ROWID) OF 'T1' (Cost=2 Card=1 Bytes=8)
       5    4       INDEX (RANGE SCAN) OF 'T1_I1' (NON-UNIQUE) (Cost=1 Card=1)
       6    1     TABLE ACCESS (BY INDEX ROWID) OF 'T1' (Cost=2 Card=1 Bytes=8)
       7    6       INDEX (RANGE SCAN) OF 'T1_I1' (NON-UNIQUE) (Cost=1 Card=1)
       8    1     TABLE ACCESS (BY INDEX ROWID) OF 'T1' (Cost=2 Card=1 Bytes=8)
       9    8       INDEX (RANGE SCAN) OF 'T1_I1' (NON-UNIQUE) (Cost=1 Card=1)
      10    1     TABLE ACCESS (BY INDEX ROWID) OF 'T1' (Cost=2 Card=1 Bytes=8)
      11   10       INDEX (RANGE SCAN) OF 'T1_I1' (NON-UNIQUE) (Cost=1 Card=1)
      12    1     TABLE ACCESS (BY INDEX ROWID) OF 'T1' (Cost=2 Card=1 Bytes=8)
      13   12       INDEX (RANGE SCAN) OF 'T1_I1' (NON-UNIQUE) (Cost=1 Card=1)
      14    1     TABLE ACCESS (BY INDEX ROWID) OF 'T1' (Cost=2 Card=1 Bytes=8)
      15   14       INDEX (RANGE SCAN) OF 'T1_I1' (NON-UNIQUE) (Cost=1 Card=1)
      16    1     TABLE ACCESS (BY INDEX ROWID) OF 'T1' (Cost=2 Card=1 Bytes=8)
      17   16       INDEX (RANGE SCAN) OF 'T1_I1' (NON-UNIQUE) (Cost=1 Card=1)
      18    1     TABLE ACCESS (BY INDEX ROWID) OF 'T1' (Cost=2 Card=1 Bytes=8)
      19   18       INDEX (RANGE SCAN) OF 'T1_I1' (NON-UNIQUE) (Cost=1 Card=1)
      20    1     TABLE ACCESS (BY INDEX ROWID) OF 'T1' (Cost=2 Card=1 Bytes=8)
      21   20       INDEX (RANGE SCAN) OF 'T1_I1' (NON-UNIQUE) (Cost=1 Card=1)This is the execution plan I got from 9.2.0.8, which doesn't transform to the UNION ALL, and only needs to optimise one query block.
    Execution Plan
       0      SELECT STATEMENT Optimizer=ALL_ROWS (Cost=3 Card=10 Bytes=80)
       1    0   INLIST ITERATOR
       2    1     TABLE ACCESS (BY INDEX ROWID) OF 'T1' (Cost=3 Card=10 Bytes=80)
       3    2       INDEX (RANGE SCAN) OF 'T1_I1' (NON-UNIQUE) (Cost=2 Card=10)Regards
    Jonathan Lewis
    http://jonathanlewis.wordpress.com
    http://www.jlcomp.demon.co.uk

Maybe you are looking for

  • Why would I get an important message that says, "I can see you" It was a loud alarm with a message box.

    My Iphone 4s had a loud alarm with a text box that said I can see you!  I was trying to copy and paste in my contacts at the time it occured.  It kinda freaked me out.

  • How to change fonts in Gvim's menus

    How can I change the fonts in Gvim's menues? I tried to add this in the .Xdefaults file Vim*useSchemes: all Vim*sgiMode: true Vim*useEnhancedFSB: true Vim.foreground: Black Vim.background: Wheat Vim*fontList: 7x13 Vim.menuFontSet: -*-bitstream vera s

  • IPhoto 6 Hangs When "Recovered Photos" Are Opened

    iPhoto 6 "recovered" 300+ photos when I updated from iPhoto 5. I can still see the recovered photos in both my library and albums, but if I double-click on the photos, iPhoto 6 hangs with a spinning ball icon requiring a Force Quit to recover. I can

  • Firefox crash when open an oracle e-business suite form

    after install j initiator run oracle application add user name & password selcect any form , then the firefox crash and quit == User Agent == Mozilla/4.0 (compatible; MSIE 8.0; Windows NT 5.1; Trident/4.0; .NET CLR 1.1.4322; .NET CLR 2.0.50727; .NET

  • Blowout program !!

    In this simple program, i used the logical operator OR ||, but i can't get it to run right!! is there is something i'm doing wrong here in this source code?