Upgrade 9i to 10g Performance Issue

Hi All,
DBA team recently upgraded the database from 9i to 10g (Oracle Database 10g Enterprise Edition Release 10.2.0.2.0 - 64bi).
There is a process as follows
Mainframe --> Flat file --> (Informatica) --> Oracle DB --> (processing through SPs) --> Report
The whole process normally used to take 1.5 - 2 hrs which is now taking 4-5 hrs after 9i to 10g upgrade.
I tried searching on Google but could not find any detailed procedure to root out the cause. Below is the link of such an instance-
http://jonathanlewis.wordpress.com/2007/02/02/10g-upgrade/
Can someone suggest where to start digging to find out the reason behind the slowdown (if possible, please tell the detailed procedure or reference)?
It is to be noted that there was no other change besides Database upgrade. Also, for about 20 days after the upgrade the whole process took near 1 hr (which was a pleasnt gain) but afterward it shoot up to 4-5 hrs suddenly and since then its taking same time.
Thanks.

Without more information (Statspack reports, AWR reports, or at least query plans), it's going to be exceptionally hard to help you.
There are dozens, more likely hundreds, of parameters which change between 9i and 10g. And one of which could cause an individual query to perform worse. Plus, obviously something changed after the process was running quickly for 20 days, so there is some change on your side that would need to be identified and accounted for. There is no way for anyone to guess which of the hundreds of parameter and behavioral changes might have contributed to the problem, nor for us to guess what might have changed after 20 days in your database.
If you can at least provide Statspace/ AWR reports, we may be able to help you out with a general tuning exercise. If you can identify what changed between the time the process was running well to the time it stopped running well, we may be able to help there as well.
Justin

Similar Messages

  • Oracle 10g performance issues

    Hi,
    We were using Oracle 9i in Solaris 5.8 and it was working fine with some minor performance issues. We formatted the Solaris server with new Solaris 5.10 and installed Oracle 10g.
    Now we are experiencing some performance issues in Oracle 10g. This issue is arising when using through Websphere 5.1.
    We have analyzed the schema, index is rebuild, SGA is 4.5 GB, PGA is 2.0 GB, Solaris RAM is 16 GB. Also we are having some Mat Views (possibly this may cause performance issues - not sure) due to refresh.
    Also I have changed some parameters in init.ora file like query_rewrite = STALE_TOLERATED, open_cursors = 1500 etc.
    Is is something due to driver from which the data is accessed. I guess it is not utilizing the indexes on the table.
    Can anyone please suggest, what could be the issue ?

    <p>There are a lot of changes to the optimizer in the upgrade from 9i to 10g, and you need to be aware of them. There are also a number of changes to the default stats collection mechanism, so after your upgrade your statistics (hence execution paths) could change dramatically.
    </p>
    <p>
    Greg Rahn has a useful entry on his blog about stats collection, and the blog al,so points to an Oracle white paper which will give you a lot of ideas about where the optimizer changes - which may help you spot your critical issues.
    </p>
    <p>Otherwise, follow triggb's advice about using Statspack to find the SQL that is the most expensive - it's reasonably likely to be this SQL that has changed execution plans in the upgrade.
    </p>
    Regards
    Jonathan Lewis
    http://jonathanlewis.wordpress.com
    http://www.jlcomp.demon.co.uk

  • Disco Plus 10g Performance Issue When Moving Table Headings to Page Items

    Hello,
    We are experiencing an performance anomaly in Discoverer Plus (the latest release: 10.1.2.48.18), and are wondering if anyone else out there has noticed similar behavior. We're having trouble identifying the cause, and have a TAR openned with Oracle support but they are not able to reproduce our issue, and have been slow to offer suggestions or help.
    The issue happens when a user drags a table heading up into the page items area in a worksheet. When we move a heading up to the page items it works quickly one time, but then, every time after that, it takes anywhere from a few minutes up to HOURS to move any additional headings up to the page items area.
    The problem occurrs on approximately 3 out of every 5 of our workbooks. We've tried different sizes of workbooks, with large (several million records) and small (a few thousand records) datasets, and none of this seems to affect the issue at hand. The problem only occurs in Discoverer Plus, not Desktop. We've also spent some time researching caching and memory configuration, and believe that we have setup all of the recommended options for maximum performance on our systems.
    I would just like to know if anyone else out there in the community has experienced the same issue, and if anyone has any advice for us.
    Thank you,
    -Scott

    Hi,
    I found the following on Best Practices of Orcle Discoverer 10g by
    Mike Donohue - Product Management - Oracle Business Intelligence
    Performance – Parameters and Page Items
    Page Items provide very responsive, interactive manipulation of data
    At a cost:
    Forces retrieval of all Detail values
    Incremental increases of memory as indices are built
    Parameters reduce result set, improve performance
    Use Page Items only when needed – 2 to 3 with < 12 values each
    Performance – Parameters and Page Items - example
    Loc(3), Dept(10), ProdType(50), Prod(1,000), Date(365)
    547,500,000 potential rows/combinations (3*10*50*1000*365)
    Use parameters for Loc, Dept, ProdType, and 90 day date range
    90,000 potential rows/combinations (1000*90)
    Reduce data retrieved by ~ 6000 X
    Improve performance by several orders of magnitude
    I hope this will help you!

  • 10g performance issue + RAC

    Hi,
    we are running oracle e-business suite 111 with 10g database and RAC.
    Please let us know what can be tuned, following is a AWR report statistics:
    Snap Id Snap Time Sessions Cursors/Session
    Begin Snap: 2904 20-Jan-09 08:30:17 362 55.3
    End Snap: 2905 20-Jan-09 09:30:43 525 79.0
    Elapsed: 60.42 (mins)
    DB Time: 214.04 (mins)
    Report Summary
    Cache Sizes
    Begin End
    Buffer Cache: 2,992M 2,896M Std Block Size: 8K
    Shared Pool Size: 1,008M 1,104M Log Buffer: 14,352K
    Load Profile
    Per Second Per Transaction
    Redo size: 95,171.90 7,079.15
    Logical reads: 83,587.58 6,217.48
    Block changes: 696.50 51.81
    Physical reads: 163.22 12.14
    Physical writes: 18.76 1.40
    User calls: 429.26 31.93
    Parses: 156.87 11.67
    Hard parses: 5.60 0.42
    Sorts: 146.05 10.86
    Logons: 0.88 0.07
    Executes: 991.77 73.77
    Transactions: 13.44
    % Blocks changed per Read: 0.83 Recursive Call %: 83.42
    Rollback per transaction %: 39.30 Rows per Sort: 10.42
    Instance Efficiency Percentages (Target 100%)
    Buffer Nowait %: 99.99 Redo NoWait %: 100.00
    Buffer Hit %: 99.81 In-memory Sort %: 100.00
    Library Hit %: 98.77 Soft Parse %: 96.43
    Execute to Parse %: 84.18 Latch Hit %: 99.90
    Parse CPU to Parse Elapsd %: 17.28 % Non-Parse CPU: 99.64
    Shared Pool Statistics
    Begin End
    Memory Usage %: 75.51 84.38
    % SQL with executions>1: 92.27 86.87
    % Memory for SQL w/exec>1: 90.30 88.03
    Top 5 Timed Events
    Event Waits Time(s) Avg Wait(ms) % Total Call Time Wait Class
    CPU time 5,703 44.4
    db file sequential read 399,573 1,732 4 13.5 User I/O
    gc current block 2-way 380,370 846 2 6.6 Cluster
    gc cr grant 2-way 202,084 325 2 2.5 Cluster
    gc cr multi block request 273,702 315 1 2.5 Cluster
    RAC Statistics
    Begin End
    Number of Instances: 2 2
    Global Cache Load Profile
    Per Second Per Transaction
    Global Cache blocks received: 207.45 15.43
    Global Cache blocks served: 201.99 15.02
    GCS/GES messages received: 491.52 36.56
    GCS/GES messages sent: 592.97 44.11
    DBWR Fusion writes: 3.18 0.24
    Estd Interconnect traffic (KB) 3,487.35
    Global Cache Efficiency Percentages (Target local+remote 100%)
    Buffer access - local cache %: 99.56
    Buffer access - remote cache %: 0.25
    Buffer access - disk %: 0.19
    Global Cache and Enqueue Services - Workload Characteristics
    Avg global enqueue get time (ms): 0.6
    Avg global cache cr block receive time (ms): 3.1
    Avg global cache current block receive time (ms): 3.4
    Avg global cache cr block build time (ms): 0.0
    Avg global cache cr block send time (ms): 0.0
    Global cache log flushes for cr blocks served %: 12.5
    Avg global cache cr block flush time (ms): 3.4
    Avg global cache current block pin time (ms): 0.0
    Avg global cache current block send time (ms): 0.0
    Global cache log flushes for current blocks served %: 0.0
    Avg global cache current block flush time (ms): 1.0
    Global Cache and Enqueue Services - Messaging Statistics
    Avg message sent queue time (ms): 0.1
    Avg message sent queue time on ksxp (ms): 1.5
    Avg message received queue time (ms): 0.0
    Avg GCS message process time (ms): 0.0
    Avg GES message process time (ms): 0.0
    % of direct sent messages: 46.02
    % of indirect sent messages: 49.25
    % of flow controlled messages: 4.73
    Regards
    Arora

    Achyot,
    Apart from the comments by Justin a few other general comments
    - RAC is mainly suited for OLTP
    - if your OLTP application doesn't scale in a single instance configuration, RAC will make this worse. You may want to tune your application without RAC first (set CLUSTER_DATABASE to FALSE)
    - You need to find out whether the delays are resulting from a badly configured Interconnect. This will show up as events starting with 'gc', above all 'gc cache cr request'
    Hth
    Sybrand Bakker
    Senior Oracle DBA

  • Oracle 10g performance issue

    Hi all,
    In sunfire v890 we have installed oracle 10g release 2 on solaris 10.
    prstat -a command shows :
    NPROC USERNAME SIZE RSS MEMORY TIME CPU
    105 root 9268M 6324M 20% 1:21:57 0.4%
    59 oracle 24G 22G 71% 0:04:33 0.1%
    2 nobody4 84M 69M 0.2% 0:11:32 0.0%
    2 esbuser 13M 9000K 0.0% 0:00:46 0.0%
    1 smmsp 7560K 1944K 0.0% 0:00:00 0.0%
    4 daemon 12M 7976K 0.0% 0:00:00 0.0%
    and top utility shows :
    last pid: 8639; load avg: 0.09, 0.09, 0.09; up 2+06:05:29 17:07:50
    171 processes: 170 sleeping, 1 on cpu
    CPU states: 98.7% idle, 0.7% user, 0.7% kernel, 0.0% iowait, 0.0% swap
    Memory: 32G phys mem, 22G free mem, 31G swap, 31G free swap
    therefore from prstat we come to know that memory used by oracle is 71%
    where as top says 31.25% used.....
    which one is true in this scenario.....
    shall we go ahead in trusting top utility????
    Advance thanks to you.

    Hi darren
    The main thing is,, prstat -a command showing oracle
    user occupied 70%.In top utility showing 22gb memory
    free out of 32 gb.That means 10gb was occupied by all
    users.In percentage calculation its comes
    31.25%...i.e top shows all users occupied only
    31.25%Right. That's all memory in use, correct? From your first message I thought you meant it said that was the amount used by Oracle.
    It's easy to calculate total memory in use.
    It's hard to calculate memory in use by a subset of processes (perhaps those owned by a particular user).
    but prstat -t command shows 70% occupied by oracle.
    which one i want to believe??????The prstat command showing memory in use by a user will be incorrect because it does not calculated shared pages properly.
    As far as I am aware, 'top' has no similar display.
    Darren

  • AS + Oracle DB 10g performance issue

    Hi,
    My company is currently exploring Oracle Portal and Discoverer. I have a question and would really appreciate if anyone can help me.
    I install the AS infrastructure and the database in the same machine. Currently when I log in to the portal for the first time of the day, it is always very slow. If I install the IAS and database in separate machines, will this give a significant improvement in the response time for my portal? What are the implications of installing this in 2 machines instead of 1?
    Thanks :)

    Two things -
    With Portal and Discoverer, your AS setup probably includes both mid-tier and Infrastructure installations. (is this a BI & F type mid-tier?) The Infrastructure commonly installs its own database for Metadata repositories etc. So this machine has 2 x AS instances and 2 x Database server, correct? An awful lot for a single machine, I'd guess.
    Do you restart the machine first thing each day? The first time a web app loads it has a startup latency. Monday morning, sort of. There is a lot of components involved and they all need to load up some stuff in memory to be able to do some work. This takes time (disk-> mem is very very slow, for one thing).
    2 machines means more complex setup and hopefully a 'quicker' solution.
    Depending on requirements, you may want one machine for customer/your own database, one machine for AS database and one for AS mid-tier.... What kind of platform are you planning on (hardware + os)?

  • 11.2.0.0 upgrade- performance issue when commits

    We have upgraded our data warehouse system from 10.2.0.4 for 11.2.0.2. After upgrade ETL system has performance issue. Extraction is fine because we compared the explain plans before and after upgrade. However, loading is taking very long. In loading there is a commit after every 1000 records, that why its slow. Even when we import data with commit=y option its very slow. Is there any know issue on this?
    Any help pl?
    Thanks

    We are solaris 10. ETL job on 10.2.0.3 was taking 2 hours and now its taking 11 hours.
    I beleive commit is a issue because
    1. Source query (Extraction) has same explain plan.
    2. Source database is still on 10.2.0.3
    3. Servers and network is same for 10g and 11g.
    4. I already experienced COMMIT=y performance issue in import.
    Thanks

  • Performance issue in oracle 11.1.0.7 version

    Hi ,
    In production environment we have some cronjobs are scheduled, they will run every Saturday. One of the cronjob is taking more time to finish the job.
    Previous oracle version is 10.2.0.4, that time it was taking 36hrs to complete it. After upgrading to 11gr1, now it's taking 47hrs some time 50hrs to finish.
    I have asked my production DBA take AWR report after finish the cronjob.
    Now he sent the AWR report, but i don't know how to read it. Can you please help me to read the AWR reports, and i need to give some recommendations to reduce the overall running time.
    I don't know how to attach the AWR report here.
    Please help me on this.
    Thanks
    Shravan Kumar

    Hi,
    Now he sent the AWR report, but i don't know how to read it. Can you please help me to read the AWR reports, and i need to give some recommendations to reduce the overall running time.An't you a DBA? Probably you should seek help of you DBA to read the AWR and mean while you should also have AWR of 10g where this job was running previously so that you can compare the things.
    Did you do a testing before upgrade? you SHOULD have done a thorough testing of your applications/reports before the upgrade and resolve the performance issues before the production upgrade.
    Mean while you do investigation, you can set optimizer_features_enable=10.2.0.4 for only cron job session to check whether you job returns to same 26 hours time
    alter session set optimizer_features_enable='10.2.0.4';Salman

  • RFBIBL00 Performance issues Post ECC6

    Hi Folks,
    Have just completed ECC6 upgrade am seeing terrible performance issues when using RFBIBL00 for loading. All loads seem to stick reading table TBIZU. DB Is oracle and table is quite small. Stats are valid and when checking explain plan in oracle the query seems to be fine.
    Anybody seen anything similar.
    ta

    hi,
    Please check this link
    [Re: rfbibl00;
    regards,
    suhas

  • DB Performance Issues after 10g Upgrade in EBS Instance

    We have upgraded our Database from 9i to 10g as first part of EBS 11.5.9 to 11.5.10.2 upgrade. Currently our production is running on 11.5.9 apps with 10g DB.
    Facing performance problems now. one of the them is, one Valueset query not using funcion based index while fired from the front end. but the same query when collected from SQL trace tkprofed file and executed from SQL Plus, it uses all proper indexes. We are not getting the cause of this.
    Had anyone faced same kind of issues before. please suggest.
    thanks,
    Raj.

    Make sure you have all of the recommended performance patches for 11.5.9, and gather stats for SYS and SYSTEM in the following manner:
    Oracle E-Business Suite Recommended Performance Patches
    http://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=244040.1
    Collecting Statistics with Oracle Apps 11i
    http://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=368252.1
    execute dbms_stats.unlock_schema_stats('SYS');
    execute dbms_stats.unlock_schema_stats('SYSTEM');
    exec dbms_stats.gather_schema_stats('SYSTEM',options=>'GATHER', estimate_percent => 100, method_opt => 'FOR ALL COLUMNS SIZE AUTO', cascade => TRUE);
    exec dbms_stats.gather_schema_stats('SYS',options=>'GATHER', estimate_percent => 100, method_opt => 'FOR ALL COLUMNS SIZE AUTO', cascade => TRUE);
    exec dbms_stats.gather_fixed_objects_stats();
    commit;
    exec dbms_stats.DELETE_TABLE_STATS('SYS','X$KCCRSR');
    exec dbms_stats.LOCK_TABLE_STATS('SYS','X$KCCRSR');
    commit;
    The last 3 commands resolve problems with RMAN, in case you are using it.
    Rman Backup is Very Slow selecting from V$RMAN_STATUS
    http://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=375386.1
    Poor performance when accessing V$RMAN_BACKUP_JOB_DETAILS
    http://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=420200.1
    Troubleshooting Oracle Applications Performance Issues
    http://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=169935.1
    Debugging General Performance Issues with Oracle Apps
    http://blogs.oracle.com/schan/newsItems/departments/optimizingPerformance/2007/05/18#a1548
    Performance Tuning the Apps Database Layer
    http://blogs.oracle.com/schan/newsItems/departments/optimizingPerformance/2007/05/17#a1562
    Preventing Apps 11i Performance Issues in Four Steps
    http://blogs.oracle.com/schan/newsItems/departments/optimizingPerformance/2007/05/21#a1566

  • Performance issue with form after 10g upgrade

    Hi Team,
    Last week we have upgraded our systems to 10g database.
    Ever since we did an upgrade there is a huge performance issues with the custom forms and this is causing a major setback to our business. Before the upgrade the forms were running with our any performance issues.
    Can anyone please help me in finding out the reason behind the performance issue(May be a tar or performance tuning).
    Many Thanks in Advance.
    Regards
    Kumar

    Like Jan said,
    You must supply more information so we can help you, like where the degradation happens?, in processing? in navigation? in forms loading?? where?
    You may also do a little test. Create a one button form, just a canvas and a button, you can include a message in when-button-pressed trigger.
    run it and see what happens.
    Tony

  • Performance issues executing process flows after upgrading db to 10G

    We have installed OWF 2.6.2, and initially our database was at 9.2. Last week we updated our database to 10g, and process flow executions are taking a lot longer, from 1 minute to 15 minutes.
    Any ideas anyone what could be the cause of this performance issue?
    Thanks,
    Yanet

    Hi,
    Oracle10g database behaves differently on the statistics of tables and indexes. So check these and check wether the mappings are updating these statistics at the right moments with respect to the ETL-proces and with the right interval.
    Also, check your generated sources on how statistics are gathered (dmbs_stats.gather....). Does the index that might play a vital role in Oracle9i get new statistics, or only the table? Or only the table where doubled in amount of rows by this mapping?
    You can always take matter into your own hands, by letting OWB NOT generate the source for gathering statistics, and call your own procedure in a post-mapping.
    Regards,
    André

  • Database migrated from Oracle 10g to 11g Discoverer report performance issu

    Hi All,
    We are now getting issue in Discoverer Report performance as the report is keep on running when database got upgrade from 10g to 11g.
    In database 10g the report is working fine but the same report is not working fine in 11g.
    The query i have changed as I have passed the date format TO_CHAR("DD-MON-YYYY" and removed the NVL & TRUNC function from the existing query.
    The report is now working fine in Database 11g backhand but when I am using the same query in Discoverer it is not working and report is keep on running.
    Please advise.
    Regards,

    Pl post exact OS, database and Discoverer versions. After the upgrade, have statistics been updated ? Have you traced the Discoverer query to determine where the performance issue is ?
    How To Find Oracle Discoverer Diagnostic and Tracing Guides [ID 290658.1]
    How To Enable SQL Tracing For Discoverer Sessions [ID 133055.1]
    Discoverer 11g: Performance degradation after Upgrade to Database 11g [ID 1514929.1]
    HTH
    Srini

  • Performance issue with Crystal when upgrading Oracle to 11g

    Dear,
    I am facing performance issue in crystal report and oracle 11g as below:
    In ther report server, I have created a ODBC for connect to another Oracle 11g server. also in report server I have created and published a folder to content all of my crystal report. These report can connect to oracle 11g server via ODBC.
    and I have a tomcat server to run my application in my application I refer to report folder in report server.
    This way can work with SQL server and oracle 9 or 10g but it facing performance issue in oracle 11g.
    please let me know the root cause.
    Notes: report server, tomcate server are win 32bit, but oracle is in win 64bit, and i have upgraded DataDirect connect ODBC version 6.1 but the issue can not resolve.
    Please help me to solve it.
    Thanks so much,
    Anh

    Hi Anh,
    Use a third party ODBC test tool now. SQL Plus will be using the Native Oracle client so you can't compare performance.
    Download our old tool called SQLCON: https://smpdl.sap-ag.de/~sapidp/012002523100006252882008E/sqlcon32.zip
    Connect and then click on the SQL tab and paste in the SQL from the report and time that test.
    I believe the issue is because the Oracle client is 64 bit, you should install the 32 bit Oracle Client. If using the 64 bit client then the client must thunk ( convert 64 bit data to 32 bit data format ) which is going to take more time.
    If you can use OLE DB or using the Oracle Server driver ( native driver ) should be faster. ODBC puts another layer on top of the Oracle client so it too takes time to communicate between the layers.
    Thank you
    Don

  • Performance Problem After upgrade to oracle 10g

    Hi
    I have upgrade one of my datawarehouse database from oracle 9.2.0.8 to oracle 10.2.0.4 running on solaris 9
    After the upgrade jobs which were running in the database is taking hell lot of time.
    The jobs are accessing the views which is used to get the monthly report data from the database.
    what could be the solution and where to start from to get the RCA to resolve this performance issue
    Please let me know if you require any other information
    database is currently running in the automatic shared memory management mode ie SGA_MAX and SGA_TARGET parameters are defined for that

    There are a lot of differences between 10g and 9i in this regard, among these are:
    - There is a default job that gathers statistics every night which is not there in 9i. You might have totally different statistics as in 9i due to that job, depending on how and if at all you used to collect statistics in 9i
    - The 10g DBMS_STATS package collects histograms on some columns by default (parameter METHOD_OPT=>'FOR ALL COLUMNS SIZE AUTO' default in 10g whereas 'FOR ALL COLUMNS SIZE 1' in 9i) which can have a significant effect on the execution plans
    - The 10g optimizer has CPU costing enabled by default which can make significant changes to your execution plans due to different costing of table scans and order of predicate evaluation. In addition it uses NOWORKLOAD system statistics if system statistics have not been gathered explicitly
    - 10g checks the min and max values stored for columns in the data dictionary. If your predicates are way off compared to these values then 10g begins to adjust the calculated selectivity of the predicate which can again significantly affect your execution plans
    - 10g introduces the "Cost Based Query Transformation (CBQT)" feature which means that rather than applying heuristic transformation rules transformations are costed and potentially discarded whereas 9i applies transformations unconditionally whenever possible
    Check also the following note resp. white paper:
    http://optimizermagic.blogspot.com/2008/02/upgrading-from-oracle-database-9i-to.html
    Regards,
    Randolf
    Oracle related stuff blog:
    http://oracle-randolf.blogspot.com/
    SQLTools++ for Oracle (Open source Oracle GUI for Windows):
    http://www.sqltools-plusplus.org:7676/
    http://sourceforge.net/projects/sqlt-pp/

Maybe you are looking for

  • Hwo to get the spool number from report output

    Hi, I am displaying some output in the report using write statements and within my program I need to collect the output written by write statements and send it as an email.So for that I need to generate the spool number and I am using the below code

  • HT1577 error message while downloading

    I am trying to download an update to my operating system from the app store.  Although my machine says I am connected to the internet, about 3 GB into the download, I get an error message that I am not connected to the internet.  I have no clue what'

  • Physcial Inventory by mass update

    Hi We have created many Physical Inventory Document in MI31 Exported to Excel To enter the count details we want to use MI34, but system is asking for logical file. Do we need to create any logical files for these? How to update by mass the count det

  • Returning the number of records retrieved by a query

    Hi everyone, I did a search on this question, but I'm still unclear as to why it's so difficult to retrieve the number of records returned by a query. What is the easiest and most straight-foward way ? Thanks, Bob

  • Facing problem in saving data without overlapping in for loop

    Hi, I am facing problem in writting data withou overlapping if i run outer loop for 2 or more times and in inner for loop i am getting array in a way I want but when i try to build that array with logging temperature i am not able to do it. Please gu