Database open (recovery) taking too long

Hi,
Ive been using your awesome BerkeleyDB Java Edition for a couple of years, and have been very happy with it.
I am currently facing an issue with trying to open the database after a disk-full issue (which resulted in the database being unable to write, and hence not closed properly).
While recovery seems to be operating, it has been taking an inordinate amount of time - 16 hours so far. My database has data of around 200GB, which inflated to over 450GB during deletion of entries, hence gobbling up all free space on disk.
My questions are:
* Should i continue to wait for recovery?
* Is there any chance that recovery is looping?
* Is there an easier way (DBDump?) to extract data from the database without having to perform recovery?
Some other information that may help:
* The recovery has decreased the size of the last significant file, and created 3 new files since it started running.
* I have been monitoring the open files (using lsof), and they change every now and then to other files, though a good amount of its time is spent near the end of the database.
Thus, i feel like recovery is running normally, just taking too long. Please let me know your opinion.
A few other things i should mention regarding my issue:
* The database was, till yesterday, running on bdb java 3.3.75. After running several hours of recovery, i upgraded to 4.1.10 (since i read about a possible recovery looping bug in one of the versions)
* Once 4.1.10 started recovery, it spat out errors regarding the last 2 files. Only on deleting those 2 files (the last being 0 bytes, the 2nd-last being about 5k) did the recovery start. Note that the older 3.3.75's recovery never complained about those files. I can post the errors here if relevant.
* Some of the jdb files (about 500 files out of the 47,000 files that make up the database) are 100 MB files, since i had experimented with larger sized files for a few days, then reverted the setting.
Would any of these above affect a successful recovery?
My setup is:
OS:Linux CentOS 5.2, 64-bit, kernel 2.6.18-92.el5
JVM: Sun Java 1.6.0_20, 64-bit
Memory: 16 GB RAM, of which 8 GB is allocated to the java process (-Xmx8000M -Xms8000M)
BDB cache set to use 6GB RAM (envconfig.setCacheSize(6000000000))
Only the BDB basic API is being used (Environment, database, cursors). We do not use DPL, or HA features.
Awaiting your kind response,
Sushant A

Hi Sushant,
* Should i continue to wait for recovery?* Is there any chance that recovery is looping?>
I'm not aware of a bug that would cause recovery to loop, however, you may want to take thread dumps to see if it is progressing. It isn't easy to tell, however, since each phase of recovery is in fact a loop. What you can tell easily from the thread dumps is whether recovery is blocked (completely stopped) for some reason. I don't know of a bug that would cause this, but it's something I would check for.
Assuming it is not blocked, I suggest that you leave recovery running, and additionally (in parallel) try to obtain some information about your log. While recovery is running you can run the DbPrintLog utility, which does not itself run recovery. I suggest running the following command, which will tell us in general what your log looks like and in particular how far apart the checkpoints are:
java -jar je-x.y.z.jar DbPrintLog -h <envHome> -S > <output>Please post the output.
If checkpoints are not running in your application for some reason, or they are running very infrequently, this can cause VERY long recoveries. Unfortunately, you may have such a problem in your app and not be aware of it, until you crash and have to recover. To guard against this sort of thing in the future, you should keep an eye on the checkpoint frequency. EnvironmentStats.getNCheckpoints and getEndOfLog can together be used to tell how much log is written between checkpoints. We will also be able to see this from the DbPrintLog -S output.
* Is there an easier way (DBDump?) to extract data from the database without having to perform recovery?DbDump normally runs recovery. DbDump with the -r or -R option does not run recovery, but has other drawbacks. With -r, a large amount of memory may be necessary to dump an accurate representation of your data set. If this fails because you run out of memory, -R can be used, but this will dump multiple versions of each record and it will be up to you to interpret the output.
If regular recovery does not succeed, then DbDump -r is the next thing to try.
Would any of these above affect a successful recovery?No, I don't believe so.
--mark                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                           

Similar Messages

  • WebI Report is taking too long time to opening

    Hi All,
    When iam trying to open the WebI report in Infoview , it is taking long time to open and refresh,
    Please suggest me a solution.
    Thanks in advance..
    Regards,
    Mahesh

    Hi,
    As the issue you are facing is that the webi report is taking too long to open and refresh, I would recommend the below steps.
    1. Check whether the webi report is set to "Refreh on Open" if yes probably you need to uncheck, save the report and open it again.
    2. Try to run the same query in the backend database and see if it returns the data.
    3. Try to run refresh the report for a smaller data selection.
    4. make the report run on a specific webi server, and when refreshing have your BOBJ admin monitor that process to see if the process is going in a hung state, using High memory etc.
    5. restart webi process and run again
    Thanks,
    aKs

  • Discoverer report taking too long time to open.

    HI,
    Discovere reports are taking too long time to open. Please help to resolve this.
    Regards,
    Bhatia

    What is the Dicoverer and the Application release?
    Please refer to the following links (For both Discoverer 4i and 10g). Please note that some Discoverer 4i notes also apply to Discoverer 10g.
    Note: 362851.1 - Guidelines to setup the JVM in Apps Ebusiness Suite 11i and R12
    https://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=362851.1
    Note: 68100.1 - Discoverer Performance When Running On Oracle Applications
    https://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=68100.1
    Note: 465234.1 - Recommended Client Java Plug-in (JVM/JRE) For Discoverer Plus 10g (10.1.2)
    https://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=465234.1
    Note: 329674.1 - Slow Performance When Opening Plus Workbooks from Oracle 11.5.10 Applications Home Page
    https://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=329674.1
    Note: 190326.1 - Ideas for Improving Discoverer 4i Performance in an Applications 11i Environment
    https://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=190326.1
    Note: 331435.1 - Slow Perfomance Using Disco 4.1 Admin/Desktop in Oracle Applications Mode EUL
    https://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=331435.1
    Note: 217669.1 - Refreshing Folders and opening workbooks is slow in Apps 11i environment
    https://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=217669.1

  • Moving the 80 Million records from Conversion database to System Test database (Just for one transaction table) taking too long.

    Hello Friends,
    The background is I am working as conversion manager and we move the data from oracle to SQL Server using SSMA and then we will apply the conversion logic and then move the data to system test ,UAT and Production.
    Scenario:
    Moving the 80 Million records from Conversion database to System Test database (Just for one transaction table) taking too long. Both the databases are in the same server.
    Questions are…
    What is best option?
    IF we use the SSIS it’s very slow and taking 17 hours (some time it use to stuck and won’t allow us to do any process).
    I am using my own script (Stored procedure) and it’s taking only 1 hour 40 Min. I would like know is there any better process to speed up and why the SSIS is taking too long.
    When we move the data using SSIS do they commit inside after particular count? (or) is the Microsoft is committing all the records together after writing into Transaction Log
    Thanks
    Karthikeyan Jothi

    http://www.dfarber.com/computer-consulting-blog.aspx?filterby=Copy%20hundreds%20of%20millions%20records%20in%20ms%20sql
    Processing
    hundreds of millions records can be done in less than an hour.
    Best Regards,Uri Dimant SQL Server MVP,
    http://sqlblog.com/blogs/uri_dimant/
    MS SQL optimization: MS SQL Development and Optimization
    MS SQL Consulting:
    Large scale of database and data cleansing
    Remote DBA Services:
    Improves MS SQL Database Performance
    SQL Server Integration Services:
    Business Intelligence

  • I shut down when an upgrade was taking too long yesterday, now Itunes will not open today

    I shut down when an upgrade was taking too long yesterday, now Itunes will not open today

    Hello James,
    Thank you for using Apple Support Communities!
    From what you are describing, I have 2 recommendations. First, troubleshoot for 3rd party plugins that may be causing an issue after the interrupted update:
    iTunes: Troubleshooting issues with third-party iTunes plug-ins
    http://support.apple.com/kb/TS3430
    If the issue persists after followign the article, I would next download iTunes from the download page, and install it.
    http://www.apple.com/itunes/download/
    Take care,
    Sterling

  • My iPad is taking too long to contact update software server during recovery.

    My iPad is taking too long to contact update software server during recovery. It can go on for 30 minutes and it would not load. What can I do?

    If the links are blocked by who-ever, you will keep getting those errors.
    About the only option would be to use a proxy.
    http://www.bing.com/search?q=proxy

  • RMAN - upgrade catalog taking too long

    Hello ,
    I am trying to upgrade the rman catalog and it is taking too long to upgrade , infact after 4 hours wait time it is still excuting the upgrade command .
    Target database is : 11.2.0.3 PSU1
    catalog database : 11.2.0.2
    upgrading the rman catalog to 11.2.0.3 .
    I tried executing the command at a very quiet time of the day when catalog is not been used by any database . Please advice if you have encountered the similar issue and fixed it .
    Thanks
    Venkat

    After some time into execution of "upgrade catalog" i got the output as below , and after that i let all the jobs of rman execute completely from diff databases that are connected to catalog and re-executed the "upgrade catalog" command , which finally completed the upgrade process with out the issues .
    Error :
    RMAN> upgrade catalog;
    error creating upgcat_strt_0
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-06004: ORACLE error from recovery catalog database: ORA-00060: deadlock detected while waiting for resource
    Re-executed when no sessions are connected to catalog and the upgrade process is successful .
    RMAN> connect catalog rman/rman@database
    connected to recovery catalog database
    PL/SQL package RMAN.DBMS_RCVCAT version 11.02.00.02 in RCVCAT database is not cu
    PL/SQL package RMAN.DBMS_RCVMAN version 11.02.00.02 in RCVCAT database is not cu
    RMAN> upgrade catalog;
    recovery catalog owner is RMAN
    enter UPGRADE CATALOG command again to confirm catalog upgrade
    RMAN> upgrade catalog;
    recovery catalog upgraded to version 11.02.00.03
    DBMS_RCVMAN package upgraded to version 11.02.00.03
    DBMS_RCVCAT package upgraded to version 11.02.00.03
    Thanks
    Venkat

  • Importing a table with a BLOB column is taking too long

    I am importing a user schema from 9i (9.2.0.6) database to 10g (10.2.1.0) database. One of the large tables (millions of records) with a BLOB column is taking too long to import (more that 24 hours). I have tried all the tricks I know to speed up the import. Here are some of the setting:
    1 - set buffer to 500 Mb
    2 - pre-created the table and turned off logging
    3 - set indexes=N
    4 - set constraints=N
    5 - I have 10 online redo logs with 200 MB each
    6 - Even turned off logging at the database level with disablelogging = true
    It is still taking too long loading the table with the BLOB column. The BLOB field contains PDF files.
    For your info:
    Computer: Sun v490 with 16 CPUs, solaris 10
    memory: 10 Gigabytes
    SGA: 4 Gigabytes

    Legatti,
    I have feedback=10000. However by monitoring the import, I know that its loading average of 130 records per minute. Which is very slow considering that the table contains close to two millions records.
    Thanks for your reply.

  • When query is taking too long time

    When query is taking too long time,Where and how to start tuning it?
    Here i've listed few things need to be considered,out of my knowledge and understanding
    1.What the sql is waiting for(wait events)
    2.Parameter modification need to be done at system/session level
    3.The query has to be tuned (using hints )
    4.Gathering/deleting statistics
    List out any other things that need to be taken into account?
    Which approach must be followed and on what basis that approach must be considered?

    When query is taking too long time,Where and how to start tuning it?explain plan will be good start . trace also
    Here i've listed few things need to be considered,out of my knowledge and understanding
    1.What the sql is waiting for(wait events)When Oracle executes an SQL statement, it is not constantly executing. Sometimes it has to wait for a specific event to happen befor it can proceed.
    Read
    http://www.adp-gmbh.ch/ora/tuning/event.html
    2.Parameter modification need to be done at system/session levelDepend on parameter , define parameter , trace done on session level for example
    3.The query has to be tuned (using hints )Could be help you but you must know how to use .
    4.Gathering/deleting statisticsDo it in non working hours , it will impact on database performance , but its good
    List out any other things that need to be taken into account?Which account ?
    Which approach must be followed and on what basis that approach must be considered?you could use lot of tools , Trace , AWR

  • Query taking too long

    I am running a fairly complex query with several table joins
    and it is taking too long. What can I do to improve performance?
    Thanks.
    Frank

    Dan's first suggestion is key - if you are doing multiple
    table joins, you want to make sure your indexes are set up on your
    tables correctly. If you have access to the database, this should
    be your first step. Rationalize's stored procedure suggestion is
    also a great idea (again, if you have access to create and manage
    stored procedures on your DB).
    Other than than, most databases usually have some sort of SQL
    efficiency analysis tool. SQL server has one built into their Query
    Analyzer tool. I would recommend using something like that to
    streamline your SQL. Like Dan said, something as simple as the
    order of elements in your where clause might make a big
    difference.

  • Query taking too long on Oracle9i

    Hi All
    I am running a query on our prod database (Oracle8i 8.1.7.4) and again running the same query on Test db (Oracle9i version 4). The query is taking too long(more then 10 min) in test db. Both the database are installed on the same machine IBM AIX V4 and table schema and data are same.
    Any help would be appreciated.
    Here are the results.
    FASTER ONE
    ORACLE 8i using Production
    Statistics
    864 recursive calls
    68 db block gets
    159855 consistent gets
    20297 physical reads
    0 redo size
    1310148 bytes sent via SQL*Net to client
    68552 bytes received via SQL*Net from client
    1036 SQL*Net roundtrips to/from client
    28 sorts (memory)
    1 sorts (disk)
    15525 rows processed
    SLOWER ONE
    ORACLE 9i using Test
    Statistics
    819 recursive calls
    80 db block gets
    22981568 consistent gets
    1361 physical reads
    0 redo size
    1194902 bytes sent via SQL*Net to client
    34193 bytes received via SQL*Net from client
    945 SQL*Net roundtrips to/from client
    0 sorts (memory)
    1 sorts (disk)
    14157 rows processed

    319404-
    To help us better understand the problem,
    1) Could you post your execution plan on the two different databases?
    2) Could you list indexes (if any, on these tables)?
    3) Are any of the objects in the 'from list' a view?
    If so, are you using a user defined function to create the view?
    4) Why are you using the table 'cal_instance_relationship' twice in the 'from ' clause'?
    5) Can't your query be the following?
    SELECT f.person_id, f.course_cd, cv.responsible_org_unit_cd cowner, f.fee_cal_type Sem, f.fee_ci_sequence_number seq_no,
    sua.unit_cd, uv.owner_org_unit_cd uowner, uv.supervised_contact_hours hours, 0 chg_rate, sum(f.transaction_amount) tot_fee,
    ' ' tally
    FROM unit_version uv,
    cal_instance_relationship cir1,
    chg_method_apportion cma,
    student_unit_attempt sua,
    course_version cv,
    fee_ass f
    WHERE f.fee_type = 'VET-MATFEE'
    AND f.logical_delete_dt IS NULL
    AND f.s_transaction_type IN ('ASSESSMENT', 'MANUAL ADJ')
    AND f.fee_ci_sequence_number > 400
    AND f.course_cd = cv.course_cd
    AND cv.version_number = (SELECT MAX(v.version_number) FROM course_version v
    WHERE v.course_cd = cv.course_cd)
    AND f.person_id = sua.person_id
    and f.course_cd = sua.course_cd
    AND f.fee_type = cma.fee_type
    AND f.fee_ci_sequence_number = cma.fee_ci_sequence_number
    AND cma.load_ci_sequence_number = cir1.sub_ci_sequence_number
    AND cir1.sup_cal_type = 'ACAD-YR'
    AND cir1.sub_cal_type = sua.cal_type
    AND cir1.sub_ci_sequence_number = sua.ci_sequence_number
    AND sua.unit_attempt_status NOT IN ('DUPLICATE','DISCONTIN')
    AND sua.unit_cd = uv.unit_cd
    AND sua.version_number = uv.version_number
    GROUP BY f.person_id, f.course_cd, cv.responsible_org_unit_cd , f.fee_cal_type, f.fee_ci_sequence_number,
    sua.unit_cd, uv.owner_org_unit_cd, uv.supervised_contact_hours;

  • Network event is taking too long (100%)

    Hi everybody. We have a 10g DB on Windows. We're using OEM to manage the DB, and it has started to show an alert about database time spend waiting for "Network" event. It arises when we execute one module that updates several tables, that is taking too long. Before we had this app on 8i, also on Windows, and that operation was much more faster than now. The indexes on tables are valid, and I've gathered statistics for the CBO, so I suppose the problem is, as the OEM says, related with network, but I don't know why, because the connection speed is the same tan before, and the two machines are in the same LAN.
    Any ideas??

    Here is the output requested:
    SQL> select * from v$system_event
    2 where event like 'SQL%';
    EVENT TOTAL_WAITS TOTAL_TIMEOUTS TIME_WAITED
    AVERAGE_WAIT TIME_WAITED_MICRO EVENT_ID
    SQL*Net message to client 1159200 0 252
    0 2516408 2067390145
    SQL*Net message to dblink 2234 0 1
    0 5590 3655533736
    SQL*Net more data to client 5753 0 166
    0 1657387 554161347
    EVENT TOTAL_WAITS TOTAL_TIMEOUTS TIME_WAITED
    AVERAGE_WAIT TIME_WAITED_MICRO EVENT_ID
    SQL*Net more data to dblink 12 0 0
    0 548 1958556342
    SQL*Net message from client 1159181 0 218341084
    188 2,1834E+12 1421975091
    SQL*Net more data from client 23299 0 180602
    8 1806015123 3530226808
    EVENT TOTAL_WAITS TOTAL_TIMEOUTS TIME_WAITED
    AVERAGE_WAIT TIME_WAITED_MICRO EVENT_ID
    SQL*Net message from dblink 2234 0 3693
    2 36934861 4093028837
    SQL*Net more data from dblink 4021 0 39
    0 390002 1136294303
    SQL*Net break/reset to client 182986 0 2740
    0 27397165 1963888671
    9 filas seleccionadas.

  • Export PDF taking too long to upload to Adobe export online. Any suggestions?

    I just recently purchased adobe exportPDF and I am trying to convert the PDF to word, but it is just taking too long even after opening multiple times.

    Hi there Huapopicante,
    How large/complex is the file that you're trying to convert? It may be that the file is too complex for the ExportPDF service to convert before timing out. Here are a few things that you can try, however:
    Clear the browser cache and try again.
    Make sure that you're using a supported web browser (http://www.adobe.com/acom/systemreqs/)
    Try a different browser.
    Make sure that there are no firewall/proxy settings that are limiting your access to the Internet.
    Convert the PDF file from within Reader with OCR disabled, as described in this document: How to disable Optical Character Recognition (OCR) when converting PDF to Word or Excel.
    Please let us know how it goes.
    Best,
    Sara

  • HT5191 (Windows 7/64bit) iTunes Taking too long to Backup.

    Hi guys/gurus/"i" masters,
    Got a MAJOR issue with itunes. I don't know why this happens on windows but this is one issue i can't solve.
    I'm on an iPhone 4s running ios 5.0.1. I want to backup all my stuff before I bring to the latest firmware (5.1.1)
    I've had to force close the program many times as the backup was taking too long, and it didn't even move a bit (backup bar).
    Can anyone please help me to resolve this issue.
    Appreciate the help very much.
    Regards,
    TJB!

    Start by making sure you have the latest version of iTunes.
    If that doesn't do it, you may be seeing something like what is described in this Apple article -> Apple software on Windows: May see performance issues and blank iTunes Store
    If it's what you're seeing, you can try resetting Winsock as a temporary fix or folllow along with the article.
    1. To do so, please do the following in an administrator user:
    Windows Vista and Windows 7: Right click All Programs and choose Accessories > Command Prompt. Choose to open as an administrator.
    Windows XP: Click Start > Run and type in cmd.
    2. Enter the following command: 
    netsh winsock reset
    3. Restart the computer and test the issue.
    Note:  If the issue returns, third-party software may be reinserting the LSP.

  • Update statement taking too long to execute

    Hi All,
    I'm trying to run this update statement. But its taking too long to execute.
        UPDATE ops_forecast_extract b SET position_id = (SELECT a.row_id
            FROM s_postn a
            WHERE UPPER(a.desc_text) = UPPER(TRIM(B.POSITION_NAME)))
            WHERE position_level = 7
            AND b.am_id IS NULL;
            SELECT COUNT(*) FROM S_POSTN;
            214665
            SELECT COUNT(*) FROM ops_forecast_extract;
            49366
    SELECT count(*)
            FROM s_postn a, ops_forecast_extract b
            WHERE UPPER(a.desc_text) = UPPER(TRIM(B.POSITION_NAME));
    575What could be the reason for update statement to execute so long?
    Thanks

    polasa wrote:
    Hi All,
    I'm trying to run this update statement. But its taking too long to execute.
    What could be the reason for update statement to execute so long?You haven't said what "too long" means, but a simple reason could be that the scalar subquery on "s_postn" is using a full table scan for each execution. Potentially this subquery gets executed for each row of the "ops_forecast_extract" table that satisfies your filter predicates. "Potentially" because of the cunning "filter/subquery optimization" of the Oracle runtime engine that attempts to cache the results of already executed instances of the subquery. Since the in-memory hash table that holds these cached results is of limited size, the optimization algorithm depends on the sort order of the data and could suffer from hash collisions it's unpredictable how well this optimization works in your particular case.
    You might want to check the execution plan, it should tell you at least how Oracle is going to execute the scalar subquery (it doesn't tell you anything about this "filter/subquery optimization" feature).
    Generic instructions how to generate a useful explain plan output and how to post it here follow:
    Could you please post an properly formatted explain plan output using DBMS_XPLAN.DISPLAY including the "Predicate Information" section below the plan to provide more details regarding your statement. Please use the {noformat}[{noformat}code{noformat}]{noformat} tag before and {noformat}[{noformat}/code{noformat}]{noformat} tag after or the {noformat}{{noformat}code{noformat}}{noformat} tag before and after to enhance readability of the output provided:
    In SQL*Plus:
    SET LINESIZE 130
    EXPLAIN PLAN FOR <your statement>;
    SELECT * FROM TABLE(DBMS_XPLAN.DISPLAY);Note that the package DBMS_XPLAN.DISPLAY is only available from 9i on.
    In 9i and above, if the "Predicate Information" section is missing from the DBMS_XPLAN.DISPLAY output but you get instead the message "Plan table is old version" then you need to re-create your plan table using the server side script "$ORACLE_HOME/rdbms/admin/utlxplan.sql".
    In previous versions you could run the following in SQL*Plus (on the server) instead:
    @?/rdbms/admin/utlxplsA different approach in SQL*Plus:
    SET AUTOTRACE ON EXPLAIN
    <run your statement>;will also show the execution plan.
    In order to get a better understanding where your statement spends the time you might want to turn on SQL trace as described here:
    When your query takes too long ...
    and post the "tkprof" output here, too.
    Regards,
    Randolf
    Oracle related stuff blog:
    http://oracle-randolf.blogspot.com/
    SQLTools++ for Oracle (Open source Oracle GUI for Windows):
    http://www.sqltools-plusplus.org:7676/
    http://sourceforge.net/projects/sqlt-pp/

Maybe you are looking for

  • Not able to send any SMS

    Hi All,  I am not able to send any SMS  from my blackberry device from past one day. I have checked all possible things including my network, the message settings etc. I am able to choose contacts and type messages but when I am trying to send it it

  • How to view recent threads

    In case it's any use to anyone, I discovered that it's possible to get a list of all recent message threads by going to URL http://community.bt.com/t5/forums/recentpostspage/post-type/thread They are sorted by the date/time of the most recent post to

  • Change NAME of User Account, Home Directory, Time Capsule and Network

    Hi All. I have a small home network and I would like to do this for all four of these. I think I have instructions on doing this for User Account, Home Directory and Time Capsule. Can anyone give me the "it is OK to do this" so I don't have to worry

  • Clear a table from cache

    hi there, I am currently trying to analyse the number of physical reads needed to perform certain queries to see which one is best. I am not getting accurate results because the table is being cached after the first query. Basically, my question is h

  • Live Monitoring and Recording for VPN Remote CAD Agents in UCCX

    Dear all, We have deployed UCCX 5.0(2) for one of our customers. Silent Monitoring and Recording are working fine for local agents. We have deployed new Remote agents using IP Communicator and Cisco VPN on their PC. Everything is working fine but cou