Global Cache Configuration - CacheGlobalConfigurationPG

Hi All,
I have 11.5.10.2 running on SLES 10.
Had a question - I want to run the following Functional Administrator work via some sort of script as a scheduled job:
1. Login with Functional Administrator responsibility.
2. Go to caching framework tab; then click on 'Global Configuration'
3. Click on 'Clear All Cache' button
so basically want to run an html form based task via some Linux shell script. I am very comfortable with cron and shell scripting; all I am looking for is what package/procedure or something similar could I call in my script to do 'Clear All Cache' that this button does.
Thanks all,

Please see (How To Clear Caches (Apache/iAS, Cabo, Modplsql, Browser, Jinitiator, Java, Portal, WebADI) for E-Business Suite? [ID 742107.1]), under "Apache / iAS" section.
Thanks,
Hussein

Similar Messages

  • How Insert Work on global cache group?

    Hi all , i'm doing some test about how many transactions for second TimesTen can process.
    With a normal configuration "direct" i reached 5200 transaction for second, on my machine (OS windows normal work station).
    now i'm using the global cache groups because we need more then one DataSource , and they have to be syncro, one with each other.
    And how i read in the guide the global cache group are perfect for this purpose.
    After configured the 2 environment with different DataBase TimesTen (those machine are server SUN, much better of my work station :P), i tried a simple test
    of insert on a single node.
    But i reached only 1500 as maximum value of transactions for second.
    The 5200 value when testing on my work station was with normal Dynamic Cache Group, not Global. So i was thinking if this performance issue was related on how the Insert statement work on a global cache group.
    Some questions:
    1) before the insert is done on Oracle, the Cache Group do some query on the other cache global group to avoid some conflicts on primary key?
    2) there is any operation performed from global cache to others when a statement is sendend?
    The 2 global cache anyway are working well, locking and changing owner on a instance cache so no problems detected atm are about " how they have to work":).
    The problem is only that we need that the global cache do it more and more faster :P at last the 5200 transaction for second i reached on my work station.
    Thanks in advance for any suggestion.
    P.S.:I don't know much about the server configuraion (SO solaris some version) but anyway good machines :).

    Okay, the rows here are quite large so you need to do some tuning. In the ODBC (DSN) parameters I see that you are using the default log buffer abd log file sizes. these are totally inadequate for this kind of workload. You should increase both to a larger value. For this kind of workloads typial values would be in the range of 256 MB to 1024 MB for both log buffer and log file size. If you are using 32-bit TimesTen you may be constrained on how large you can make these sicne the log buffer is part of the overall datastore memory allocation wh9ich on 32-bit platforms is quite limited. On 64-bit TimesTen you have no such restriction (as long as the machine has enough memory). Here is an example of the directives you would use to set both to 1 GB. The key one is the log buffer size but it is important that LogFileSize is >= LogBufMB.
    [my_ds]
    LogBufMB=1024
    LogFileSize=1024
    For this change to take effect you need to shutdown (unload from memory) and restart (load back into memory) the datastore.
    Secondly, it's hard to be sure from your example code but it looks like maybe you are pre-paring the INSERT each time you execute it? If that is the case this is very expensive and unnecessary. You only need to prepare once and then you can execute many times as follows:
    insPs = connection.prepareStatement("Insert into test.transactions (ID_ ,NUMBE,SHORT_CODE,REQUEST_TIME) Values (?,?,?,?)");
    for (int i=1; i < 1000000; i++)
    insPs.setString(1,""+getSequence());
    insPs.setString(2,"TEST_CODE");
    insPs.setString(3,"TT Insert test");
    insPs.setTimestamp(4, new Timestamp(System.currentTimeMillis()));
    insPs.execute();
    connection.commit();
    This should improve performance noticeably. mif you can get away with only comiting every 'N' inserts you will see a further uplift. For example:
    int COMMIT_INTVL = 100;
    for (int i=1; i < 1000000; i++)
    insPs.setString(1,""+getSequence());
    insPs.setString(2,"TEST_CODE");
    insPs.setString(3,"TT Insert test");
    insPs.setTimestamp(4, new Timestamp(System.currentTimeMillis()));
    insPs.execute();
    if ( (i % COMMIT_INTVL) == 0 )
    connection.commit();
    connection.commit();
    And lastly, the fastest way of all is to use JDBC batch operations; see the JDBC documentation about batch operations. That will improve insert performance still more.
    Lastly, a word of caution. Although you will probably be able to easily achieve more than 5000 inserts per second into TimesTen, TimesTen may not be able to push the data to oracle at this rate. the rate of push to Oracle is likely to be significantly slower. Thus if you are executing a continuous high volume insert workload into TimesTen two things will happen; (a) the datastore will become fiull and unable to accepot any more inserts until you explicitly remove some data and (b) a backlog will build up (in the TT transaction logs on disk) of data waiting to be pushed to Oracle.
    This kind of setup is not really suited to support sustained high insert levels; you need to look at the maximum that can be sustained for the whole application -> TimesTen -> Oracle pathway. Of course, if the workload is 'bursty' then this may not be an issue at all.
    Chris

  • EM Alert: Critical: Metrics "Global Cache Average Current Get Time" is at X

    I frequently get this notification email alert from the enterprise manager. What does this mean and should I ignore it or work on it?
    Name=<sid>_<instancename>
    Type=Database Instance
    Host=<servername>.<domainname>
    Metric=Global Cache Average Current Block Request Time (centi-seconds)
    Timestamp=Jan 27, 2009 9:27:20 PM GST
    Severity=Critical
    Message=Metrics "Global Cache Average Current Get Time" is at 1.46154
    Rule Name=XXProd
    Rule Owner=SYS
    [email protected]

    those metrics shows the effect of accessing blocks in the global cache and maintaining cache coherency.
    basically, the response time for Cache Fusion transfers is determined by the messaging time and
    processing time imposed by the physical interconnect components, the IPC protocol, and the
    GCS protocol.
    therefore, Inter-Instance performance issues can be caused by:
    1) Under configured network settings at the OS
    2) dropped packets,retransmits, or cyclic redundancy check errors (CRC)
    3) large number of processes in the run queue waiting for CPU
    4) high value for the DB_FILE_MULTIBLOCK_READ_COUNT
    Note also that poor SQL or bad optimization paths can cause additional block gets via the interconnect, just as not using ASSM for Locally Managed Tablespaces, or not using CACHE NOORDER for sequence, ...

  • Express to Global Cache IR repeater

    I want to have a global cache IR send signals from Cinemar DVD Lobby to my Sony DVD changer. The Global Cache can get input from ethernet and convert it to IR.
    My computer is in the office, the DVD player in the living room.
    Can I connect the Global Cache to the Airport Express and send the data wirelessly from my computer? Or is the ethernet port on the Express only for connecting a modem?
    I need the Express to receive and send to my Airport Base Station and send the info out to the controller via the ethernet port. Make sense?

    To activate the Ethernet port on the AirPort Express (AX), it must be wirelessly connected to a network using WDS.
    You can find directions for configure your AirPort Extreme base station (AEBS) and AX in KB 107454, AirPort Extreme and Express: Using WDS to create a network from multiple base stations.

  • Global-Cache-Manager for Multi-Environment Applications

    Hi,
    Within our server implementation we provide a "multi-project" environment. Each project is fully isolated from the rest of the server e.g. in terms of file-system usage, backup and other ressources. As one might expect the way to go is using a single VM with multiple BDB environments.
    Obviously each JE-Environment uses its own cache. Within a our environment with dynamic numbers of active projects this causes a problem because the optimal cache configuration within a given memory frame depends on the JE-Environments in use BUT there is no way to define a global JE cache for ALL JE-Environments.
    Our "plan of attack" is to implement a Global-Cache-Manager to dynamicly configure the cache sizes of all active BDB environments depending on the given global cache size.
    Like Federico proposed the starting point for determining the optimal cache setting at load time will be a modification to the DbCacheSize utility so that the return value can be picked up easily, rather than printed to stdout. After that the EnvironmentMutableConfig.setCacheSize will be used to set the cache size. If there is enough Cache-RAM available we could even set a larger cache but I do not know if that really makes sense.
    If Cache-Memory is getting tight loading another BDB environment means decreasing cache sizes for the already loaded environments. This is also done via EnvironmentMutableConfig.setCacheSize. Are there any timing conditions one should obey before assuming the memory is really available? To determine if there are any BDB environments that do not use their cache one could query each cache utilization using EnvironmentStats.getCacheDataBytes() and getCacheTotalBytes().
    Are there any comments to this plan? Is there perhaps a better solution or even an implementation?
    Do you think a global cache manager is something worth back-donating?
    Related Postings: Multiple envs in one process?
    Stefan Walgenbach

    Here is the updated DbCacheSize.java to allow calling it with an API.
    Charles Lamb
    * See the file LICENSE for redistribution information.
    * Copyright (c) 2005-2006
    *      Oracle Corporation.  All rights reserved.
    * $Id: DbCacheSize.java,v 1.8 2006/09/12 19:16:59 cwl Exp $
    package com.sleepycat.je.util;
    import java.io.File;
    import java.io.PrintStream;
    import java.math.BigInteger;
    import java.text.NumberFormat;
    import java.util.Random;
    import com.sleepycat.je.Database;
    import com.sleepycat.je.DatabaseConfig;
    import com.sleepycat.je.DatabaseEntry;
    import com.sleepycat.je.DatabaseException;
    import com.sleepycat.je.Environment;
    import com.sleepycat.je.EnvironmentConfig;
    import com.sleepycat.je.EnvironmentStats;
    import com.sleepycat.je.OperationStatus;
    import com.sleepycat.je.dbi.MemoryBudget;
    import com.sleepycat.je.utilint.CmdUtil;
    * Estimating JE in-memory sizes as a function of key and data size is not
    * straightforward for two reasons. There is some fixed overhead for each btree
    * internal node, so tree fanout and degree of node sparseness impacts memory
    * consumption. In addition, JE compresses some of the internal nodes where
    * possible, but compression depends on on-disk layouts.
    * DbCacheSize is an aid for estimating cache sizes. To get an estimate of the
    * in-memory footprint for a given database, specify the number of records and
    * record characteristics and DbCacheSize will return a minimum and maximum
    * estimate of the cache size required for holding the database in memory.
    * If the user specifies the record's data size, the utility will return both
    * values for holding just the internal nodes of the btree, and for holding the
    * entire database in cache.
    * Note that "cache size" is a percentage more than "btree size", to cover
    * general environment resources like log buffers. Each invocation of the
    * utility returns an estimate for a single database in an environment.  For an
    * environment with multiple databases, run the utility for each database, add
    * up the btree sizes, and then add 10 percent.
    * Note that the utility does not yet cover duplicate records and the API is
    * subject to change release to release.
    * The only required parameters are the number of records and key size.
    * Data size, non-tree cache overhead, btree fanout, and other parameters
    * can also be provided. For example:
    * $ java DbCacheSize -records 554719 -key 16 -data 100
    * Inputs: records=554719 keySize=16 dataSize=100 nodeMax=128 density=80%
    * overhead=10%
    *    Cache Size      Btree Size  Description
    *    30,547,440      27,492,696  Minimum, internal nodes only
    *    41,460,720      37,314,648  Maximum, internal nodes only
    *   114,371,644     102,934,480  Minimum, internal nodes and leaf nodes
    *   125,284,924     112,756,432  Maximum, internal nodes and leaf nodes
    * Btree levels: 3
    * This says that the minimum cache size to hold only the internal nodes of the
    * btree in cache is approximately 30MB. The maximum size to hold the entire
    * database in cache, both internal nodes and datarecords, is 125Mb.
    public class DbCacheSize {
        private static final NumberFormat INT_FORMAT =
            NumberFormat.getIntegerInstance();
        private static final String HEADER =
            "    Cache Size      Btree Size  Description\n" +
        //   12345678901234  12345678901234
        //                 12
        private static final int COLUMN_WIDTH = 14;
        private static final int COLUMN_SEPARATOR = 2;
        private long records;
        private int keySize;
        private int dataSize;
        private int nodeMax;
        private int density;
        private long overhead;
        private long minInBtreeSize;
        private long maxInBtreeSize;
        private long minInCacheSize;
        private long maxInCacheSize;
        private long maxInBtreeSizeWithData;
        private long maxInCacheSizeWithData;
        private long minInBtreeSizeWithData;
        private long minInCacheSizeWithData;
        private int nLevels = 1;
        public DbCacheSize (long records,
                   int keySize,
                   int dataSize,
                   int nodeMax,
                   int density,
                   long overhead) {
         this.records = records;
         this.keySize = keySize;
         this.dataSize = dataSize;
         this.nodeMax = nodeMax;
         this.density = density;
         this.overhead = overhead;
        public long getMinCacheSizeInternalNodesOnly() {
         return minInCacheSize;
        public long getMaxCacheSizeInternalNodesOnly() {
         return maxInCacheSize;
        public long getMinBtreeSizeInternalNodesOnly() {
         return minInBtreeSize;
        public long getMaxBtreeSizeInternalNodesOnly() {
         return maxInBtreeSize;
        public long getMinCacheSizeWithData() {
         return minInCacheSizeWithData;
        public long getMaxCacheSizeWithData() {
         return maxInCacheSizeWithData;
        public long getMinBtreeSizeWithData() {
         return minInBtreeSizeWithData;
        public long getMaxBtreeSizeWithData() {
         return maxInBtreeSizeWithData;
        public int getNLevels() {
         return nLevels;
        public static void main(String[] args) {
            try {
                long records = 0;
                int keySize = 0;
                int dataSize = 0;
                int nodeMax = 128;
                int density = 80;
                long overhead = 0;
                File measureDir = null;
                boolean measureRandom = false;
                for (int i = 0; i < args.length; i += 1) {
                    String name = args;
    String val = null;
    if (i < args.length - 1 && !args[i + 1].startsWith("-")) {
    i += 1;
    val = args[i];
    if (name.equals("-records")) {
    if (val == null) {
    usage("No value after -records");
    try {
    records = Long.parseLong(val);
    } catch (NumberFormatException e) {
    usage(val + " is not a number");
    if (records <= 0) {
    usage(val + " is not a positive integer");
    } else if (name.equals("-key")) {
    if (val == null) {
    usage("No value after -key");
    try {
    keySize = Integer.parseInt(val);
    } catch (NumberFormatException e) {
    usage(val + " is not a number");
    if (keySize <= 0) {
    usage(val + " is not a positive integer");
    } else if (name.equals("-data")) {
    if (val == null) {
    usage("No value after -data");
    try {
    dataSize = Integer.parseInt(val);
    } catch (NumberFormatException e) {
    usage(val + " is not a number");
    if (dataSize <= 0) {
    usage(val + " is not a positive integer");
    } else if (name.equals("-nodemax")) {
    if (val == null) {
    usage("No value after -nodemax");
    try {
    nodeMax = Integer.parseInt(val);
    } catch (NumberFormatException e) {
    usage(val + " is not a number");
    if (nodeMax <= 0) {
    usage(val + " is not a positive integer");
    } else if (name.equals("-density")) {
    if (val == null) {
    usage("No value after -density");
    try {
    density = Integer.parseInt(val);
    } catch (NumberFormatException e) {
    usage(val + " is not a number");
    if (density < 1 || density > 100) {
    usage(val + " is not betwen 1 and 100");
    } else if (name.equals("-overhead")) {
    if (val == null) {
    usage("No value after -overhead");
    try {
    overhead = Long.parseLong(val);
    } catch (NumberFormatException e) {
    usage(val + " is not a number");
    if (overhead < 0) {
    usage(val + " is not a non-negative integer");
    } else if (name.equals("-measure")) {
    if (val == null) {
    usage("No value after -measure");
    measureDir = new File(val);
    } else if (name.equals("-measurerandom")) {
    measureRandom = true;
    } else {
    usage("Unknown arg: " + name);
    if (records == 0) {
    usage("-records not specified");
    if (keySize == 0) {
    usage("-key not specified");
         DbCacheSize dbCacheSize = new DbCacheSize
              (records, keySize, dataSize, nodeMax, density, overhead);
         dbCacheSize.caclulateCacheSizes();
         dbCacheSize.printCacheSizes(System.out);
    if (measureDir != null) {
    measure(System.out, measureDir, records, keySize, dataSize,
    nodeMax, measureRandom);
    } catch (Throwable e) {
    e.printStackTrace(System.out);
    private static void usage(String msg) {
    if (msg != null) {
    System.out.println(msg);
    System.out.println
    ("usage:" +
    "\njava " + CmdUtil.getJavaCommand(DbCacheSize.class) +
    "\n -records <count>" +
    "\n # Total records (key/data pairs); required" +
    "\n -key <bytes> " +
    "\n # Average key bytes per record; required" +
    "\n [-data <bytes>]" +
    "\n # Average data bytes per record; if omitted no leaf" +
    "\n # node sizes are included in the output" +
    "\n [-nodemax <entries>]" +
    "\n # Number of entries per Btree node; default: 128" +
    "\n [-density <percentage>]" +
    "\n # Percentage of node entries occupied; default: 80" +
    "\n [-overhead <bytes>]" +
    "\n # Overhead of non-Btree objects (log buffers, locks," +
    "\n # etc); default: 10% of total cache size" +
    "\n [-measure <environmentHomeDirectory>]" +
    "\n # An empty directory used to write a database to find" +
    "\n # the actual cache size; default: do not measure" +
    "\n [-measurerandom" +
    "\n # With -measure insert randomly generated keys;" +
    "\n # default: insert sequential keys");
    System.exit(2);
    private void caclulateCacheSizes() {
    int nodeAvg = (nodeMax * density) / 100;
    long nBinEntries = (records * nodeMax) / nodeAvg;
    long nBinNodes = (nBinEntries + nodeMax - 1) / nodeMax;
    long nInNodes = 0;
         long lnSize = 0;
    for (long n = nBinNodes; n > 0; n /= nodeMax) {
    nInNodes += n;
    nLevels += 1;
    minInBtreeSize = nInNodes *
         calcInSize(nodeMax, nodeAvg, keySize, true);
    maxInBtreeSize = nInNodes *
         calcInSize(nodeMax, nodeAvg, keySize, false);
         minInCacheSize = calculateOverhead(minInBtreeSize, overhead);
         maxInCacheSize = calculateOverhead(maxInBtreeSize, overhead);
    if (dataSize > 0) {
    lnSize = records * calcLnSize(dataSize);
         maxInBtreeSizeWithData = maxInBtreeSize + lnSize;
         maxInCacheSizeWithData = calculateOverhead(maxInBtreeSizeWithData,
                                  overhead);
         minInBtreeSizeWithData = minInBtreeSize + lnSize;
         minInCacheSizeWithData = calculateOverhead(minInBtreeSizeWithData,
                                  overhead);
    private void printCacheSizes(PrintStream out) {
    out.println("Inputs:" +
    " records=" + records +
    " keySize=" + keySize +
    " dataSize=" + dataSize +
    " nodeMax=" + nodeMax +
    " density=" + density + '%' +
    " overhead=" + ((overhead > 0) ? overhead : 10) + "%");
    out.println();
    out.println(HEADER);
    out.println(line(minInBtreeSize, minInCacheSize,
                   "Minimum, internal nodes only"));
    out.println(line(maxInBtreeSize, maxInCacheSize,
                   "Maximum, internal nodes only"));
    if (dataSize > 0) {
    out.println(line(minInBtreeSizeWithData,
                   minInCacheSizeWithData,
                   "Minimum, internal nodes and leaf nodes"));
    out.println(line(maxInBtreeSizeWithData,
                   maxInCacheSizeWithData,
    "Maximum, internal nodes and leaf nodes"));
    } else {
    out.println("\nTo get leaf node sizing specify -data");
    out.println("\nBtree levels: " + nLevels);
    private int calcInSize(int nodeMax,
                   int nodeAvg,
                   int keySize,
                   boolean lsnCompression) {
    /* Fixed overhead */
    int size = MemoryBudget.IN_FIXED_OVERHEAD;
    /* Byte state array plus keys and nodes arrays */
    size += MemoryBudget.byteArraySize(nodeMax) +
    (nodeMax * (2 * MemoryBudget.ARRAY_ITEM_OVERHEAD));
    /* LSN array */
         if (lsnCompression) {
         size += MemoryBudget.byteArraySize(nodeMax * 2);
         } else {
         size += MemoryBudget.BYTE_ARRAY_OVERHEAD +
    (nodeMax * MemoryBudget.LONG_OVERHEAD);
    /* Keys for populated entries plus the identifier key */
    size += (nodeAvg + 1) * MemoryBudget.byteArraySize(keySize);
    return size;
    private int calcLnSize(int dataSize) {
    return MemoryBudget.LN_OVERHEAD +
    MemoryBudget.byteArraySize(dataSize);
    private long calculateOverhead(long btreeSize, long overhead) {
    long cacheSize;
    if (overhead == 0) {
    cacheSize = (100 * btreeSize) / 90;
    } else {
    cacheSize = btreeSize + overhead;
         return cacheSize;
    private String line(long btreeSize,
                   long cacheSize,
                   String comment) {
    StringBuffer buf = new StringBuffer(100);
    column(buf, INT_FORMAT.format(cacheSize));
    column(buf, INT_FORMAT.format(btreeSize));
    column(buf, comment);
    return buf.toString();
    private void column(StringBuffer buf, String str) {
    int start = buf.length();
    while (buf.length() - start + str.length() < COLUMN_WIDTH) {
    buf.append(' ');
    buf.append(str);
    for (int i = 0; i < COLUMN_SEPARATOR; i += 1) {
    buf.append(' ');
    private static void measure(PrintStream out,
    File dir,
    long records,
    int keySize,
    int dataSize,
    int nodeMax,
    boolean randomKeys)
    throws DatabaseException {
    String[] fileNames = dir.list();
    if (fileNames != null && fileNames.length > 0) {
    usage("Directory is not empty: " + dir);
    Environment env = openEnvironment(dir, true);
    Database db = openDatabase(env, nodeMax, true);
    try {
    out.println("\nMeasuring with cache size: " +
    INT_FORMAT.format(env.getConfig().getCacheSize()));
    insertRecords(out, env, db, records, keySize, dataSize, randomKeys);
    printStats(out, env,
    "Stats for internal and leaf nodes (after insert)");
    db.close();
    env.close();
    env = openEnvironment(dir, false);
    db = openDatabase(env, nodeMax, false);
    out.println("\nPreloading with cache size: " +
    INT_FORMAT.format(env.getConfig().getCacheSize()));
    preloadRecords(out, db);
    printStats(out, env,
    "Stats for internal nodes only (after preload)");
    } finally {
    try {
    db.close();
    env.close();
    } catch (Exception e) {
    out.println("During close: " + e);
    private static Environment openEnvironment(File dir, boolean allowCreate)
    throws DatabaseException {
    EnvironmentConfig envConfig = new EnvironmentConfig();
    envConfig.setAllowCreate(allowCreate);
    envConfig.setCachePercent(90);
    return new Environment(dir, envConfig);
    private static Database openDatabase(Environment env, int nodeMax,
    boolean allowCreate)
    throws DatabaseException {
    DatabaseConfig dbConfig = new DatabaseConfig();
    dbConfig.setAllowCreate(allowCreate);
    dbConfig.setNodeMaxEntries(nodeMax);
    return env.openDatabase(null, "foo", dbConfig);
    private static void insertRecords(PrintStream out,
    Environment env,
    Database db,
    long records,
    int keySize,
    int dataSize,
    boolean randomKeys)
    throws DatabaseException {
    DatabaseEntry key = new DatabaseEntry();
    DatabaseEntry data = new DatabaseEntry(new byte[dataSize]);
    BigInteger bigInt = BigInteger.ZERO;
    Random rnd = new Random(123);
    for (int i = 0; i < records; i += 1) {
    if (randomKeys) {
    byte[] a = new byte[keySize];
    rnd.nextBytes(a);
    key.setData(a);
    } else {
    bigInt = bigInt.add(BigInteger.ONE);
    byte[] a = bigInt.toByteArray();
    if (a.length < keySize) {
    byte[] a2 = new byte[keySize];
    System.arraycopy(a, 0, a2, a2.length - a.length, a.length);
    a = a2;
    } else if (a.length > keySize) {
    out.println("*** Key doesn't fit value=" + bigInt +
    " byte length=" + a.length);
    return;
    key.setData(a);
    OperationStatus status = db.putNoOverwrite(null, key, data);
    if (status == OperationStatus.KEYEXIST && randomKeys) {
    i -= 1;
    out.println("Random key already exists -- retrying");
    continue;
    if (status != OperationStatus.SUCCESS) {
    out.println("*** " + status);
    return;
    if (i % 10000 == 0) {
    EnvironmentStats stats = env.getStats(null);
    if (stats.getNNodesScanned() > 0) {
    out.println("*** Ran out of cache memory at record " + i +
    " -- try increasing the Java heap size ***");
    return;
    out.print(".");
    out.flush();
    private static void preloadRecords(final PrintStream out,
    final Database db)
    throws DatabaseException {
    Thread thread = new Thread() {
    public void run() {
    while (true) {
    try {
    out.print(".");
    out.flush();
    Thread.sleep(5 * 1000);
    } catch (InterruptedException e) {
    break;
    thread.start();
    db.preload(0);
    thread.interrupt();
    try {
    thread.join();
    } catch (InterruptedException e) {
    e.printStackTrace(out);
    private static void printStats(PrintStream out,
    Environment env,
    String msg)
    throws DatabaseException {
    out.println();
    out.println(msg + ':');
    EnvironmentStats stats = env.getStats(null);
    out.println("CacheSize=" +
    INT_FORMAT.format(stats.getCacheTotalBytes()) +
    " BtreeSize=" +
    INT_FORMAT.format(stats.getCacheDataBytes()));
    if (stats.getNNodesScanned() > 0) {
    out.println("*** All records did not fit in the cache ***");

  • Error installing EM10gR2 Configuration assistant "Web Cache Configuration A

    I am getting an error when the web cache configuration assistant is running. Here is the output of the message:
    Output generated from configuration assistant "Web Cache Configuration Assistant" (attempt 2):
    Command = /oracle/oms10g/jdk/bin/java -DORACLE_HOME=/oracle/oms10g -Djava.net.preferIPv4Stack=true -classpath /oracle/oms10g/sysman/webapps/emd/WEB-INF/lib/webcachecfg.jar:/oracle/oms10g/sysman/webapps/emd/WEB-INF/lib/emd.jar:/oracle/oms10g/lib/emSDK.jar:/oracle/oms10g/dcm/lib/dcm.jar:/oracle/oms10g/jlib/repository.jar:/oracle/oms10g/lib/xmlparserv2.jar:/oracle/oms10g/jlib/emConfigInstall.jar:/oracle/oms10g/jdbc/lib/classes12.zip:/oracle/oms10g/jdbc/lib/nls_charset12.jar:/oracle/oms10g/jlib/jndi.jar:/oracle/oms10g/jlib/rt.jar:/oracle/oms10g/jdk/lib/i18n.jar:/oracle/oms10g/jlib/ldapjclnt10.jar:/oracle/oms10g/jlib/ojmisc.jar:/oracle/oms10g/j2ee/home/jazn.jar:/oracle/oms10g/j2ee/home/jaas.jar oracle.ias.webcache.config.WebcacheConfig -oh /oracle/oms10g
    Failed in smi manipulation().
    A problem has occurred reading the initial configuration and storing it into repository
    Resolution:
    Please refer to the base exception for resolution, or call Oracle support.
    Base Exception:
    /oracle/oms10g/Apache/Apache/bin/httpd: error while loading shared libraries: libdb.so.2: cannot open shared object file: No such file or directory
    Resolution:
    Please make sure the values entered in OHS configuration files are correct.
    oracle.ias.sysmgmt.exception.InvalidConfigurationException: Base Exception:
    /oracle/oms10g/Apache/Apache/bin/httpd: error while loading shared libraries: libdb.so.2: cannot open shared object file: No such file or directory
    Resolution:
    Please make sure the values entered in OHS configuration files are correct.
         at oracle.ias.sysmgmt.repository.plugin.advanced.apache.StateTranslator.checkConfigFileValidity(Unknown Source)
         at oracle.ias.sysmgmt.repository.plugin.advanced.apache.StateTranslator.validateConfigDuringEvaluate(Unknown Source)
         at oracle.ias.sysmgmt.repository.plugin.advanced.apache.PlugInImpl.localConfigValidation(Unknown Source)
         at oracle.ias.sysmgmt.repository.DcmPlugin.localConfigValidation(Unknown Source)
         at oracle.ias.sysmgmt.repository.RepositoryImpl.performLocalValidation(Unknown Source)
         at oracle.ias.sysmgmt.repository.SyncUpHandler._updatePluginConfigData(Unknown Source)
         at oracle.ias.sysmgmt.repository.SyncUpHandler.syncUpFromLocalFiles(Unknown Source)
         at oracle.ias.sysmgmt.repository.RepositoryImpl.syncUpFromLocalFiles(Unknown Source)
         at oracle.ias.sysmgmt.utility.editpropagator.PropagateLocalEdit.repositoryInit(Unknown Source)
         at oracle.ias.sysmgmt.persistence.utility.PMUtility.initConfiguration(Unknown Source)
         at oracle.ias.sysmgmt.task.TaskMaster.initConfiguration(Unknown Source)
         at oracle.ias.sysmgmt.task.TaskMaster.sysInit(Unknown Source)
         at oracle.ias.sysmgmt.task.TaskMaster.sysInit(Unknown Source)
         at oracle.ias.sysmgmt.task.InstanceManager.sysInit(Unknown Source)
         at oracle.ias.sysmgmt.task.InstanceManager.init(Unknown Source)
         at oracle.ias.sysmgmt.EntryPoint.init(Unknown Source)
         at oracle.ias.webcache.config.WebcacheConfig.smiSetup(WebcacheConfig.java:483)
         at oracle.ias.webcache.config.WebcacheConfig.smiEnableWebcache(WebcacheConfig.java:454)
         at oracle.ias.webcache.config.WebcacheConfig.enableConfiguration(WebcacheConfig.java:299)
         at oracle.ias.webcache.config.WebcacheConfig.main(WebcacheConfig.java:87)
    Configuration assistant "Web Cache Configuration Assistant" failed
    The "/oracle/agent10g/cfgtoollogs/configToolFailedCommands" script contains all commands that failed, were skipped or were cancelled. This file may be used to run these configuration assistants outside of OUI. Note that you may have to update this script with passwords (if any) before executing the same.-----------------------------------------------------------------------------
    My etc/hosts file looks like:
    cat /etc/hosts
    # Do not remove the following line, or various programs
    # that require network functionality will fail.
    127.0.0.1 localhost.localdomain localhost
    172.24.254.25 k25
    Ugh!

    sorry, a little bit late but i had the same problem.
    /oracle/oms10g/Apache/Apache/bin/httpd: error while loading shared libraries: libdb.so.2: cannot open shared object file: No such file or directory
    simple make the link "ln -s libdb.so.3 libdb.so.2" in the /usr/lib directory.
    regard´s
    roland

  • Top time events showing global cache  buffer busy waits

    can any one guide me to find the root cause for global cache buffer waits

    "Segments by Global Cache Buffer Busy" output from an AWR report.
    And let us know how many CPUs you have If you don't want to reveal the names of the objects, then change them (but do it in a way that means if two indexes are for the same table then it's visible). The distribution of waits is significant.
    How many of the indexes in the "segments" are based on an Oracle Sequence ? Check the values for the CACHE size of those sequences they probably ought to be at least 1,000.
    See note below about producing readable output.
    Regards
    Jonathan Lewis
    http://jonathanlewis.wordpress.com
    http://www.jlcomp.demon.co.uk
    To post code, statspack/AWR report, execution plans or trace files, start and end the section with the tag {noformat}{noformat} (lowercase, curly brackets, no spaces) so that the text appears in fixed format.
    There is a +"Preview"+ tab at the top of the text entry panel. Use this to check what your message will look like before you post the message. If it looks a complete mess you're unlikely to get a response. (Click on the +"Plain text"+ tab if you want to edit the text to tidy it up.)                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                           

  • Global cache

    Hi Experts
    I want to create Global Cache on a Query to improve the performance of the Query.In RSRT we have the options of
    Read Mode
             A Query to read all data at once
             X Query to read data during navigation
             H Quert to read when you navigate or expand hierarchies
    Cache Mode
               1 main memory cache without swapping
               2 main memory cache with swapping
               3 persistent cache per application server
               4 persistent cache across each application server
    Persistence Mode
               1 flat file
               2 cluster table
               3 transparent table(BLOB)
    Optimization Mode
          0 query will be optimised after generation
    Could you tell me which Mode I need to choose for the Query and how it affects.
    I have created one of these options and executed , and then reun my query many times, But still it takes same amount of time to run, Could you tell me whether it is caching or not and what is the procedure to set query for caching.

    files are only loaded in the system cache when specifically imported into the system cache using:
    javaws -import -system <url>
    /Andy

  • Node display off , Central data cache configured but not active

    Hi
    We get an error "Node display off , Central data cache configured but not active" in Central system while trying to display the monitored the abap instance of system.We just completed the CCMS configuration.
    we are getting this error in RZ20 (central system)
    Agent regristration is complete in monitored system and it is running fine.connection from central to monitored (abap instance) system is also ok.Please help asap.
    Operating System --AIX
    Sap version--ECC 6
    Database-db2
    Regards
    Shashi

    Hi,
    central data cache to improve the performance of the Alert Monitor when creating the alert monitoring trees.
    Check [This link|http://help.sap.com/saphelp_sm32/helpdata/en/de/13622d770b5e4f8398fc536b690e93/content.htm]
    And [This link|http://help.sap.com/saphelp_sm32/helpdata/en/10/75ea3b6d73e807e10000000a11402f/content.htm]
    Hope this solves your problem.
    Feel free to revert back.
    --Ragu

  • Node display off ,  Central data cache configured but not active while tryi

    Hi
    We get an error "Node display off ,  Central data cache configured but not active" in Central system while trying to display the monitored the abap instance of system.We just completed the CCMS configuration.
    Agent regristration is complete in monitored system and it is running fine.connection from central to monitored (abap instance) system is also ok.Please help asap.
    Operating System --AIX
    Sap version--ECC 6
    Database-db2
    Regards
    Chamba Shashi
    Edited by: SHASHI SAMAL on Apr 28, 2009 6:25 AM

    Hi,
    central data cache to improve the performance of the Alert Monitor when creating the alert monitoring trees.
    Check [This link|http://help.sap.com/saphelp_sm32/helpdata/en/de/13622d770b5e4f8398fc536b690e93/content.htm]
    And [This link|http://help.sap.com/saphelp_sm32/helpdata/en/10/75ea3b6d73e807e10000000a11402f/content.htm]
    Hope this solves your problem.
    Feel free to revert back.
    --Ragu

  • "Web Cache Configuration Assistant" failed during installation

    Hi,
    I'm trying to install Grid Control on Fedora Core 4 but it failes during the installation at the "Web Cache Configuration Assistant".
    I already createt a loopback device on localhost.
    Thanks for any surggestions!!
    End output from spawned process.
    Configuration assistant "Web Cache Configuration Assistant" failed
    2
    ORACLE_HOME is /opt/oracle/EM/10.1.0.3
    Failed in smi manipulation().
    TaskMaster initialization failed.
    Root Cause: Subscription request timed out after 120000 millseconds. Possible causes: OPMN may not be running, you may have an OPMN running in an alternate ORACLE_HOME using duplicate port values, or OPMN may be misconfigured.
    oracle.ons.SubscriptionException: Subscription request timed out after 120000 millseconds. Possible causes: OPMN may not be running, you may have an OPMN running in an alternate ORACLE_HOME using duplicate port values, or OPMN may be misconfigured.
    at oracle at oracle.ons.ONS.addSubscriber(ONS.java:333)
    at oracle.ons.Subscriber.realStartup(Subscriber.java:92)
    at oracle.ons.Subscriber.<init>(Subscriber.java:80)
    at oracle.ons.ONS.createNewSubscriber(ONS.java:687)
    at oracle.ias.sysmgmt.task.TaskMaster.sysInit(Unknown Source)
    at oracle.ias.sysmgmt.task.TaskMaster.sysInit(Unknown Source)
    at oracle.ias.sysmgmt.task.InstanceManager.sysInit(Unknown Source)
    at oracle.ias.sysmgmt.task.InstanceManager.init(Unknown Source)
    at oracle.ias.sysmgmt.EntryPoint.init(Unknown Source)
    at oracle.ias.webcache.config.WebcacheConfig.smiSetup(WebcacheConfig.java:243)
    at oracle.ias.webcache.config.WebcacheConfig.smiEnableWebcache(WebcacheConfig.java:215)
    at oracle.ias.webcache.config.WebcacheConfig.enableConfiguration(WebcacheConfig.java:191)
    at oracle.ias.webcache.config.WebcacheConfig.main(WebcacheConfig.java:66)
    .ons.SubscriptionNotification.waitForReply(SubscriptionNotification.java:82)
    cat /etc/hosts:
    127.0.0.1 loopback localhost
    127.0.0.1 fedora01 fedora01.barmenia.de localhost.localdomain localhost
    172.xxx.xxx.xxx fedora01.barmenia.de fedora01 oracleinstall

    hi,
    i tried to reinstall but it is still the same problem.
    before i started the installer there was no oracle process left
    fedora01.oracle[emrep]>ps -x
    Warning: bad syntax, perhaps a bogus '-'? See /usr/share/doc/procps-3.2.5/FAQ
    PID TTY STAT TIME COMMAND
    2976 pts/2 S 0:00 -ksh
    17573 pts/2 R+ 0:00 ps -x
    fedora01.oracle[emrep]>.1.0.3/Disk1/runInstaller PRE_REQUISITE=TRUE <
    Starting Oracle Universal Installer...
    before the assistant fails it looked like:
    Warning: bad syntax, perhaps a bogus '-'? See /usr/share/doc/procps-3.2.5/FAQ
    PID TTY STAT TIME COMMAND
    2976 pts/2 S+ 0:00 -ksh
    17588 pts/2 S 2:32 /tmp/OraInstall2005-09-27_02-31-00PM/jre/1.4.2/bin/java -Doracle.installer.librar
    17589 pts/2 S 0:00 /tmp/OraInstall2005-09-27_02-31-00PM/jre/1.4.2/bin/java -Doracle.installer.librar
    17590 pts/2 S 0:26 /tmp/OraInstall2005-09-27_02-31-00PM/jre/1.4.2/bin/java -Doracle.installer.librar
    17591 pts/2 S 0:00 /tmp/OraInstall2005-09-27_02-31-00PM/jre/1.4.2/bin/java -Doracle.installer.librar
    17592 pts/2 S 0:00 /tmp/OraInstall2005-09-27_02-31-00PM/jre/1.4.2/bin/java -Doracle.installer.librar
    17593 pts/2 S 0:00 /tmp/OraInstall2005-09-27_02-31-00PM/jre/1.4.2/bin/java -Doracle.installer.librar
    17594 pts/2 S 0:00 /tmp/OraInstall2005-09-27_02-31-00PM/jre/1.4.2/bin/java -Doracle.installer.librar
    17595 pts/2 S 0:02 /tmp/OraInstall2005-09-27_02-31-00PM/jre/1.4.2/bin/java -Doracle.installer.librar
    17596 pts/2 S 0:00 /tmp/OraInstall2005-09-27_02-31-00PM/jre/1.4.2/bin/java -Doracle.installer.librar
    17597 pts/2 S 0:00 /tmp/OraInstall2005-09-27_02-31-00PM/jre/1.4.2/bin/java -Doracle.installer.librar
    17598 pts/2 S 0:02 /tmp/OraInstall2005-09-27_02-31-00PM/jre/1.4.2/bin/java -Doracle.installer.librar
    17600 pts/2 S 0:00 /tmp/OraInstall2005-09-27_02-31-00PM/jre/1.4.2/bin/java -Doracle.installer.librar
    17601 pts/2 S 0:07 /tmp/OraInstall2005-09-27_02-31-00PM/jre/1.4.2/bin/java -Doracle.installer.librar
    17607 pts/2 S 0:02 /tmp/OraInstall2005-09-27_02-31-00PM/jre/1.4.2/bin/java -Doracle.installer.librar
    17677 pts/2 S 0:00 /tmp/OraInstall2005-09-27_02-31-00PM/jre/1.4.2/bin/java -Doracle.installer.librar
    26185 ? Ss 0:00 /opt/oracle/EM/10.1.0.3/opmn/bin/opmn -d
    26301 pts/2 S 0:00 /tmp/OraInstall2005-09-27_02-31-00PM/jre/1.4.2/bin/java -Doracle.installer.librar
    26381 pts/2 S 0:00 /tmp/OraInstall2005-09-27_02-31-00PM/jre/1.4.2/bin/java -Doracle.installer.librar
    26409 pts/2 S 0:00 /opt/oracle/EM/10.1.0.3/bin/tnslsnr LISTENER -inherit
    26669 pts/4 S 0:00 -ksh
    26717 ? S 0:00 /opt/oracle/EM/10.1.0.3/opmn/bin/opmn -d
    26718 ? S 0:00 /opt/oracle/EM/10.1.0.3/opmn/bin/opmn -d
    26719 ? S 0:00 /opt/oracle/EM/10.1.0.3/opmn/bin/opmn -d
    26720 ? S 0:00 /opt/oracle/EM/10.1.0.3/opmn/bin/opmn -d
    26721 ? S 0:00 /opt/oracle/EM/10.1.0.3/opmn/bin/opmn -d
    26722 ? S 0:00 /opt/oracle/EM/10.1.0.3/opmn/bin/opmn -d
    26723 ? S 0:00 /opt/oracle/EM/10.1.0.3/opmn/bin/opmn -d
    26724 ? S 0:00 /opt/oracle/EM/10.1.0.3/opmn/bin/opmn -d
    26725 ? S 0:00 /opt/oracle/EM/10.1.0.3/opmn/bin/opmn -d
    26726 ? S 0:00 /opt/oracle/EM/10.1.0.3/opmn/bin/opmn -d
    26727 ? S 0:00 /opt/oracle/EM/10.1.0.3/opmn/bin/opmn -d
    26728 26729 ? S 0:00 /opt/oracle/EM/10.1.0.3/opmn/bin/opmn -d
    26730 ? S 0:00 /opt/oracle/EM/10.1.0.3/opmn/bin/opmn -d
    26731 ? S 0:00 /opt/oracle/EM/10.1.0.3/opmn/bin/opmn -d
    26732 ? S 0:00 /opt/oracle/EM/10.1.0.3/opmn/bin/opmn -d
    26770 pts/2 S 0:00 /tmp/OraInstall2005-09-27_02-31-00PM/jre/1.4.2/bin/java -Doracle.installer.librar
    26771 pts/2 S 0:00 /tmp/OraInstall2005-09-27_02-31-00PM/jre/1.4.2/bin/java -Doracle.installer.librar
    26772 pts/2 S 0:01 /opt/oracle/EM/10.1.0.3/jdk/bin/java -DORACLE_HOME=/opt/oracle/EM/10.1.0.3 -class
    26773 pts/2 S 0:00 /tmp/OraInstall2005-09-27_02-31-00PM/jre/1.4.2/bin/java -Doracle.installer.librar
    26774 pts/2 S 0:00 /tmp/OraInstall2005-09-27_02-31-00PM/jre/1.4.2/bin/java -Doracle.installer.librar
    26775 pts/2 S 0:00 /opt/oracle/EM/10.1.0.3/jdk/bin/java -DORACLE_HOME=/opt/oracle/EM/10.1.0.3 -class
    26776 pts/2 S 0:00 /opt/oracle/EM/10.1.0.3/jdk/bin/java -DORACLE_HOME=/opt/oracle/EM/10.1.0.3 -class
    26777 pts/2 S 0:00 /opt/oracle/EM/10.1.0.3/jdk/bin/java -DORACLE_HOME=/opt/oracle/EM/10.1.0.3 -class
    26778 pts/2 S 0:00 /opt/oracle/EM/10.1.0.3/jdk/bin/java -DORACLE_HOME=/opt/oracle/EM/10.1.0.3 -class
    26779 pts/2 S 0:00 /opt/oracle/EM/10.1.0.3/jdk/bin/java -DORACLE_HOME=/opt/oracle/EM/10.1.0.3 -class
    26780 pts/2 S 0:00 /opt/oracle/EM/10.1.0.3/jdk/bin/java -DORACLE_HOME=/opt/oracle/EM/10.1.0.3 -class
    26781 pts/2 S 0:00 /opt/oracle/EM/10.1.0.3/jdk/bin/java -DORACLE_HOME=/opt/oracle/EM/10.1.0.3 -class
    26782 pts/2 S 0:00 /opt/oracle/EM/10.1.0.3/jdk/bin/java -DORACLE_HOME=/opt/oracle/EM/10.1.0.3 -class
    26783 pts/2 S 0:00 /opt/oracle/EM/10.1.0.3/jdk/bin/java -DORACLE_HOME=/opt/oracle/EM/10.1.0.3 -class
    26791 pts/2 S 0:00 /opt/oracle/EM/10.1.0.3/jdk/bin/java -DORACLE_HOME=/opt/oracle/EM/10.1.0.3 -class
    26792 pts/2 S 0:00 /opt/oracle/EM/10.1.0.3/jdk/bin/java -DORACLE_HOME=/opt/oracle/EM/10.1.0.3 -class
    26834 pts/2 S 0:00 /opt/oracle/EM/10.1.0.3/jdk/bin/java -DORACLE_HOME=/opt/oracle/EM/10.1.0.3 -class
    26835 pts/2 S 0:00 /bin/sh /opt/oracle/EM/10.1.0.3/opmn/bin/opmnctl stopall
    26843 pts/2 S 0:00 /opt/oracle/EM/10.1.0.3/opmn/bin/opmn -a -q shutdown
    26844 ? S 0:00 /opt/oracle/EM/10.1.0.3/opmn/bin/opmn -d
    ? S 0:00 /opt/oracle/EM/10.1.0.3/opmn/bin/opmn -d
    I'm running out of ideas!!
    thanks for your help!!!
    daniel

  • Central data cache configured but not active

    Hi,
    I am currently at a customer site and am configuring the solution manager system as a CEN.
    The sapccm4x en ccmsr agents have been registered and memory has been allocated in the CEN system for the central data cache, but whenever I open a monitor the following message is listed at the top of the screen:
    Central data cache configured but not active
    Suggestions on how to configure this cache and the impact of not doing so will be rewarded.
    Regards,
    Sander

    Hi,
    central data cache to improve the performance of the Alert Monitor when creating the alert monitoring trees.
    Check [This link|http://help.sap.com/saphelp_sm32/helpdata/en/de/13622d770b5e4f8398fc536b690e93/content.htm]
    And [This link|http://help.sap.com/saphelp_sm32/helpdata/en/10/75ea3b6d73e807e10000000a11402f/content.htm]
    Hope this solves your problem.
    Feel free to revert back.
    --Ragu

  • IFS and Data Cache Configuration

    Hi!!
    i installed 9iAs Enterprice Edition on my machine that already had Oracle db 8.1.7, i'm trying to set up ifs, but when i choose the option "Oracle8i is on THIS machine" in the ifs wizard i've got a protocol adapter error, if i choose the option "oracle8i on another machine" i solved the problem, why is this happening???
    the other problem it's with theb database cache i started the wizard, but in 16% when it's creating the cache it stoped it didn't show any error, i leaved for 2 days but it just reached the 16% and stayed ther.
    could you please help me..

    For iFS configuration on Enterprise Edition you can't select the option, "Oracle8i is on THIS machine". Actually this is a typo. This should be "Oracle8i is on THIS oracle_home". This will be fixed in future releases. iFS can't be configured against database cache(icache) also. So definitly you need a database which is on another oracle_home and you should select the second option, "oracle8i on another machine".
    From the second problem, I assume you are installing 9iAS on a Windows NT platform. On NT, sometimes database cache hangs. I assume it is mainly because of less memory. How much is the memory in your system? Are you running so many processes on the system while cache configuration? If so, reduce it as much as you can and retry for cache configuration. You may need to manually shutdown one SQLPLUS.EXE process in 'Task manager' started from the current cache configuration before retrying the config.

  • Include in cache configuration?

    I would like to define a "corporate" cache configuration file that later can be included and extended by project cache configuration files. I have seen some hints that this is supported by the Common package in the incubator but I have not seen any examples or documentation about how to do it. I know that I must specify the ExtensibleEnvironment as initialisation handler but what does the XML required look like? Can (should) both the included and the including cache configuration files be complete cache configuration files (verifyable against the XML schema) or should the included file just be a fragment that is legal to include at the point where it is inserted?
    Any links or examples are very much appreciated!
    Note I am not an XML guru so try to keep the explanation simple and try to provide complete examples not just "snippets" - to get all the new schema XML stuff to just validate without the added complexity of commons name spaces is a challenge to me :-)
    /Magnus
    Edited by: MagnusE on Jul 13, 2011 6:04 PM

    Hi user738616,
    I have the same MagnusE problem,
    but in my case I have both the coherence-common jar inside my classpath and both the introduce namespace correctly defined inside my configuration file:
    Here is my cache config file, named "poc-coherence-cache-config.xml":
    <cache-config
         xmlns:introduce="class://com.oracle.coherence.environment.extensible.namespaces.IntroduceNamespaceContentHandler"
         >
         <introduce:config file="coherence-messagingpattern-cache-config.xml"/>
         <caching-scheme-mapping>
         </caching-scheme-mapping>
         <caching-schemes>
         </caching-schemes>
    </cache-config>Which I load using defining the system property to load this override file:"poc-tangosol-coherence-override.xml", which is as follows:
    <?xml version='1.0'?>
    <coherence>
         <configurable-cache-factory-config>
              <class-name>com.oracle.coherence.environment.extensible.ExtensibleEnvironment
              </class-name>
              <init-params>
                   <init-param>
                        <param-type>java.lang.String</param-type>
                        <param-value system-property="tangosol.coherence.cacheconfig">
                        poc-coherence-cache-config.xml</param-value>
                   </init-param>
              </init-params>
         </configurable-cache-factory-config>
         <cluster-config>
              <member-identity>
                   <cluster-name system-property="tangosol.coherence.cluster">
                        Coherence_PoC_Cluster
                   </cluster-name>
              </member-identity>
              <multicast-listener>
                   <time-to-live system-property="tangosol.coherence.ttl">1</time-to-live>
                   <join-timeout-milliseconds>5000</join-timeout-milliseconds>
                   <address system-property="tangosol.coherence.clusteraddress">239.1.5.0</address> <!-- 224.13.16.10 -->
                   <port system-property="tangosol.coherence.clusterport">3030</port>
              </multicast-listener>
         </cluster-config>
         <logging-config>
              <destination system-property="tangosol.coherence.log">log4j</destination>
              <severity-level system-property="tangosol.coherence.log.level">10</severity-level>
              <character-limit system-property="tangosol.coherence.log.limit">0</character-limit>
         </logging-config>
    </coherence>But I still get the same exception:
    2011-08-03 11:48:54,416 INFO [Logger@8079107 3.7.0.0] Coherence [Log4j.CDB:3] 2011-08-03 11:48:54.320/2.246 Oracle Coherence 3.7.0.0 <Info> (thread=main, member=n/a): Loaded operational configuration from "jar:file:/C:/Users/.../.m2/repository/com/oracle/oracle-coherence/3.7.0/oracle-coherence-3.7.0.jar!/tangosol-coherence.xml"
    2011-08-03 11:48:54,417 INFO [Logger@8079107 3.7.0.0] Coherence [Log4j.CDB:3] 2011-08-03 11:48:54.322/2.248 Oracle Coherence 3.7.0.0 <Info> (thread=main, member=n/a): Loaded operational overrides from "file:/C:/.../coherence-config/target/classes/poc-tangosol-coherence-override.xml"
    2011-08-03 11:48:54,498 INFO [Logger@8079107 3.7.0.0] Coherence [Log4j.CDB:3] 2011-08-03 11:48:54.494/2.420 Oracle Coherence GE 3.7.0.0 <Info> (thread=main, member=n/a): Loaded cache configuration from "file:/C:/.../coherence-config/target/classes/poc-coherence-cache-config.xml"
    Exception in thread "main" (Wrapped: Failed to load the factory) (Wrapped: Missing or inaccessible constructor "com.oracle.coherence.environment.extensible.ExtensibleEnvironment(String)"
    <configurable-cache-factory-config>
      <class-name>com.oracle.coherence.environment.extensible.ExtensibleEnvironment</class-name>
      <init-params>
        <init-param>
          <param-type>java.lang.String</param-type>
          <param-value>poc-coherence-cache-config.xml</param-value>
        </init-param>
      </init-params>
    </configurable-cache-factory-config>) java.lang.reflect.InvocationTargetException
         at com.tangosol.util.Base.ensureRuntimeException(Base.java:288)
         at com.tangosol.net.ScopedCacheFactoryBuilder.getDefaultFactory(ScopedCacheFactoryBuilder.java:308)
         at com.tangosol.net.DefaultCacheFactoryBuilder.getSingletonFactory(DefaultCacheFactoryBuilder.java:48)
         at com.tangosol.net.DefaultCacheFactoryBuilder.getFactory(DefaultCacheFactoryBuilder.java:121)
         at com.tangosol.net.ScopedCacheFactoryBuilder.getConfigurableCacheFactory(ScopedCacheFactoryBuilder.java:109)
         at com.tangosol.net.CacheFactory.getConfigurableCacheFactory(CacheFactory.java:126)
         at com.tangosol.net.CacheFactory.getCache(CacheFactory.java:204)
         at com.tangosol.net.CacheFactory.getCache(CacheFactory.java:181)
         at poc.datagrid.coherence.impl.PartitionCoherence.<init>(PartitionCoherence.java:26)
         at poc.datagrid.coherence.impl.GridCoherence.get(GridCoherence.java:14)
         at poc.eventcollector.jmx.TournamentBean.<init>(TournamentBean.java:19)
         at poc.eventcollector.main.EventCollectorMain.main(EventCollectorMain.java:52)
    Caused by: (Wrapped: Missing or inaccessible constructor "com.oracle.coherence.environment.extensible.ExtensibleEnvironment(String)"
    <configurable-cache-factory-config>
      <class-name>com.oracle.coherence.environment.extensible.ExtensibleEnvironment</class-name>
      <init-params>
        <init-param>
          <param-type>java.lang.String</param-type>
          <param-value>poc-coherence-cache-config.xml</param-value>
        </init-param>
      </init-params>
    </configurable-cache-factory-config>) java.lang.reflect.InvocationTargetException
         at com.tangosol.util.Base.ensureRuntimeException(Base.java:288)
         at com.tangosol.run.xml.XmlHelper.createInstance(XmlHelper.java:2608)
         at com.tangosol.run.xml.XmlHelper.createInstance(XmlHelper.java:2492)
         at com.tangosol.net.ScopedCacheFactoryBuilder.getDefaultFactory(ScopedCacheFactoryBuilder.java:270)
         ... 10 more
    Caused by: java.lang.reflect.InvocationTargetException
         at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
         at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
         at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
         at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
         at com.tangosol.util.ClassHelper.newInstance(ClassHelper.java:694)
         at com.tangosol.run.xml.XmlHelper.createInstance(XmlHelper.java:2567)
         ... 12 more
    Caused by: java.lang.RuntimeException: Can't instantiate the NamespaceContentHandler [com.oracle.coherence.environment.extensible.namespaces.IntroduceNamespaceContentHandler] as the class is not found
         at com.oracle.coherence.environment.extensible.DefaultConfigurationContext$Scope.establishNamespaceContentHandlerFor(DefaultConfigurationContext.java:1145)
         at com.oracle.coherence.environment.extensible.DefaultConfigurationContext.ensureNamespaceContentHandler(DefaultConfigurationContext.java:119)
         at com.oracle.coherence.environment.extensible.DefaultConfigurationContext.processElement(DefaultConfigurationContext.java:282)
         at com.oracle.coherence.environment.extensible.DefaultConfigurationContext.processDocument(DefaultConfigurationContext.java:230)
         at com.oracle.coherence.environment.extensible.ExtensibleEnvironment.setConfig(ExtensibleEnvironment.java:469)
         at com.tangosol.net.DefaultConfigurableCacheFactory.<init>(DefaultConfigurableCacheFactory.java:190)
         at com.tangosol.net.DefaultConfigurableCacheFactory.<init>(DefaultConfigurableCacheFactory.java:176)
         at com.tangosol.net.DefaultConfigurableCacheFactory.<init>(DefaultConfigurableCacheFactory.java:152)
         at com.oracle.coherence.environment.extensible.ExtensibleEnvironment.<init>(ExtensibleEnvironment.java:155)
         ... 18 more
    Caused by: java.lang.ClassNotFoundException: com.oracle.coherence.environment.extensible.namespaces.IntroduceNamespaceContentHandler
         at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
         at java.security.AccessController.doPrivileged(Native Method)
         at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
         at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
         at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
         at java.lang.ClassLoader.loadClass(ClassLoader.java:247)
         at java.lang.Class.forName0(Native Method)
         at java.lang.Class.forName(Class.java:169)
         at com.tangosol.util.ExternalizableHelper.loadClass(ExternalizableHelper.java:3056)
         at com.oracle.coherence.environment.extensible.DefaultConfigurationContext$Scope.establishNamespaceContentHandlerFor(DefaultConfigurationContext.java:1109)
         ... 26 moreDo you have any ideas ?

  • AE CS6's Global Cache and Premiere Pro

    Not sure if this is an AE or Premiere question...
    Is Premiere Pro CS6 aware of AE's Global Cache? If I have a comp that is cached in AE, and I use Dynamic Link to import that comp into Premiere Pro, does it know that it's already rendered to the cache? Or does Premiere Pro tell AE to re-render it from scratch?

    It's not that Premiere Pro knows anything about the cache. Rather, when Premiere Pro asks After Effects to render the frames and send them over Dynamic Link, the headless version of After Effects can retrieve the frames from the cache if they're there. This does, in fact, make Dynamic Link workflows much faster in CS6.
    Lutz, what bug are you referring to?
    BTW, mae sure that you've installed the latest updates (choose Help > Updates). The After Effects updates include a lot of fixes relating to the cache.

Maybe you are looking for

  • Is it possible to generate Html and Pdf report format  for multi byte charecter ???

    Hi, I have gererated a PDF report for Chines Simplified, i set the NLS_LANG = AMERICAN_AMERICA.ZHT16BIG5 in my report server, the PDF format report is working fine, But the HTML report comes junk , Suppose if i change the NLS_LANG to AMERICAN_AMERICA

  • Switching from Premiere Elements to FCE

    Greetings ... I'll very likely be switching over to an iMac soon. As I do a fair amount of video editing, I'm looking at FCE. I've been using Adobe's Premiere Elements (the consumer version of Premiere) for a number of years and was just wondering if

  • Background VI question

    I need to have a VI running on the background constantly monitoring a value measured by DAQ. I want to put the value on top of the screen, and some other VI need to pull the value from this background VI from time to time. I guess I can use a backgro

  • I chose not to renew my .Mac account

    I decided not to renew my .Mac account this year. There were two main reasons for this. Firstly I was very disappointed with the service; I was unable to receive my mail or to access my home computer during the time I was traveling last year in N and

  • DataGrid scroll positions resetting with dynamic dataprovider

    Hi - I have a DataGrid that is backed by an ArrayCollection that is fairly dynamic. Before doing my updates, i disable autoupdating, and though my size is typically fixed at 100 items, sometimes those items will be replaced. My problem is that a user