Cache inserts not reflected across different readonly cache

I am a new developer to JavaBDB. I have designed a process as below :
1. A java BDB cache loader which loads data to a environment with a home path in write mode.
2. A java BDB cache reader which point to same home path and in readonly mode.
I load data using loader and then strart the reader it works fine. I can fetch data.
Now i start reader and then load additional data, i am not able to get this newly loaded data using reader until i restart my reader.
Can you please help me  ?
Thanks in advance.

Hi Cindy,
Thanks for your time
I am using Java api and version is 5.0.73.
I dont have a reproducer program to share but my use case is as below.
I have a java program which populates data in cache store and then there is another java process which is trying to read out of this store.
So basically i am accessing same store from different jvm processes one with read only environment and another with read_write environment.
The newly added elements to store by the writing process are not available in the reader process (different jvm process) untill i restart the reader.
I hope i am able to add more clarity.

Similar Messages

  • Webmail settings not reflected in different URLs

    I'm working on setting up an OS X Lion server to handle email. I have the basic email working in and out from the Internet. That is all working fine. However, I'm having an issue with the webmail app, which is now Roundcube. Specifically, webmail is accessible from several URLs by default.
    www.example.com/webmail
    servername.example.com/webmail
    If I log into www.example.com/webmail, it works. If I go into the settings and change something in there, that works too. Now, if I logout, and then go to servername.example.com/webmail and login with the same username and password, the changes I made are not reflected here. This is a pain in the butt. I can't control what URL users are ultimately going to type in.
    Has anyone else seen this? Am I missing something? I've been banging by head against the wall on this for a while.

    Try this. Something I tried a few months ago ...
    Downloaded the 6.1 plug-in updater here: http://www.adobe.com/support/downloads/thankyou.jsp?ftpID=4728&fileID= 4398
    Unzipped the file and found two relevant folders in 'payloads':
    AdobeCameraRaw6.0All-210510124504 contained the 32-bit plug-in
    AdobeCameraRaw6.0All-x64-210510130426 contained the 64-bit plug-in
    Within these folders was a filed called Assets2_1.zip
    Unzipped the appropriate files and extracted a single file called 1003 or 1002
    Renamed the file to 'Camera Raw.8bi'
    Backed-up the old plug-ins by renaming and inserting a tilde (~) at the start
    Copied the 6.1 version into the same folder:
    (64-bit) Program Files\Common Files\Adobe\Plug-ins\File Formats\
    and/or
    (32-bit) Program Files (x86)\Common Files\Adobe\Plug-ins\File Formats\
    Message was edited by: Yammer P
    Reformatted reply, because iPhone struggled with Jive interface.

  • PSP clear cache not reflecting in all Languages

    Calling siebel forum!!..
    Clear cache of PSP is not taking effect in SVE application
    However works well in ENU. We have done clear cache from ENU but should
    Work across all server and language deployements.
    This will work after server restart but is this a known issue,
    How can it be traced and corrected?
    Thanks!!

    After doing changes in the German Form in transaction SE63, the changes do not reflect in English language after activating the german form.Please help me with your valuable suggestion to solve this problem.
    Regards
    Ankur Godre

  • Cache Insert for an already existing key is firing two update events

    Using an INamedCache I perform an insert to update the values of my item that is already in the cache. I have listeners on the cache, in particular I have a DelegatingCacheListener that listens on EntryUpdated.
    On doing this insert once (and only once, I think!) my update event handler gets invoked twice. I have put breakpoints in all conceivable places that might raise this event, do an insert or update this key but it only ever hits my insertion point once.
    I have looked at the args of the event and seen that on the first invocation my old value is A and new value is B. For the second invocation my old value is B and New value is C. The second invocation new value object is only different by a single member which is a time value that I set and this difference is only a second. This kind of implies that somehow my own application/code is inserting twice into the cache for the same key/value.
    Has anyone seen this symptom before?
    Is there somehow that I can see what is inserting into the cache?
    Let me know if I can provide more information to help.
    Edited by: Kunal on Feb 9, 2011 8:08 PM
    Changed the title grammar

    Hi Kunal -
    First, just to narrow this down, it appears that your code is running in .NET. Is there other logic deployed as part of this application, e.g. application logic on the server side (the "back end") that might be doing something? This would include anything from the incubator (e.g. push replication).
    Second, what does your listener do when it receives the event?
    Third, are there other listeners for the event that may be doing something when the first change appears?
    The event originates from within the cluster, from the server that "owns" the specific piece of data that your are inserting, so unless the value is changing there twice, you should not get two events. I would start by tracking down all of the places that change the time value, since you can see that is the value that is being changed.
    Peace,
    Cameron Purdy | Oracle Coherence
    http://coherence.oracle.com/

  • Different distributed caches within the cluster

    Hi,
    i've three machines n1 , n2 and n3 respectively that host tangosol. 2 of them act as the primary distributed cache and the third one acts as the secondary cache. i also have weblogic running on n1 and based on some requests pumps data on to the distributed cache on n1 and n2. i've a listener configured on n1 and n2 and on the entry deleted event i would like to populate tangosol distributed service running on n3. all the 3 nodes are within the same cluster.
    i would like to ensure that the data directly coming from weblogic should only be distributed across n1 and n2 and NOT n3. for e.g. i do not start an instance of tangosol on node n3. and an object gets pruned from either n1 or n2. so ideally i should get a storage not configured exception which does not happen.
    The point is the moment is say CacheFactory.getCache("Dist:n3") in the cache listener, tangosol does populate the secondary cache by creating an instance of Dist:n3 on either n1 or n2 depending from where the object has been pruned.
    from my understanding i dont think we can have a config file on n1 and n2 that does not have a scheme for n3. i tried doing that and got an illegalstate exception.
    my next step was to define the Dist:n3 scheme on n1 and n2 with local storage false and have a similar config file on n3 with local-storage for Dist:n3 as true and local storage for the primary cache as false.
    can i configure local-storage specific to a cache rather than to a node.
    i also have an EJB deployed on weblogic that also entertains a getData request. i.e. this ejb will also check the primary cache and the secondary cache for data. i would have the statement
    NamedCahe n3 = CacheFactory.getCache("n3") in the bean as well.

    Hi Jigar,
    i've three machines n1 , n2 and n3 respectively that
    host tangosol. 2 of them act as the primary
    distributed cache and the third one acts as the
    secondary cache.First, I am curious as to the requirements that drive this configuration setup.
    i would like to ensure that the data directly coming
    from weblogic should only be distributed across n1
    and n2 and NOT n3. for e.g. i do not start an
    instance of tangosol on node n3. and an object gets
    pruned from either n1 or n2. so ideally i should get
    a storage not configured exception which does not
    happen.
    The point is the moment is say
    CacheFactory.getCache("Dist:n3") in the cache
    listener, tangosol does populate the secondary cache
    by creating an instance of Dist:n3 on either n1 or n2
    depending from where the object has been pruned.
    from my understanding i dont think we can have a
    config file on n1 and n2 that does not have a scheme
    for n3. i tried doing that and got an illegalstate
    exception.
    my next step was to define the Dist:n3 scheme on n1
    and n2 with local storage false and have a similar
    config file on n3 with local-storage for Dist:n3 as
    true and local storage for the primary cache as
    false.
    can i configure local-storage specific to a cache
    rather than to a node.
    i also have an EJB deployed on weblogic that also
    entertains a getData request. i.e. this ejb will also
    check the primary cache and the secondary cache for
    data. i would have the statement
    NamedCahe n3 = CacheFactory.getCache("n3") in the
    bean as well.In this scenario, I would recommend having the "primary" and "secondary" caches on different cache services (i.e. distributed-scheme/service-name). Then you can configure local storage on a service by service basis (i.e. distributed-scheme/local-storage).
    Later,
    Rob Misek
    Tangosol, Inc.

  • How to tell if loaded file came from cache or not?

    Trying to work out how I can tell if a downloaded file came from the cache or was freshly downloaded over the net. I need this because users will download multiple files and I am trying to give make an accurate downloader for them. I have made them download a sample file which I have made not to cache by doing the following:
    private var _loader:URLLoader;
    private var _rootPath:String; = "http:/example.com/speedTest/f4v/";
    private var _startDownloadTime:int;
    private var _targetSpeed:Number //set elsewhere
    private function loadTestSpeedFile():void
                var fileRequest:URLRequest = new URLRequest();
                fileRequest.url = _fullFilePath + "speedTest.f4v";
                var variables:URLVariables = new URLVariables(); 
                variables.nocache = new Date().getTime();
                fileRequest.data = variables;
                _startDownloadTime = flash.utils.getTimer();
                _loader.load(fileRequest);
    private function speedTestHandler(e:Event):void
                var endDownloadTime:int = flash.utils.getTimer();
                var timeElapsed:Number =   endDownloadTime - _startDownloadTime;
                timeElapsed = timeElapsed/1000;
                var bytesPerSec:Number = (SIZE_OF_TEST_FILE/timeElapsed);
                var currentSpeedFactor:Number = bytesPerSec/_targetSpeed;
                trace("Speed Test file downloaded in " + timeElapsed + " secs at " + bytesPerSec + " bytes a sec : speedFactor " + currentSpeedFactor);
                   //remove this handler and download as normal from a list of files
                _loader.removeEventListener(Event.COMPLETE, speedTestHandler);
                _loader.addEventListener(Event.COMPLETE, completeHandler);
              //download the rest of the files
                download();
    I could work out if the other files are downloading much more quickly (say 20X quicker) than the speedTest rate worked out here and then discount it from calculating the download speed and estimated time remaining.
    However, it would be much better if there was a way to access properties on the URLoader on the Event.COMPLETE handler to see if the file 'downloaded' was taken from the cache or not. Anyone know of way of doing it more elegantly than the hack described in the last sentence?

    Kglad,
    Thanks for the interest and reply. Yes, you are right the file load preventing the file being cached is stictly a different matter from that which is bugging me.  I am, however,  using the Progress Event (as in flash.events.ProgressEvent ) in the following event handler listening to it:
    private function progressHandler(e:ProgressEvent):void
                var percent:Number = (e.bytesLoaded/e.bytesTotal)*100;
                  //other code to resize the gui loader ie scaling it to the percentage.
    Again this just tells me how much has been loaded each time the event is fired. On the Event.COMPLETE handler I am getting the load timings pretty much as per the
    speedTestHandler(e:Event) I posted. The only difference there is that it loads from an array of filenames (most but not all of which can change on later replays - hence some will be cached and some not) What i need to know is where the file is loaded from - either from the server or the local systems cache. The only way round this is to discard very quick downloads (which must be from the cache) from true calculating download speeds but that is a bit hit and miss. I am asking for a better solution. Hope that makes sense...

  • Cache entry not found error message in java console 1.6.0_05 - Citrix ICA

    Client information:_
    Windows XP_SP2
    Java Runtime version 1.6.0_05
    Application: Citrix ICA version 9.4.1868
    Slow citrix ICA client connection, repeated errors in the java console stating "cache entry not found". However when I downgrade to Java Runtime version 1.5.0_10 I do not see the "cache entry not found" errors and the Citrix ICA connection is much faster. Basically launches and connects in 10 seconds versus 2 minutes.
    Any ideas? Thanks!

    Hi,
    All your classes must be accessible through the web. The easiest solution is to put all your classes in the same folder as your web page.
    If your classes are in a different folder (which must be mapped as a virtual directory), try adding the codebase attribute to your applet.
    Regards,
    Kurt.

  • Calendar cache could not be moved because...permission

    after updates associated is iOS5 and icloud iCal won't open.  i get the message: Calendar cache could not be moved because you do not have permission to access calendars.  I checkd permission and my users wasn't there.  i added it, and no help.  I also repaired permissions from disk utility.  I quick search looks like this has been a common problem with most every new verion of 10.x..but the fixes are slightly different.  anyone have a fix for lion? 

    Dave,
    Welcome to Apple Discussions.
    Have you ever tried to change the "Sharing & Permissions" setting of any files/folders on your Mac?
    Where the heck do I find this mythical "Calendars" item?
    Open your Macintosh HD/Users/yourusername/Library/Calendars Folder. Using "Get Info," what is the Sharing & Permissions setting for the Calendar Cache file in that folder?
    What is the Sharing & Permissions setting for the Calendars Folder.
    I've already tried re-installing iCal with my Snow Leopard install disc and it did no good.
    Since you have reinstalled iCal you should also download and reinstall the Mac OS X v10.6.3 v1.1 Update (Combo)

  • XAER_RMERR error when configuring readonly cache group

    Hello,
    I am having some problem with XA transaction controll on websphere.
    error log looks like below.
    ====================================================================
    [09. 6. 26 13:28:15:511 KST] 0000002b ServiceLogger I com.ibm.ws.ffdc.IncidentStreamImpl open FFDC0009I: FFDC가 /was/WebSphere61/AppServer/profiles/AppSrv01/logs/ffdc/server1_0000002b_09.06.26_13.28.15_2.txt 문제 스트림 파일을 열었습니다.
    [09. 6. 26 13:28:15:536 KST] 0000002b ServiceLogger I com.ibm.ws.ffdc.IncidentStreamImpl resetIncidentStream FFDC0010I: FFDC가 /was/WebSphere61/AppServer/profiles/AppSrv01/logs/ffdc/server1_0000002b_09.06.26_13.28.15_2.txt 문제 스트림 파일을 닫았습니다.
    [09. 6. 26 13:28:15:551 KST] 0000002b ServiceLogger I com.ibm.ws.ffdc.IncidentStreamImpl open FFDC0009I: FFDC가 /was/WebSphere61/AppServer/profiles/AppSrv01/logs/ffdc/server1_0000002b_09.06.26_13.28.15_3.txt 문제 스트림 파일을 열었습니다.
    [09. 6. 26 13:28:15:555 KST] 0000002b ServiceLogger I com.ibm.ws.ffdc.IncidentStreamImpl resetIncidentStream FFDC0010I: FFDC가 /was/WebSphere61/AppServer/profiles/AppSrv01/logs/ffdc/server1_0000002b_09.06.26_13.28.15_3.txt 문제 스트림 파일을 닫았습니다.
    [09. 6. 26 13:28:15:557 KST] 0000002b RegisteredRes E WTRN0078E: 트랜잭션 관리자가 트랜잭션 자원 시작을 호출하려는 중에 오류가 발생했습니다. 오류 코드는 XAER_RMERR입니다. 예외 스택 추적은 다음과 같습니다: javax.transaction.xa.XAException: errorCode=XAER_RMERR, a resource manager error has occured in the transaction branch.
         at com.timesten.jdbc.xa.XAJdbcOdbc.XAStandardError(XAJdbcOdbc.java:298)
         at com.timesten.jdbc.xa.XAJdbcOdbc.SQLXAStart(XAJdbcOdbc.java:67)
         at com.timesten.jdbc.xa.TimesTenXAResource.start(TimesTenXAResource.java:273)
         at com.ibm.ws.rsadapter.spi.WSRdbXaResourceImpl.start(WSRdbXaResourceImpl.java:1417)
         at com.ibm.ejs.j2c.XATransactionWrapper.start(XATransactionWrapper.java:1467)
         at com.ibm.ws.Transaction.JTA.JTAResourceBase.start(JTAResourceBase.java:145)
         at com.ibm.ws.Transaction.JTA.RegisteredResources.startRes(RegisteredResources.java:1240)
         at com.ibm.ws.Transaction.JTA.RegisteredResources.enlistResource(RegisteredResources.java:648)
         at com.ibm.ws.Transaction.JTA.TransactionImpl.enlistResource(TransactionImpl.java:3294)
         at com.ibm.ws.Transaction.JTA.TranManagerSet.enlist(TranManagerSet.java:405)
         at com.ibm.ejs.j2c.XATransactionWrapper.enlist(XATransactionWrapper.java:693)
         at com.ibm.ejs.j2c.ConnectionManager.lazyEnlist(ConnectionManager.java:1909)
         at com.ibm.ws.rsadapter.spi.WSRdbManagedConnectionImpl.lazyEnlist(WSRdbManagedConnectionImpl.java:2219)
         at com.ibm.ws.rsadapter.jdbc.WSJdbcConnection.beginTransactionIfNecessary(WSJdbcConnection.java:643)
         at com.ibm.ws.rsadapter.jdbc.WSJdbcConnection.prepareStatement(WSJdbcConnection.java:2083)
         at com.ibm.ws.rsadapter.jdbc.WSJdbcConnection.prepareStatement(WSJdbcConnection.java:2038)
         at com.huni.framework.front.dao.HmmDAO.showPrg(HmmDAO.java:181)
         at com.huni.framework.ejbcommand.CommandExecuter.initCommand(CommandExecuter.java:53)
         at com.huni.framework.ejbcommand.CommandExecuter.executeCommand(CommandExecuter.java:27)
         at com.huni.framework.front.bean.HMMFacadeBean.executeCommand(HMMFacadeBean.java:104)
         at com.huni.framework.front.bean.EJSRemoteStatelessHMMFacade_905ddb17.executeCommand(Unknown Source)
         at com.huni.framework.front.bean._HMMFacade_Stub.executeCommand(_HMMFacade_Stub.java:267)
         at com.huni.common.servlet.HmmServlet.doPost(HmmServlet.java:103)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:763)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:856)
         at com.ibm.ws.webcontainer.servlet.ServletWrapper.service(ServletWrapper.java:1143)
         at com.ibm.ws.webcontainer.servlet.ServletWrapper.handleRequest(ServletWrapper.java:591)
         at com.ibm.ws.wswebcontainer.servlet.ServletWrapper.handleRequest(ServletWrapper.java:481)
         at com.ibm.ws.webcontainer.webapp.WebApp.handleRequest(WebApp.java:3453)
         at com.ibm.ws.webcontainer.webapp.WebGroup.handleRequest(WebGroup.java:267)
         at com.ibm.ws.webcontainer.WebContainer.handleRequest(WebContainer.java:815)
         at com.ibm.ws.wswebcontainer.WebContainer.handleRequest(WebContainer.java:1466)
         at com.ibm.ws.webcontainer.channel.WCChannelLink.ready(WCChannelLink.java:119)
         at com.ibm.ws.http.channel.inbound.impl.HttpInboundLink.handleDiscrimination(HttpInboundLink.java:458)
         at com.ibm.ws.http.channel.inbound.impl.HttpInboundLink.handleNewInformation(HttpInboundLink.java:387)
         at com.ibm.ws.http.channel.inbound.impl.HttpInboundLink.ready(HttpInboundLink.java:267)
         at com.ibm.ws.tcp.channel.impl.NewConnectionInitialReadCallback.sendToDiscriminators(NewConnectionInitialReadCallback.java:214)
         at com.ibm.ws.tcp.channel.impl.NewConnectionInitialReadCallback.complete(NewConnectionInitialReadCallback.java:113)
         at com.ibm.ws.tcp.channel.impl.AioReadCompletionListener.futureCompleted(AioReadCompletionListener.java:165)
         at com.ibm.io.async.AbstractAsyncFuture.invokeCallback(AbstractAsyncFuture.java:217)
         at com.ibm.io.async.AsyncChannelFuture.fireCompletionActions(AsyncChannelFuture.java:161)
         at com.ibm.io.async.AsyncFuture.completed(AsyncFuture.java:136)
         at com.ibm.io.async.ResultHandler.complete(ResultHandler.java:196)
         at com.ibm.io.async.ResultHandler.runEventProcessingLoop(ResultHandler.java:751)
         at com.ibm.io.async.ResultHandler$2.run(ResultHandler.java:881)
         at com.ibm.ws.util.ThreadPool$Worker.run(ThreadPool.java:1473)
    =================================================================================
    By seeing the similar forum message like this which is saying that XA transactions are not allowed with replication configured, i am wondering whether readonly cache group is relevant with this.
    I is not allowed too??
    best regards,
    Yu-Mi

    Cache Connect is not supported with XA transactions. This is mentioned in the Docs, albeit only in the Error Messages section (Error 11037 per my 7.0.3. copy).

  • Getting error:  java.lang.IllegalStateException: Service has been started by a different configurable cache factory.

    We are trying to implement host name based security for our Coherence cluster. This is mainly to prevent processes running in one environment (UAT for example) connecting to the Coherence cluster running in another environment (PROD for example).
    We wanted to ensure that 1) Only Coherence nodes running on specific machines can join the cluster as members. To accomplish this, we added the list of authorized hosts to the override file (see marketrisk-coherence-override.xml) This works as expected with no issues
    We also want to ensure that we allow client connections from some specific machines. All of our clients connect to the cluster through the proxy. So, we added added a host filter to the cache config file (see marketrisk-cache-config.xml) to control the list of client machines that can are allowed to connect to the cluster. The host filter is pretty simple. If the incoming host address is found in a cache, it allows the connection else it refuses the connection. Please see AuthorizedClientHostsFilter.java for more details.
    When we start the cluster, the storage nodes are failing to come up with the following error message:
    Exception in thread "main" java.lang.IllegalStateException: Service "ReferenceDataCacheService" has been started by a different configurable cache factory.
                    at com.tangosol.net.ExtensibleConfigurableCacheFactory.validateBackingMapManager(ExtensibleConfigurableCacheFactory.java:829)
                    at com.tangosol.net.ExtensibleConfigurableCacheFactory.ensureService(ExtensibleConfigurableCacheFactory.java:585)
                    at com.tangosol.net.ExtensibleConfigurableCacheFactory.startServices(ExtensibleConfigurableCacheFactory.java:666)
                    at com.tangosol.net.DefaultCacheServer.startServicesInternal(DefaultCacheServer.java:369)
                    at com.tangosol.net.DefaultCacheServer.initialStartServices(DefaultCacheServer.java:466)
                    at com.tangosol.net.DefaultCacheServer.startAndMonitor(DefaultCacheServer.java:74)
                    at com.tangosol.net.DefaultCacheServer.main(DefaultCacheServer.java:260)
    This error goes away and the cluster starts normally if we remove host filter from the cache config file or we modify AuthorizedClientHostsFilter.java to not get a handle to the cache when it is instantiated. Please see AuthorizedClientHostsFilter1.java which has this change.
    We believe this is happening because the storage node is creating two cache factories when we use the host filter (AuthorizedClientHostsFilter.java). The storage nodes creates only one cache factory if  we do not use the host filter or if I use AuthorizedClientHostsFilter1.java.
    I do not understand why this is happening. Can you please help? The attached zip file should have all the information that you will probably need.

    public class AuthorizedClientHostsFilter implements Filter {
        //private static final Logger logger = LoggerFactory.getLogger(AuthorizedClientHostsFilter.class);
        public static final String ALL_HOSTS = "ALL_HOSTS";
        private static NamedCache hostsCache = new ContinuousQueryCache(
                CacheFactory.getCache("client-auth-hosts"), AlwaysFilter.INSTANCE, false);
        public AuthorizedClientHostsFilter() {
        public boolean evaluate(Object o) {
            logger.info("hostsCache.size() = "+hostsCache.size());
            Set<String> authorizedHostsinCache = hostsCache.keySet();
            for (String hostName : authorizedHostsinCache) {
                logger.info("Host: "+hostName);
            if (hostsCache.containsKey("ALL_HOSTS")) return true;
            InetAddress address = (InetAddress) o;
            //logger.info("Incoming Host: "+address.getHostAddress());
            return hostsCache.containsKey(address.getHostAddress());
    public class AuthorizedClientHostsFilter1 implements Filter {
        //private static final Logger logger = LoggerFactory.getLogger(AuthorizedClientHostsFilter.class);
        public static final String ALL_HOSTS = "ALL_HOSTS";
        private static NamedCache hostsCache = null;
        public AuthorizedClientHostsFilter1() {
        public boolean evaluate(Object o) {
            if (hostsCache == null) {
                hostsCache = new ContinuousQueryCache(
                        CacheFactory.getCache("client-auth-hosts"), AlwaysFilter.INSTANCE, false);
            logger.info("hostsCache.size() = "+hostsCache.size());
            Set<String> authorizedHostsinCache = hostsCache.keySet();
            for (String hostName : authorizedHostsinCache) {
                logger.info("Host: "+hostName);
            if (hostsCache.containsKey("ALL_HOSTS")) return true;
            InetAddress address = (InetAddress) o;
            //logger.info("Incoming Host: "+address.getHostAddress());
            return hostsCache.containsKey(address.getHostAddress());

  • Autorefresh Limit - ReadOnly cache

    Hi,
    I have a readonly cachegroup with about 2million records, about 100 thousand records are updated in Oracle but this data is not refreshing in T10. My cache group parameters are...
    Cache Group Type: Read Only
    Autorefresh: Yes
    Autorefresh Mode: Incremental
    Autorefresh State: On
    Autorefresh Interval: 5 Minutes
    Autorefresh Limit: 10000
    what is the importance of Autorefresh Limit? What are the required settings for refreshing huge volumes of data?
    TIA..
    Regards,
    Prakash T

    Within what time period are the 100,000 records updated in Oracle?
    If you are using TimesTen 7.0 then the WITH LIMIT clause is redundant and will be ignored.
    If you are using TimesTen 6.0 or earlier, the WITH LIMIT clause sets an upper bound on the number of changes that can be tracked for INCREMENTAL autorefresh. If more than that number of changes occur within the autorefresh interval (5 minutes in your case) then a FULL refresh will be performed (i.e. all 2 million records will be pushed to the cache - ouch!)
    Do you have the TT cache agent running for the datastore? If not then that is probably why autorefresh is not working. If the agent appears to be running, consult the TimesTen daemon log and look for messages that will indicate why the refresh is not happening.
    Chris

  • Why is my cache results not highligted anymore?

    why is my cache results not highligted anymore?

    When I do a word/s search.
    Normally the words I typed and searched on Google will be highlighted in different colors each word at the top of the webpage.
    The same color will be highligted for each word/s whenever they appears all throught out the page. when I open it as cached.
    Now the words don't highlight any more and no color even.
    Here below info from About.com about what I am asking about:
    http://google.about.com/od/searchingtheweb/qt/cache_syntax.htm
    ''''''Highlight Keywords With Google Cache Search''''''
    Find Specific Information Faster With Google's Cache
    By Marziah Karch, About.com Guide
    Is it hard to find a specific piece of information on a large Web page? You can simply this by using Google's cached page to highlight your search term.
    As Google indexes Web pages, it retains a snapshot of the page contents, known as a cached page. When a cached page is available, you'll see a Cached link at the bottom of the search result.
    Clicking on the Cached link will show you the page as it was last indexed on Google, but with your search keywords highlighted. This is extremely useful if you want to find a specific piece of information without having to scan the entire page.
    Keep in mind that this shows the last time the page was indexed, so sometimes images will not show up and the information will be out of date. For most quick searches, that doesn't matter. You can always go back to the current version of the page and double check to see if the information has changed.

  • Cached data not available to clients

    Hi All ,
    I am loading table data to cache using putAll() . Data loads fine and if i try to access the data in same program , i can get it .
    but if i try to access the cached data with some other client it is not available (getting null pointer exceptions ) .
    is there any config change required to keep data in cache for other clients .I am using JPA beans for persistence .
    Code to access the data
    =========================================
    public static void main(String[] args) {   
    CacheFactory.ensureCluster();
    NamedCache currcache = CacheFactory.getCache("ClientMaster");
    //These two commented lines of code is used to load data in cache using putAll .
    // DataLoad data = new DataLoad();
    //data.preload(currcache);
    Set setKeys = currcache.keySet();
    Map mapResult = currcache.getAll(setKeys);
    ClientMaster cm =(ClientMaster)mapResult.get("ABC");
    System.out.println("key"+cm.getClientDesc());
    =========================================
    Here is my cache-config file .
    <?xml version="1.0" encoding="windows-1252" ?>
    <cache-config>
    <caching-scheme-mapping>
    <cache-mapping>
    <!-- Set the name of the cache to be the entity name -->
    <cache-name>ClientMaster</cache-name>
    <!-- Configure this cache to use the scheme defined below -->
    <scheme-name>jpa-distributed</scheme-name>
    </cache-mapping>
    <cache-mapping>
    <!-- Set the name of the cache to be the entity name -->
    <cache-name>CurrencyMaster</cache-name>
    <!-- Configure this cache to use the scheme defined below -->
    <scheme-name>jpa-distributed</scheme-name>
    </cache-mapping>
    </caching-scheme-mapping>
    <caching-schemes>
    <distributed-scheme>
    <scheme-name>jpa-distributed</scheme-name>
    <service-name>JpaDistributedCache</service-name>
    <backing-map-scheme>
    <read-write-backing-map-scheme>
    <!--
    Define the cache scheme
    -->
    <internal-cache-scheme>
    <local-scheme>
    </local-scheme>
    </internal-cache-scheme>
    <cachestore-scheme>
    <class-scheme>
    <class-name>com.tangosol.coherence.jpa.JpaCacheStore</class-name>
    <init-params>
    <!--
    This param is the entity name
    This param is the fully qualified entity class
    This param should match the value of the
    persistence unit name in persistence.xml
    -->
    <init-param>
    <param-type>java.lang.String</param-type>
    <param-value>{cache-name}</param-value>
    </init-param>
    <init-param>
    <param-type>java.lang.String</param-type>
    <param-value>project1.{cache-name}</param-value>
    </init-param>
    <init-param>
    <param-type>java.lang.String</param-type>
    <param-value>Project1</param-value>
    </init-param>
    </init-params>
    </class-scheme>
    </cachestore-scheme>
    </read-write-backing-map-scheme>
    </backing-map-scheme>
    <autostart>true</autostart>
    </distributed-scheme>
    </caching-schemes>
    </cache-config>

    Now Getting this error when trying load data in cache using putAll , This is really weird as this same piece of code worked fine earlier.
    Any ideas ?
    ============================================================
    2011-01-11 13:45:47.165/14.266 Oracle Coherence GE 3.6.1.0 <Error> (thread=Distr
    ibutedCache:JpaDistributedCache, member=6): BackingMapManager com.tangosol.net.D
    efaultConfigurableCacheFactory$Manager: failed to instantiate a cache: ClientMaster
    2011-01-11 13:45:47.165/14.266 Oracle Coherence GE 3.6.1.0 <Error> (thread=DistributedCache:JpaDistributedCache, member=6):
    (Wrapped: Missing or inaccessible constructor "com.tangosol.coherence.jpa.JpaCacheStore(String,String,String)"
    <class-scheme>
    <class-name>com.tangosol.coherence.jpa.JpaCacheStore</class-name>
    <init-params>
    <!--
    This param is the entity name
    This param is the fully qualified entity class
    This param should match the value of the
    persistence unit name in persistence.xml
    -->
    <init-param>
    <param-type>java.lang.String</param-type>
    <param-value>ClientMaster</param-value>
    </init-param>
    <init-param>
    <param-type>java.lang.String</param-type>
    <param-value>project1.ClientMaster</param-value>
    </init-param>
    <init-param>
    <param-type>java.lang.String</param-type>
    <param-value>Project1</param-value>
    </init-param>
    </init-params>
    </class-scheme>) java.lang.reflect.InvocationTargetException
    at com.tangosol.util.Base.ensureRuntimeException(Base.java:293)
    at com.tangosol.run.xml.XmlHelper.createInstance(XmlHelper.java:2542)
    at com.tangosol.run.xml.XmlHelper.createInstance(XmlHelper.java:2426)
    at com.tangosol.coherence.jpa.JpaCacheStore.<init>(JpaCacheStore.java:44
    ... 29 more
    Caused by: java.lang.ClassNotFoundException: project1.ClientMaster
    at java.net.URLClassLoader$1.run(URLClassLoader.java:200)
    at java.security.AccessController.doPrivileged(Native Method)
    at java.net.URLClassLoader.findClass(URLClassLoader.java:188)
    at java.lang.ClassLoader.loadClass(ClassLoader.java:307)
    at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
    at java.lang.ClassLoader.loadClass(ClassLoader.java:252)
    at com.tangosol.coherence.jpa.JpaCacheLoader.initialize(JpaCacheLoader.j
    ava:143)

  • DataGrid not reflecting changes after INSERT into Table. Delete from Table does.

    Wow, it's been a while.
    Hope you guys can help.
    This is my DataGrid:
    <DataGrid DataContext="{StaticResource TableAssetsViewSource}" ItemsSource="{Binding}" x:Name="TableAssetsDataGrid" AutoGenerateColumns="False" EnableRowVirtualization="True" Margin="15,10,10,10" RowDetailsVisibilityMode="VisibleWhenSelected" Grid.Column="1" HeadersVisibility="Column" CanUserResizeRows="False" IsReadOnly="True">
    <DataGrid.CellStyle>
    <Style TargetType="DataGridCell">
    <Setter Property="BorderThickness" Value="0"/>
    </Style>
    </DataGrid.CellStyle>
    <DataGrid.Columns>
    <DataGridTextColumn x:Name="NodeColumn" Binding="{Binding node}" Header="VS Number" Width="*"/>
    <DataGridTextColumn x:Name="SerialColumn" Binding="{Binding serial}" Header="Serial Number" Width="*"/>
    <DataGridTextColumn x:Name="NameColumn" Binding="{Binding name}" Header="Asset Name" Width="*"/>
    <DataGridTextColumn x:Name="TypeColumn" Binding="{Binding type}" Header="Asset Type" Width="*"/>
    <DataGridTextColumn x:Name="_dateColumn" Binding="{Binding date}" Header="Date Allocated" Width="*"/>
    <DataGridTextColumn x:Name="PoColumn" Binding="{Binding po}" Header="Purchase Order" Width="*"/>
    </DataGrid.Columns>
    </DataGrid>
    This is where I add a new Row to the Table:
    Private Sub Button_Click_1(sender As Object, e As RoutedEventArgs)
    Dim cbblocation As ComboBox = Me.FindName("LocationComboBox")
    Dim row As DataRowView = DirectCast(cbblocation.SelectedItem, DataRowView)
    Dim cbbtext As String = row.Item("node")
    Dim cbbtype As ComboBox = Me.FindName("comboBoxType")
    Dim cbbtext2 As String = cbbtype.Text
    Dim RETAILISOAMDDataSet As Retail_ISO_AMD.RETAILISOAMDDataSet = CType(Me.FindResource("RETAILISOAMDDataSet"), Retail_ISO_AMD.RETAILISOAMDDataSet)
    Dim RETAILISOAMDDataSetTableAssetsTableAdapter As Retail_ISO_AMD.RETAILISOAMDDataSetTableAdapters.tableAssetsTableAdapter = New Retail_ISO_AMD.RETAILISOAMDDataSetTableAdapters.tableAssetsTableAdapter()
    RETAILISOAMDDataSetTableAssetsTableAdapter.AddNewAsset(cbbtext, txbSerial.Text, txbName.Text, cbbtext2, Date.Today, txbPO.Text)
    node = cbbtext
    Dim mp As New MainPage
    mp.RefreshGrid(node)
    Me.Close()
    End Sub
    The RefreshGrid method:
    Public Sub RefreshGrid(node As String)
    Dim RETAILISOAMDDataSet As Retail_ISO_AMD.RETAILISOAMDDataSet = CType(Me.FindResource("RETAILISOAMDDataSet"), Retail_ISO_AMD.RETAILISOAMDDataSet)
    Dim RETAILISOAMDDataSetTableAssetsTableAdapter As Retail_ISO_AMD.RETAILISOAMDDataSetTableAdapters.tableAssetsTableAdapter = New Retail_ISO_AMD.RETAILISOAMDDataSetTableAdapters.tableAssetsTableAdapter()
    RETAILISOAMDDataSetTableAssetsTableAdapter.FillByNode(RETAILISOAMDDataSet.tableAssets, node)
    Dim TableAssetsViewSource As System.Windows.Data.CollectionViewSource = CType(Me.FindResource("TableAssetsViewSource"), System.Windows.Data.CollectionViewSource)
    Dim be As BindingExpression = BindingOperations.GetBindingExpression(TableAssetsViewSource, CollectionViewSource.SourceProperty)
    be.UpdateTarget()
    End Sub
    And this is what I use to delete a row from the table (and without doing anything special, the DataGrid auto-updates with the changes):
    Private Sub Button_Click_4(sender As Object, e As RoutedEventArgs)
    If TableAssetsDataGrid.SelectedIndex = -1 Then
    MsgBox("You have selected nothing to Delete!", MsgBoxStyle.OkOnly, "Nothing Selected")
    Else
    Dim dgv As DataGridTextColumn = Me.FindName("NodeColumn")
    Dim row As DataRowView = DirectCast(Me.TableAssetsDataGrid.SelectedItem, DataRowView)
    Dim dgvText As String = row.Item("node")
    Dim dgv2 As DataGridTextColumn = Me.FindName("SerialColumn")
    Dim row2 As DataRowView = DirectCast(Me.TableAssetsDataGrid.SelectedItem, DataRowView)
    Dim dgvText2 As String = row.Item("serial")
    Dim dgv3 As DataGridTextColumn = Me.FindName("NameColumn")
    Dim row3 As DataRowView = DirectCast(Me.TableAssetsDataGrid.SelectedItem, DataRowView)
    Dim dgvText3 As String = row.Item("name")
    Dim dgv4 As DataGridTextColumn = Me.FindName("TypeColumn")
    Dim row4 As DataRowView = DirectCast(Me.TableAssetsDataGrid.SelectedItem, DataRowView)
    Dim dgvText4 As String = row.Item("type")
    Dim dgv5 As DataGridTextColumn = Me.FindName("_dateColumn")
    Dim row5 As DataRowView = DirectCast(Me.TableAssetsDataGrid.SelectedItem, DataRowView)
    Dim dgvText5 As String = row.Item("date")
    Dim dgv6 As DataGridTextColumn = Me.FindName("POColumn")
    Dim row6 As DataRowView = DirectCast(Me.TableAssetsDataGrid.SelectedItem, DataRowView)
    Dim dgvText6 As String = row.Item("po")
    Dim RETAILISOAMDDataSet As Retail_ISO_AMD.RETAILISOAMDDataSet = CType(Me.FindResource("RETAILISOAMDDataSet"), Retail_ISO_AMD.RETAILISOAMDDataSet)
    Dim RETAILISOAMDDataSetTableAssetsTableAdapter As Retail_ISO_AMD.RETAILISOAMDDataSetTableAdapters.tableAssetsTableAdapter = New Retail_ISO_AMD.RETAILISOAMDDataSetTableAdapters.tableAssetsTableAdapter()
    RETAILISOAMDDataSetTableAssetsTableAdapter.RemoveAsset(dgvText, dgvText2, dgvText3, dgvText4, dgvText5, dgvText6)
    Dim cbb As ComboBox = Me.FindName("cbbLocation")
    Dim row7 As DataRowView = DirectCast(cbb.SelectedItem, DataRowView)
    Dim cbbtext As String = row.Item("node")
    RefreshGrid(cbbtext)
    End If
    End Sub
    --------- End of Edit
    It is bound to a Dataset which gets it's data from a SQL Database.
    Loading the Data and Filtering the data based on certain conditions work 100%. The problem I am having is as follows:
    I have a form on the Page that takes input and inserts a row into the Database. When this happens, the DataGrid won't automatically reflect the changes (even if I recall the Fill Method of the Dataset). No matter what I do, I have to refresh the entire page
    and THEN Fill the Dataset to see any changes.
    This is what strikes me as odd...
    When I do a delete row operation on the Database (Custom Method on the Dataset with conditions), and just Fill the Dataset again (without doing anything special), the row deletes and the changes is reflected IMMEDIATELY.
    What am I doing wrong here? Why would Delete Row reflect the changes on the DataGrid but no Insert Row?
    Thanks in Advance.
    (P.S. I am very rusty with my developing skills, I haven't done this in YEARS)

    >>And what about the the most important question, how exactly is TableAssetsViewSource defined in the XAML markup and what is
    its Source property set or bound to? And what about the reproducable sample...?
    Here is the Markup of the TableAssetsViewSource:
    <Page.Resources>
    <local:RETAILISOAMDDataSet x:Key="RETAILISOAMDDataSet"/>
    <CollectionViewSource x:Key="TableRegionsViewSource" Source="{Binding tableRegions, Source={StaticResource RETAILISOAMDDataSet}}"/>
    <CollectionViewSource x:Key="TableLocationsViewSource" Source="{Binding tableLocations, Source={StaticResource RETAILISOAMDDataSet}}"/>
    <CollectionViewSource x:Key="TableAssetsViewSource" Source="{Binding tableAssets, Source={StaticResource RETAILISOAMDDataSet}}"/>
    </Page.Resources>
    The ItemsSource of the DataGrid is bound to the TableAssetsViewSource:
    <DataGrid DataContext="{StaticResource TableAssetsViewSource}" ItemsSource="{Binding}" x:Name="TableAssetsDataGrid" AutoGenerateColumns="False" EnableRowVirtualization="True" Margin="15,10,10,10" RowDetailsVisibilityMode="VisibleWhenSelected" Grid.Column="1" HeadersVisibility="Column" CanUserResizeRows="False" IsReadOnly="True">
    Just remember, this code was auto-generated with the drag & drop onto the Page, so I did little in terms of setting the actual bindings.
    All in all, what I have done so far works as intended, except that when I want to add a row to the table it does not reflect, and that it only shows after refreshing/restarting the application.
    I will try and put together a reproducable sample.

  • Error ACLContainer: 65315 does NOT EXIST in  the Object Cache for parentID:

    Expert,
    I did the following steps to upgrade the OWB repository from 10g to 11g
    • Created a dummy workspace in the 11.2 repository
    • Created users in the destination environment
    • Run Repository Assistant against the 11.1 source database
    • Then selected *“Export entire repository to a file”* and selected the MDL(10g) to import
    After 99% of completing I have got the below error
    Error ACLContainer: 65315 does NOT EXIST in  the Object Cache for parentID: 65314
    Please let me know the solution.
    Thanks,
    Balaa...

    Hi I had the same error and worked on it for almost a week with no success but there is few work around for this try it it might work for you.
    Step 1>Run a health check on OWB 10.2 enviroment(make sure you clone your OWB 10.2 enviroment and then do this healthcheck ,these checks are tricky )
    refer
    Note 559542.1 Health Check of the Oracle Warehouse Builder 10.2 Metadata Repository
    This will give you info about your missing ACL
    Step 2> download these two scripts fixLostACLContainer.sql ,fixAllACLContainers.sql
    please refer :
    Note 559542.1 Health Check of the Oracle Warehouse Builder 10.2 Metadata Repository
    OWB 10.2 - Internal ERROR: Can not find the ACL container for object (Doc ID 460411.1)
    Note 754763.1 Repository Cleanup Script for OWB 10.2 and OWB 11.1
    Note 460411.1 Opening Map Returns Cannot find ACL Containter for Object
    Note 1165068.1 Internal Error: Can Not Find The ACL Containter For for object:CMPPhysical Object
    It might resolve this ACL issue but it did not work for me.
    If none of these work then
    Perform export from design center of OWB 10.2 and import through design center of OWB 11.2.(ONLY OPTION)
    It worked for me.
    Varun

Maybe you are looking for